added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:35:22.104557
| 2016-09-26T09:28:39
|
179179595
|
{
"authors": [
"challiwill",
"ktogo",
"robfig"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10382",
"repo": "robfig/cron",
"url": "https://github.com/robfig/cron/pull/76"
}
|
gharchive/pull-request
|
Fix: Cron.run() was resetting the timezone
time.Timer() does not return a timezone within time.Time, since Timer is unaware of the custom timezone as it only accepts time.Duration when it's initialized.
Thus the receiver needs to take care of the timezone.
// NewTimer only receives time.Duration
func NewTimer(d Duration) *Timer
// And time.Duration only holds int64 but no timezone specifier
type Duration int64
Interesting.. your change to the tests causes them to fail without your change to the code?
Great find, thanks
+1 I was about to try and track down what was happening here also. Good find.
|
2025-04-01T04:35:22.131692
| 2017-08-23T09:47:35
|
252222908
|
{
"authors": [
"AndiLeni",
"robiso"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10383",
"repo": "robiso/wondercms",
"url": "https://github.com/robiso/wondercms/issues/39"
}
|
gharchive/issue
|
Persistant update notification | Summernote button size
Hello,
I found several issues which I could not resolve.
The first issue is the notification which says "New WonderCMS update available." When I click on "Update WonderCMS" the success message shoows up but the notification itself won't disappear.
The second one has to deal with the summernote plugin. The buttons of the editor don't have the same height. In the developer console I could not find any css class which overrides the default values for the buttons.
Hopefully someone can help me :)
Regards
AndiLeni
Hello AndiLeni.
Please answer a couple of question to make the assistance faster.
Have you ever successfully used the update before?
When you open the settings panel, what version of WonderCMS does it show you?
Are you coming from a clean install of WonderCMS?
Do you know what version of PHP you're running and on what server (Apache, NGINX, IIS)?
I'm sure more question will arise such as (if you're coming from an older version):
Did you update the summernote plugin manually, or is this a clean install?
Were the summernote buttons fine before the update (if you are in fact coming from an update)?
No, it is the first time I tried WonderCMS
2.3.0
see 1.
I installed Wondercraft on my local webserver using MAMP. Apache, PHP 7.0.9
Everything was installed freshly.
Somebody had a similar issue before on a local installation, because of a missing certificate on the local machine:
https://github.com/robiso/wondercms/issues/33
@halojoy solved this with
I downloaded a certificate from the cURL website.
cacert.pem
So now it works alright.
You could turn of the cerificate check for cURL, but it's bad practice and considered unsafe as it enables MITM (Man In The Middle) attacks, so please use the above solution to solve your problem.
Just to add, this can be fixed on your local machine by install the certificate as posted above. So it would work with your MAMP installation without any problems, thus reducing the need to test it on a "real" host.
I skipped the prozesss to install a certificate and simply added the line
curl_setopt($ch, CURLOPT_SSL_VERIFYPEER, 0);
to the _getExternalFile function.
The update notification is now working as expected but the Summernote plugin still does not work.
UPDATE:
For further reference and if someone might has this problem too here is the solution.
In my php.ini file display_errors was set to on which forces the browser to render a warning, notice, error, etc. before the <!DOCTYPE html> tag. As seen in your link to stackoverflow the problem was caused by a wrong doctype.
Example:
<br />
<b>Notice</b>: Undefined index: token in <b>C:\MAMP\htdocs\wCMS\index.php</b> on line <b>408</b><br />
<br />
<b>Notice</b>: Undefined index: token in <b>C:\MAMP\htdocs\wCMS\index.php</b> on line <b>365</b><br />
<!DOCTYPE html>
As I stated before: turning of the certificate check (with curl_setopt($ch, CURLOPT_SSL_VERIFYPEER, 0);) instead of actually installing the certificate is considered bad practice, unsafe - it enables MITM (Man In The Middle) attacks.
Regarding the Summernote plugin, we have nothing to do with the actual styling of the plugin except for turning off the background to transparent. I tried Googling for your problem and the only thing I came up with was: https://stackoverflow.com/questions/24594828/summernote-buttons-varying-height.
You can open a seperate issue about this as this is not a part of WonderCMS. Summernote is a seperate project, and a WonderCMS plugin was made by Prakai Nadee, I only patched up the vulnerabilities and turned off the Air mode. When you open a new issue in the plugins section, make sure to post a screenshot or a demo example on your server.
The cause for those errors is because we've missed to check if the REQUEST[TOKEN] exists before putting into an if statement. Closing this issue and issuing a patch soon.
Note: the patch won't fix the cURL certificate error you're having. Updating WonderCMS to a newer version overwrites any changes you've made to index.php. If you need to make changes, you can do so with the functions.php file or a plugin, check our wiki (https://github.com/robiso/wondercms/wiki) for more info.
// Edit, here's the fixed index.php, you should not be getting any more errors. Please confirm.
index.php.zip
Fatal error: Cannot use isset() on the result of an expression (you can use "null !== expression" instead) in C:\MAMP\htdocs\wCMSold\index.php on line 406
Found the type, please try again.
index.php.zip
If everything goes well, let me know. Can you please post your first name / last name and link to website so we can link to it on https://wondercms.com/whatsnew
I guess you posted the wrong file. The first one you gave me and the second one are the same :)
My name is Andreas Lenhardt.
Here comes the right one:
index.php.zip
We'll list you just by name, any website links, or maybe Twitter?
Parse error: syntax error, unexpected ')' in C:\MAMP\htdocs\wCMSold\index.php on line 406
A ) too much. If I remove it it's fine.
You can link to: https://twitter.com/AndiLenhardt
Sorry for the hassle, one last time attaching the corrected ZIP.
index.php.zip
If you don't mind running it for a test one more time. This will be dispatched as version 2.3.1 within the next 24 hours.
Thanks @AndiLeni.
Patch for the two "is sets" deployed with 2.3.1.
@AndiLeni, we have thanked/linked to you on the following locations:
https://wondercms.com/whatsnew
https://github.com/robiso/wondercms/releases/tag/2.3.1
https://wondercms.com/forum/viewtopic.php?f=8&p=1833#p1833
We have additionally added you to our special contributors list
https://www.wondercms.com/special-contributors
Also on the WonderCMS download page
https://www.wondercms.com/latest
Thank you. :)
There are still some warnings which didn't occur before. Are you sure you didn't mix up some files accidentally?
Notice: Undefined index: token in C:\MAMP\htdocs\wCMS\index.php on line 407
Notice: Undefined index: token in C:\MAMP\htdocs\wCMS\index.php on line 364
Both of them can be fixed using isset()
Are you sure this happens with the latest 2.3.1 version? Here's the official index.php:
https://github.com/robiso/wondercms/blob/master/index.php
Both of those lines you're reporting have an isset check in the official repository.
Well, obviously this time I messed stuff up.
My fault, sorry!
We're only human. :)
|
2025-04-01T04:35:22.170058
| 2015-10-30T18:52:58
|
114332405
|
{
"authors": [
"emanchado"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10384",
"repo": "robohydra/robohydra",
"url": "https://github.com/robohydra/robohydra/issues/17"
}
|
gharchive/issue
|
Implement plugin configuration defaults
Many plugins might share configuration keys, and it's tedious to repeat their values. Thus, there should be some kind of plugin configuration defaults in the configuration file.
See https://groups.google.com/forum/?hl=es&fromgroups#!topic/robohydra/Bsf8Wo0efvE for background.
This should do the trick. I'm not sure when I'll make a new release (I'd like to try to get the proxy support in there, too), but for now you can use master.
|
2025-04-01T04:35:22.177225
| 2018-05-01T05:22:48
|
319119509
|
{
"authors": [
"cmeury",
"gtaylor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10385",
"repo": "roboll/helmfile",
"url": "https://github.com/roboll/helmfile/issues/131"
}
|
gharchive/issue
|
KubeCon Europe 2018
I am not sure whether an issue is the right place to discuss this, but I wouldn't know of any other communication channel (Chat, ML). Is any maintainer, contributor or user at KubeCon Europe in Copenhagen this week? If yes, care to have a coffee? Drop me a mail.
This issue might be changed to: establish and/or communicate discussion channel for helmfile devs and users.
Could be worth seeing if we could get a helmfile channel created on the Kubernetes Slack org.
Yeah, I asked the admins already, will check once in the hotel :)
On Tue, May 1, 2018, 17:48 Greg Taylor<EMAIL_ADDRESS>wrote:
Could be worth seeing if we could get a helmfile channel created on the
Kubernetes Slack org.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/roboll/helmfile/issues/131#issuecomment-385705351,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABQgSmLxYyp7Z4q8zXj4qP-P4u7-u0x9ks5tuIPfgaJpZM4TtlsX
.
Didn't have a response from the admins - maybe try again in the future, closing for now.
|
2025-04-01T04:35:22.182029
| 2022-11-07T08:01:30
|
1437924411
|
{
"authors": [
"Danfoa",
"stephane-caron"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10386",
"repo": "robot-descriptions/robot_descriptions.py",
"url": "https://github.com/robot-descriptions/robot_descriptions.py/issues/12"
}
|
gharchive/issue
|
Missing dependency package yourdfpy
Good Day,
This issue is just to call attention to a missing dependency during package build and install. Specifically, yourdfpy package is missing as a description and throws an error when using the CLI.
I will submit a pull request with the change soon if it is not easily fixed.
Thanks for your feedback :slightly_smiling_face: If everything is implemented correctly yourdfpy should be an optional dependency (e.g. pip install robot_descriptions[cli] installs it):
There shouldn't be any yourdfpy-related import error when doing import robot_descriptions
If an example or sub-module uses it, it should raise a user-friendly instruction such as "Importing this module requires yourdfpy, you can install it by..."
I see that point 2 was lacking, so I started a PR for it: https://github.com/robot-descriptions/robot_descriptions.py/pull/13 @Danfoa feel free to review and let me know if that would fix it for you.
|
2025-04-01T04:35:22.183017
| 2017-12-13T20:09:44
|
281879695
|
{
"authors": [
"jmauss"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10387",
"repo": "robotattackorg/robot-detect",
"url": "https://github.com/robotattackorg/robot-detect/issues/9"
}
|
gharchive/issue
|
Scanning subnets
I would love to have the ability to scan entire subnets with this tool, unless I am missing something on how this is done? (Using .* and 0/24 both do not work)
I have closed this as I just used a bash script to fix the problem. Reopen if wanting to implement.
|
2025-04-01T04:35:22.187995
| 2016-08-24T00:57:58
|
172841721
|
{
"authors": [
"arunku2x",
"pekkaklarck"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10388",
"repo": "robotframework/robotframework",
"url": "https://github.com/robotframework/robotframework/issues/2429"
}
|
gharchive/issue
|
Values of Robot Framework Automatic Variables (${TEST STATUS} and ${SUITE STATUS}) not correct in certain scenarios.
Hi,
I have found that the values of Robot Framework Automatic Variables (${TEST STATUS} and ${SUITE STATUS}) not correct in the following scenarios.
Added Test Timeout parameter under Test Suite settings so that it becomes applicable for all test cases under the test suite. Observed that in case a test case is timed out, its status ${TEST STATUS} still return PASS in Test Case Teardown section. This happens if a keyword in Test Case Teardown is timed out.
There was a failure in Test Suite Teardown section but its status ${SUITE STATUS} returned PASS.
I want to perform some actions in Test Case and Test Suite teardown sections on the basis of test case and test suite status but due to inconsistent status of these variable, not able to do so.
Could you please look into these issues.
This is a variation of #2078. Test timeout occurring during the teardown is a bit more problematic than a normal failure, though, becuase test timeouts are, by design, disabled during the teardown to allow teardowns to run fully. Robot only marks the test failed after the test ended if it in the end took too much time. I'll update #2078 accordingly and close this issue.
Notice that the required changes to actually fix this are pretty large. They cannot be done in RF 3.0.1 release but need to wait for a major version. Unfortunately I cannot give any promises when RF 3.1 would be release nor would these fixes make into that release.
Hi Pekka,
As an alternative, I switched to listeners. I have added listener for end_test and end_suite and taking decision on the basis of attrs['status'] field inside these listener functions. Here as well, I have observed one weird issue. I had a test suite containing 7 test cases. Now all the test cases inside this test suite failed and I got correct test case status inside "end_test" listener. The issue which I have observed is that test suite status inside "end_suite" listener shows up as PASS. Wouldn't it be FAIL even if 1 test case inside the test suite fails.
Regards,
Arun Kumar Saini
Suite status is PASS if all critical tests pass. Most likely cause for your issue is that tests are non-critical. If that's not the case, please submit a separate issue with instructions how to reproduce the problem. An executable example would be best.
|
2025-04-01T04:35:22.201267
| 2017-03-24T19:54:59
|
216889288
|
{
"authors": [
"m3d",
"mdspk",
"zwn"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10389",
"repo": "robotika/simple-sfm",
"url": "https://github.com/robotika/simple-sfm/pull/1"
}
|
gharchive/pull-request
|
Calculate reprojection error.
This is the "simplest thing that could possibly work" regarding reprojection error. Calculates gradient with respect to 3D points positions in preparation for bundle adjustment.
Open question is about the best representation of orientation in 3D wrt optimization by gradient descent. Quaternions seem somewhat straight forward. Euler angles could be also used but one is never quite sure about which rotation is which.
Could you please change README.md to direct links for installation? (I suppose that requirements.txt would not help in this case). Thanks
https://conda.io/docs/install/quick.html
d:\>conda install autograd
Fetching package metadata ...........
PackageNotFoundError: Package not found: '' Package missing in current win-64 channels:
- autograd
You can search for packages on anaconda.org with
anaconda search -t conda autograd
You may need to install the anaconda-client command line client with
conda install anaconda-client
Done. In the future I intend to add travis and appveyor unit tests, so that will always contain up-to-date install instructions.
What I am most interested in the review is the readability of the code. I'd like the code to be easy to understand with minimum comments.
Thanks, now it works :). I would already split it into two files (bundle.py and test_bundle.py) and somehow I am used to 2 empty lines from PEP8 (sorry for stupid comments). Maybe test_project() and test_rotate() into unittest and test_reprojection_error() keep with prints in the bundle.py
Open question is about the best representation of orientation in 3D wrt optimization by gradient descent. Quaternions seem somewhat straight forward. Euler angles could be also used but one is never quite sure about which rotation is which.
Any opinions about this?
I would start with Quaternions
? delete branch?
|
2025-04-01T04:35:22.206145
| 2018-01-15T12:17:34
|
288575475
|
{
"authors": [
"jgvictores",
"nigno17"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10390",
"repo": "robotology/QA",
"url": "https://github.com/robotology/QA/issues/270"
}
|
gharchive/issue
|
How to change control mode in python
Hello everyone,
How can I change back the joints control mode in position after using the cartesian control in python?
I need to first control the robot in position mode and then use the cartesian controller to control in task space. This process needs to be repeated in a loop.
Thanks in advance,
Nino
Hi @nigno17 @pattacini !!
Using the old IControlMode interface, you should still be able to use:
props = yarp.Property()
props.put("device","remote_controlboard")
props.put("local","/client/right_arm")
props.put("remote","/icubSim/right_arm")
armDriver = yarp.PolyDriver(props)
iPos = armDriver.viewIPositionControl()
jnts=iPos.getAxes()
...
...
iMode = armDriver.viewIControlMode()
for j in range(1,jnts):
iMode.setPositionMode(j)
Bindings for IControlMode2 are still not in master, but in devel as of https://github.com/robotology/yarp/commit/554b7f64418a1949a2cce053d099457a11899807. I'll brew up an example if I have time and remember. :-)
Thank a lot.
The old IControlMode interface works perfectly.
|
2025-04-01T04:35:22.217998
| 2021-11-09T10:03:47
|
1048411930
|
{
"authors": [
"GiulioRomualdi",
"YuRan-W",
"tmacattank"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10391",
"repo": "robotology/osqp-eigen",
"url": "https://github.com/robotology/osqp-eigen/issues/107"
}
|
gharchive/issue
|
Why the solver returns "the problem is non-convex" or " solve successfull"
I am now using quadratic programming to solve a trajectory planning problem, and the test scenario is to overtake a vehicle. I try OSQP diresctly and OSQP eigen. OSQP works well, while when i use osqp eigen, sometimes it works well, sometimes it suddenly return "the problem is non-convex" or "solve unccessful". I check the error code from OSQP, and only when the Hessian matrix is not positive definite, can the problem becomes non-convex. I have follow the OSQP eigen MPC example grammar rules to use this tool, is there any suggestions on the randomly occured crash?
Hi @tmacattank. Could you please give us a minimum example that we can use to better debug the problem?
I am using an online simulation platform based on ROS to develop the algorithm, I will try to find an easy way so that you can run my code
Hello, I also encountered this problem, did you solve it?
|
2025-04-01T04:35:22.306725
| 2023-10-17T04:25:29
|
1946514913
|
{
"authors": [
"CLAassistant",
"michMartineau"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10392",
"repo": "robusta-dev/robusta",
"url": "https://github.com/robusta-dev/robusta/pull/1129"
}
|
gharchive/pull-request
|
Add custom labels on pods
We would like to add custom labels on forwarder and runner pods. We will use them in order to show which team is responsible for this pods.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T04:35:22.335441
| 2018-04-19T08:39:26
|
315779087
|
{
"authors": [
"coveralls",
"fanc999",
"rockdaboot"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10393",
"repo": "rockdaboot/libpsl",
"url": "https://github.com/rockdaboot/libpsl/pull/98"
}
|
gharchive/pull-request
|
Enable autotools builds to build Windows binaries
Hi,
Apparently I was working on my branch at the same time when it was being merged, sorry! The other changes that was done in libpsl.h.in shouldn't affect things at this point.
This updates the autotools builds (which I am not too good at) for enabling builds with MinGW. I will also work on Visual Studio NMake Makefiles shortly afterwards.
With blessings, thank you!
Coverage remained the same at 93.648% when pulling 917be0ffa82090006fe5f8dec575f60827ea2fe5 on fanc999:master into 44256b1a3aa1309a2fe60af942a8c7ace17dbcea on rockdaboot:master.
The commit is too windows-centric. So here are some questions:
We link to -lws2_32, which is necessary for the networking functions on Windows.
Why don't you use AC_SEARCH_LIBS(inet_ntop, ws2_32) or something like this ?
We default to using libicu on Windows...
That's what we have configure flags for. Please describe the problem you have using them.
...not have fmemopen()
So please add fmemopen to AC_CHECK_FUNCS in configure.ac, and modify the fuzzer code to exit(77) if HAVE_FMEMOPEN is not defined. That displays a 'SKIP' when running the test suite.
Hi Tim,
I updated the check for inet_pton(), which was what was being used, as you suggested.
For the libicu part, I am leaving that intact for now, as basically the issue is where we set the default runtime and builtin as libidn2. The thing is code that uses libidn2 (and libidn as well) make use of items in langinfo.h, which is a header that is not available on native Windows builds (be it MinGW or Visual Studio). This means that --enable-runtime=libicu and --enable-builtin=libicu is required for the code to build on Windows at the current state (or we disable the runtime annd builtins). Another way is to update README.md to tell people that currently it is required to pass in --enable-runtime=libicu and --enable-builtin=libicu (or use --disable-runtime and --disable-builtin) to build libpsl on Windows.
Please let me know which is your preferred way to this issue.
Also, I am doing another commit here for the fuzz tests part.
With blessings, thank you!
make use of items in langinfo.h
Right, there also should be a guard around the cal to nl_langinfo() and in the else case call the appropriate Windows function. Do you have it at hand ? I have it somewhere, must search...
Another way is to update README.md to tell people that currently it is required to pass in --enable-runtime=libicu and --enable-builtin=libicu (or use --disable-runtime and --disable-builtin) to build libpsl on Windows.
That sounds good for now. On the other hand, if libidn2 development files aren't installed, ./configure should fallback to libidn and then to libicu automatically. I thought it was working that way, but will have a look. Maybe it works if you fix the nl_langinfo() thing ?
Hi Tim,
I will look into nl_langinfo() with the appropriate Windows replacement later.
For the default runtime/builtin, unfortunately configure errors out, saying something like requested libidn2 but could not find it, so no automatic fallback is being done there.
I think for the Visual Studio builds, getting libidn2/libidn and libunistring is going to be tedious, due to their use of gnulib (big autotools-using stuff there)... So, it is likely that for the Visual Studio builds, I will focus on using ICU and the case where the runtime/builtin is disabled.
With blessings, thank you!
OK, just let me know if you are interested in a MinGW build script for libunistring, libidn2 and libidn. I use one for wget2 CI builds. Might help before you start from scratch.
@fanc999 Pushed two of your commits and added another one implementing the fallback thing I mentioned. Indeed, it wasn't implemented (must have been a dream ;-)). Thanks for your work.
Hi,
I posted a new patch to enable DLL builds on Windows, as we need the -no-undefined linker flag on Windows for building DLLs, and enabled linking to libintl and libiconv when we are using the libidn/libidn2 code, otherwise the build will fail to link since APIs from these two libraries are being used (libintl and libiconv are things that are not in the MinGW/Windows CRT library).
Let me know if these two changes could be done in a better way, since I am by no means good in autotools.
With blessings, thank you!
Can you remove the OS specific code and try this instead, please:
diff --git a/configure.ac b/configure.ac
index f0e627b..501d83e 100644
--- a/configure.ac
+++ b/configure.ac
@@ -10,7 +10,7 @@ AC_CONFIG_HEADERS([config.h])
AC_PROG_CC
m4_ifdef([AM_PROG_AR], [AM_PROG_AR])
#LT_INIT([disable-static])
-LT_INIT
+LT_INIT([win32-dll])
AC_CONFIG_MACRO_DIR([m4])
m4_ifdef([AM_SILENT_RULES], [AM_SILENT_RULES([yes])])
diff --git a/src/Makefile.am b/src/Makefile.am
index e18d754..c645b70 100644
--- a/src/Makefile.am
+++ b/src/Makefile.am
@@ -11,7 +11,7 @@ libpsl_la_CPPFLAGS = -I$(top_srcdir)/include -I$(top_builddir)/include -DPSL_DIS
$(CFLAG_VISIBILITY) -DBUILDING_PSL
# include ABI version information
-libpsl_la_LDFLAGS = -version-info $(LIBPSL_SO_VERSION)
+libpsl_la_LDFLAGS = -no-undefined -version-info $(LIBPSL_SO_VERSION)
if WITH_LIBICU
libpsl_la_LDFLAGS += $(LIBICU_LIBS)
endif
It's from https://www.gnu.org/software/libtool/manual/html_node/LT_005fINIT.html, scroll down to 'win32-ddl'. The code works here. If it works for you we should use it.
The second issue, libintl and libiconv, should also be done automatically by libtool. We just have to add $(LTLIBICONV) $(LTLIBINTL) @INTL_MACOSX_LIBS@ in Makefile.am at the right place(s). I'll try to find a machine/OS for testing.
Added two commits that address the above. Please check and give me feedback.
Hi Tim,
The two commits do work.
With blessings, thank you!
|
2025-04-01T04:35:22.359168
| 2023-01-29T13:23:57
|
1561287706
|
{
"authors": [
"aidenlangley",
"tversteeg"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10395",
"repo": "rockerBOO/awesome-neovim",
"url": "https://github.com/rockerBOO/awesome-neovim/pull/673"
}
|
gharchive/pull-request
|
Add ~nedia/auto-save.nvim
Checklist:
[x] The plugin is specifically built for Neovim, or if it's a colorscheme, it supports treesitter syntax.
[x] The lines end with a .. This is to conform to awesome-list linting and requirements.
[x] It's not already on the list.
[x] The title of the pull request is Add/Update/Remove `username/repo` when adding a new plugin.
[x] The description doesn't start with A Neovim plugin for... or A plugin for..., and doesn't end with ... for Neovim.
[x] The description doesn't mention that it's a Neovim plugin, it's obvious from the rest of the document.
[x] Neovim is spelled as Neovim (not nvim, NeoVim or neovim), Vim is spelled as Vim (capitalized), Lua is spelled as Lua (capitalized).
Thanks for the PR!
|
2025-04-01T04:35:22.478877
| 2023-02-01T16:33:58
|
1566433949
|
{
"authors": [
"daniel-ciocirlan",
"niklasuhrberg"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10396",
"repo": "rockthejvm/rockthejvm.github.io",
"url": "https://github.com/rockthejvm/rockthejvm.github.io/pull/25"
}
|
gharchive/pull-request
|
Article on Akka Streams to actors interoperability
Here is the article on Akka Streams and actors interoperability.
Please note that I used html for the image of the creek and specified its width. It was rendered too big without this I think.
Also, will make sure the final picture (whether it's changed or not) is guaranteed to be freely used.
When I added a Title below the fron matter , the title appeared twice so I only have the title specified by the front matter.
Thanks!
|
2025-04-01T04:35:22.495297
| 2024-09-30T20:25:49
|
2557574316
|
{
"authors": [
"Piotrk39",
"gulfaraz"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10397",
"repo": "rodekruis/IBF-system",
"url": "https://github.com/rodekruis/IBF-system/issues/1599"
}
|
gharchive/issue
|
Create issues for E2E tests in trigger scenarios
Who requested this task?
@gulfaraz
What needs to be done
[x] Identify differences between trigger and no-trigger state on the UI (start with Uganda flood)
[x] Update test plan
[x] Create follow-up items
Acceptance Criteria
[x] Follow-up items created
Relevant links
Use prioritization done for the no-trigger state for reference.
All the above test were added in "Qase" as test scenarios and separate items were created in "New" tab
|
2025-04-01T04:35:22.499490
| 2020-03-29T01:35:51
|
589695683
|
{
"authors": [
"rodrigo-brito",
"rrfaria"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10398",
"repo": "rodrigo-brito/backtrader-binance-bot",
"url": "https://github.com/rodrigo-brito/backtrader-binance-bot/issues/6"
}
|
gharchive/issue
|
how to use it in a different broker
I was taking a look on your code
and it is pretty good
but it is to binance
I need use it in iqoption as a broker
there is a community behind iqoption trying to use iqoption api
but no one know how to put it on backtrader
do you know how to create a broker on backtrader
Is there any video on internet about it ?
Hi @rrfaria,
You can create a custom class with iqoption operations like this: https://gist.github.com/rodrigo-brito/8c82020f04e946e3f0c39c7243cfe1ee. In the code, you can create an instance of this custom class in the constructor of your strategy and execute the buy or sell function instead of backtrader built-in functions.
The main problem is the data feed. You should create a class wich provides some fields: https://www.backtrader.com/docu/datafeed-develop-general/datafeed-develop-general/
thanks @rodrigo-brito
but if I use this class on strategy I'll not lose boker functionalities ?
I need mensure if it is getting profit or not ?
iqoption uses spread on forex and crypto I would like to use a strategy where use it as a important property to do buy or sell because some time when spread is too high doesn't compensate do any thing and it is hold time
I found oanda code:
https://github.com/ftomassetti/backtrader-oandav20/tree/master/btoandav20
he implements store, broker, sizer and feed
I tried to read backtrader docs but it is not so clear
|
2025-04-01T04:35:22.506651
| 2017-02-08T09:43:48
|
206143873
|
{
"authors": [
"rofrischmann",
"tiagojsalmeida"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10399",
"repo": "rofrischmann/fela",
"url": "https://github.com/rofrischmann/fela/issues/195"
}
|
gharchive/issue
|
Monolithic mode?
Version: 3.0.x
Environment: Both
Type: Feature
Description
Atm you guys only support atomic classnames.
Would it be possible to implement an monolithic mode?
Proposal
Code Example:
createRenderer({
mode: 'monolithic'
})
Not sure about that. I don't see any advantages actually. But it might be possible to just have an alternative renderer. The React Native renderer e.g. is also a totally different renderer.
@rofrischmann In our case it would be ideal to use the monotithic version on development as it's just easy to change styles.
On our implementation we made our own render method as we don't want to use sheet.insertRule for multiple reasons (specially because of third-party applications).
I would totally help. Could you point me on the right direction?
You could basically take the existing renderer and simply replace renderRule. I may give you a list of "things" which you need to do in order to achieve correct classnames etc.
Added with #200
|
2025-04-01T04:35:22.575279
| 2014-12-06T10:58:53
|
51179036
|
{
"authors": [
"bhh",
"dgilperez",
"guzart",
"ianheggie",
"jondeandres",
"rokob"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10400",
"repo": "rollbar/rollbar-gem",
"url": "https://github.com/rollbar/rollbar-gem/issues/193"
}
|
gharchive/issue
|
Configuration option to disable Rails support
i know its just lame to still have 2.3 applications but i got this "new" old application i have to take care of and first thing is i want to log errors nicely.
is there a way to disable the default rails stuff per config?
the only problem i have is that it wants to autoload railtie and active_record_extension
if i could disable it per config everything would work fine and i could just have to rescue the errors and stuff i want to and send it with the gem (this part seems to work perfectly because there is not rails integration)
if you think wtf he is crazy wanting rails 2.3 support just delete it ;)
i dont really need support for it just a method to disable rails support
would it be a problem if i try to get it into the config with a PR?
ok new approach.
what do u think of the idea of spliting the rails features.
so we have a rollbar-gem and rollbar-rails with rollbar dependecy of course. like rspec-rails and stuff.
additionally we could implement the javascript client side integration also in the rails gem
I think environment flag is probably the easiest way to go, and best practices ATM suggest your app gets the configuration from environment variables (running context).
easy definitly but i like the idea of extracting the rails part way more. and if the owners approve, i have no problem of doing a bit more work to keep everything clean.
for me it feels nicer to have a core gem with just the rollbar api in it
@bhh feel free to open a PR with your solution, we'll be looking at this at soon as possible
I have just added rollbar into a rails 2.3.18 LTS project (ruby 1.9.7, capistranio v3) - suggestions on how to improve it welcome.
I disable the require in Gemfile so I can be sure it is initialised if it has been loaded.
Gemfile:
gem 'rollbar', require: false
config/initializers/rollbar.rb:
require 'rollbar'
rollbar_enabled = (Rails.env.test? || Rails.env.development?)
Rollbar.configure do |config|
require 'rollbar'
rollbar_enabled = (Rails.env.test? || Rails.env.development?)
Rollbar.configure do |config|
# Without configuration, Rollbar is enabled in all environments.
# To disable in specific environments, set config.enabled=false.
config.enabled = rollbar_enabled
# ... etc etc ...
end
if rollbar_enabled
require 'rollbar/rails23/notifier'
class ActionController::Base
include Rollbar::Rails23::Notifier
def rescue_action_with_rollbar(exception)
notify_rollbar_from_within_controller(exception)
rescue_action_without_rollbar exception
end
alias_method_chain :rescue_action, :rollbar
end
end
lib/rollbar/rails23/notifier.rb:
module Rollbar
module Rails23
module Notifier
include Rollbar::ExceptionReporter
include Rollbar::RequestDataExtractor
def notify_rollbar_from_within_controller(exception)
env = @_request.env
request_data = extract_request_data_from_rack(env)
request_data[:route] = {
controller: params[:controller],
action: params[:action],
format: params[:format]
}
scope = {
request: request_data,
#context: { vertical: @vertical } # FIXME: Work out how to pass context
}
if current_user
scope[:person]= {:id => current_user.id, :username => current_user.screenname, :email => current_user.email}
scope[:person][:username] = current_user.screenname if !current_user.screenname.blank? && current_user.screenname != 'anonymous'
end
Rollbar.scoped(scope) do
Rollbar.error(exception)
end
rescue Exception => exception2 # report if we stuffed up reporting with details
Rollbar.error(exception)
Rollbar.error(exception2)
end
end
end
end
i'm so sorry i totally forgot about that topic. i think will look into it again upcoming week and maybe start a rollbar-rails project
Two of my clients are on https://railslts.com/ - rails 2.3 LTS (and 3.2) is worthwhile supporting.
Hey @ianheggie, have you taken a look at https://github.com/rollbar/rollbar-gem/blob/master/docs/plugins.md?
It's the architecture for the plugins system. We are wondering if that rails 2.3 support could be rewriten using the new plugins interface.
BTW, nice work so far on that.
Thanks for the input @jondeandres - I will have a look at the plugin support - I have a few further things I need to do (missing action is being caught earlier in the process and thus is not reporting controller). Also in my project I am disabling router errors unless a user is logged in - I get way to many bots submitting mangled urls.
Superb job @ianheggie! Did you eventually get a running solution? I also have a new old rails 2 app to maintain :)
Yes, it works well enough (A few little niggles, but I am happy to share the code) - and re new/old rails 2 - I have two clients on 2.3.18 LTS - and I can see them staying on it because the cost to upgrade and keep upgrading would be too much.
I'd love to see that final code if I may 😄
wondering if people are still interested in this.
had some time to kill today and removed everything rails related as i suggested. as far as i can tell everything is still working. tests + some examples i tried.
https://github.com/rubberandglue/rollbar-gem
i copied every piece of removed code and tests into another repository (i guess this one will be a bit more tricky)
i still missed out on the documentation for now. and i'm not sure if i did everything right for travis because rails tests aren't necessary anymore for this repository if the solution is accepted. there are still some "rails" strings but they should all be just string matching (not really rails related as far as i can see)
i really like the solution so far. splitting everything up really contributes to nice replaceable modules. we wont be able to just turn rails support on/off with the new gem we also could make multiple versions for rails 2,3,4,5
hope you guys like it too
I like the idea, but I'm still not sure about the benefit of doing a full split into two repos versus trying to do this via the plugin mechanism. I need to look at the code a bit more. I think the split into two repos is actually probably easier, just not sure if it is better from a maintenance perspective going forward.
as i discovered the chronlogy is seems to be weird. thats why i came up with a solution i like most
railtie loading
getting config
deciding to disable rails
how not to load rails!? everything loaded/crashed already
with the plugin mechanism and railtie you always have to load everything first then get the config and the decide to disable!? but the crash already happens during loading. why load it in the first place when i want to disable a specific part(not efficient anyway) and thats fine because i dont think it should be there anyway
imo maintaining the rollbar part seperate from the rails part means you dont have to handle multiple rails versions in the rollbar plugin. no more RAILS_VERSION compare stuff
and its even easier to handle different rails versions because you even can remove all the rails version if/switch statements and just use different rollbar-rails versions.
|
2025-04-01T04:35:22.617762
| 2017-01-03T17:47:35
|
198529245
|
{
"authors": [
"mrship",
"solnic"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10401",
"repo": "rom-rb/rom-sql",
"url": "https://github.com/rom-rb/rom-sql/issues/123"
}
|
gharchive/issue
|
Composing relations through many-to-many join
We have a setup where we have a many-to-many relationship via a join table, e.g.
module Relations
class Questions < ROM::Relation[:sql]
dataset :questions
schema(infer: true) do
associations do
has_many :charts, through: :questions_charts
end
end
end
end
module Relations
class QuestionsCharts < ROM::Relation[:sql]
dataset :questions_charts
schema(infer: true) do
associations do
belongs_to :questions
belongs_to :charts
end
end
end
end
I can query this in my questions repository using a combine_children call to pull out all the charts and that works well, e.g.
questions
.combine_children(
many: {
charts: charts,
}
)
However, what I'd like to do is add an order attribute to the chart which is actually a column on the questions_tables join table, not on the chart table, alongside the attributes that come from the chart table itself, e.g.
[{ id: 1, name: 'my chart', order: 1 }, { id: 2, name: 'my other chart', order: 2 }]
...where id and name are attributes of the chart table.
At the moment the associations definition is happily pulling the charts back through that has_many through call, so I'm expecting to have to write a custom view in the questions_charts relation that does the same thing, but also includes the order attribute from the questions_tables table.
However, I can't work out how to construct this view so that the join still works as it currently does automagically with the association setup. Could you please offer some advice on how to achieve that?
Thanks in advance for your help.
In rom-sql 1.0.0 you will be able to extend an association's relation with your own query logic, but for now you can define a custom view. To make it simpler for you, you can re-use association relation:
class Charts < ROM::Relation[:sql]
schema(infer: true) do
associations do
has_many :questions_charts
has_many :questions, through: :questions_charts
end
end
view(:for_questions, %[list of attributes in resulting relation]) do |questions|
# not pretty, I know, there will be a nice API for that in 1.0.0
rel = associations[:questions].call(__registry__)
rel.order(:questions__that_column_you_want_to_use)
end
end
Combine will automatically use this view by convention, so it should just work.
It'll be much simpler in 1.0.0 though. I'm releasing first beta in the upcoming days so stay tuned :)
ps. just updated the example to be a bit nicer
Thanks for the help; I probably confused things a little using order as my example attribute, but that is actually a column on the questions_tables join table, so I want to be able to select all the columns from chart as well as the order column from the join table.
I'll see if I can figure that out from your example, but I'm still not 100% sure how to use the join_table as part of the for_combine.
@mrship the association used under the hood joins that table for you automatically, so you can just do select_append(:questions_tables__some_col_1, :questions_tables__some_col_2).
ps. this will be so much nicer in rom-sql 1.0.0 :)
Thanks for the help so far and apologies for the delay in coming back to you but I'm only just coming back around to this.
When I'm looking at the custom view that you wrote above, it appears to be selecting the question attributes, not chart attributes; e.g: if I strip this back to make Charts#for_questions just replicate the existing association (before adding in any other attributes)
view(:for_questions, %i(id type created_at updated_at snapshot insight)) do |questions|
q = for_combine(associations[:questions])
.where(questions__id: questions.pluck(:id))
require 'pry'; binding.pry
q
end
q (above) produces this relation SQL:
"SELECT `questions`.`id`, `questions`.`topic_id`, `questions`.`title`, `questions`.`short_title`, `questions`.`methodology`, `questions`.`order`, `questions`.`created_at`, `questions`.`updated_at`, `chart_id` FROM `questions` INNER JOIN `questions_charts` ON (`questions_charts`.`question_id` = `questions`.`id`) INNER JOIN `charts` ON (`charts`.`id` = `questions_charts`.`chart_id`) ORDER BY `questions`.`id`"
implemented like that, I then end up with this error:
ArgumentError:
Relations::Questions#preload arity is 3 (2 args given)
when calling the combine_children (note I also have to call the view directly as it doesn't work automatically as you implied)
questions
.combine_children(
many: {
charts: charts.for_questions,
}
)
So, sorry the news isn't better but I think I'm missing something fundamental here.
Unfortunately, the view doesn't get called automatically if I combine_children(many: charts) only if I am explicit about it with combine_children(many: questions.for_charts).
@mrship oh right I forgot that it needs keys to passed explicitly, damn :) see docs for #combine where it's explained
Ok, you mean something like, this?
questions.combine(many: { charts: [questions.for_charts, id: :chart_id] })
unfortunately, that gives a similar error but on charts.
ArgumentError:
Relations::Charts#preload arity is 3 (2 args given)
Apologies if I'm being dense here; I feel like I'm so close.
I thought you wanted questions.combine(many: { charts: [charts.for_questions, id: question_id])?
I've created a gist - https://gist.github.com/mrship/3f59b5ab6c40e7cacc0a03eba92742d0 - that shows the problem and hopefully makes it a little clearer what is happening.
The ultimate aim is to have a Question with many Charts, and the Charts have their order from the join table.
If you run the gist, you'll get the same arity issue that I reported above.
Thanks again for the extensive help.
That's very helpful. I'll check it out and get back to you soon
Cheers
Piotr
On 13 Jan 2017, 7:59 PM +0100, Andy Shipman<EMAIL_ADDRESS>wrote:
I've created a gist - https://gist.github.com/mrship/3f59b5ab6c40e7cacc0a03eba92742d0 - that shows the problem and hopefully makes it a little clearer what is happening.
The ultimate aim is to have a Question with many Charts, and the Charts have their order from the join table.
If you run the gist, you'll get the same arity issue that I reported above.
Thanks again for the extensive help.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub (https://github.com/rom-rb/rom-sql/issues/123#issuecomment-272518667), or mute the thread (https://github.com/notifications/unsubscribe-auth/AAAEKtKYsI_P2qBepmyX8AW1vqjXEqgYks5rR8mcgaJpZM4LZ2sT).
@mrship pheeew, there we go: https://gist.github.com/46584d53fb1f6901b3d29381cec1ccd1
Wow! That works a treat. There's no way I'd ever have got to that point without your help though, so (a) thanks and (b) I'm looking forward to seeing how that becomes easier in v1.0.0!
The good news is that this approach is going to get used a lot throughout my current project, and simplify a lot of features, so thanks again for the help.
@mrship cool! I'll let you know once there's a nicer way, in 1.0.0.beta there's already support for defining associations extended with custom views (ie has_many :questions, view: :ordered) and the only missing part is having ability to extend views's schema with attributes from other relations (this order col in your case).
@mrship hey man, I added a couple of features to make this easier for you. Here's an updated gist: https://gist.github.com/solnic/46584d53fb1f6901b3d29381cec1ccd1
This will be available in rom 3.0.0 and rom-sql 1.0.0
That looks really nice. Thanks for following up. I hope to try 1.0.0 when it comes out of beta.
|
2025-04-01T04:35:22.626921
| 2023-04-01T19:16:58
|
1650608775
|
{
"authors": [
"Abashinos",
"mike-pisman",
"roman-right"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10402",
"repo": "roman-right/beanie",
"url": "https://github.com/roman-right/beanie/issues/522"
}
|
gharchive/issue
|
Design the back-ref interfaces
Interfaces:
back refs attributes and ways to map
fetching
Attributes Design:
from beanie import Document, Link, BackLink
class Door(Document):
height: int
width: int
house: BackLink["House"] # Back-ref
class House(Document):
door: Link[Door] # Original reference
Seems useful for model ops decoupling
Hi, thank you for this feature! Could you please clarify the use of it?
Am I correct to understand that the BackLink will be populated automatically after creating the link? I am trying to create a BackLink, but there is no data in the mongo document after creation(besides name, etc). I assume, there should be some metadata like id and document type, right? Here is a simplified example of what I am trying to do:
class Resource(Document):
id: ResourceID = Field(default_factory=ResourceID, alias="_id")
resource_type = ""
name: str = Field(title="Name", description="Name of the resource", min_length=3, max_length=16)
groups: list[Link["Group"]] = []
class Workspace(Resource):
resource_type = "workspace"
class Group(Resource):
resource_type = "group"
# workspace: Link[Workspace] # I was using Link before, but decided to try BackLink
workspace: BackLink[Workspace] = Field(original_field="groups")
# Assuming I already created a workspace
def create_group(workspace: Workspace):
new_group = Group(name=input_data.name)
workspace.groups.append(new_group)
await Group.insert(new_group, link_rule=WriteRules.WRITE)
await Workspace.save(workspace, link_rule=WriteRules.WRITE)
Hi @mike-pisman ,
No, it doesn't store any additional data in the Mongo document. I use the lookup aggregation pipeline to find documents in the collection where the links were created and check if the original_field there has the id or ids of the documents in the collection with documents that have backlinks. So everything is happening on the db layer but without additional data to be stored there.
The feature was implemented
|
2025-04-01T04:35:22.643754
| 2015-04-10T19:58:58
|
67670283
|
{
"authors": [
"andrewcstewart",
"romansanchez"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10403",
"repo": "romansanchez/Calaca",
"url": "https://github.com/romansanchez/Calaca/issues/13"
}
|
gharchive/issue
|
Unexpected CSS token: }
I keep getting the following in the debugger. Dunno it's a big deal but thought I'd report.
Unexpected CSS token: }
calaca.css:115
What browser/os are you using?
On Friday, April 10, 2015, Andrew Stewart<EMAIL_ADDRESS>wrote:
I keep getting the following in the debugger. Dunno it's a big deal but
thought I'd report.
Unexpected CSS token: }
calaca.css:115
—
Reply to this email directly or view it on GitHub
https://github.com/romansanchez/Calaca/issues/13.
Safari 8.0.4
If there’s anything in my session info that’d be helpful let me know.
On Fri, Apr 10, 2015 at 1:22 PM, Roman Sanchez<EMAIL_ADDRESS>wrote:
What browser/os are you using?
On Friday, April 10, 2015, Andrew Stewart<EMAIL_ADDRESS>wrote:
I keep getting the following in the debugger. Dunno it's a big deal but
thought I'd report.
Unexpected CSS token: }
calaca.css:115
—
Reply to this email directly or view it on GitHub
https://github.com/romansanchez/Calaca/issues/13.
Reply to this email directly or view it on GitHub:
https://github.com/romansanchez/Calaca/issues/13#issuecomment-91675903
Yeah line 114 had some extra chars. Fixed it. Thanks.
|
2025-04-01T04:35:22.658447
| 2020-06-16T18:49:38
|
639897708
|
{
"authors": [
"electrodude",
"emagnier",
"tukusejssirs"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10404",
"repo": "romcal/romcal",
"url": "https://github.com/romcal/romcal/issues/193"
}
|
gharchive/issue
|
Make GitHub description say e.g. "Roman Rite of the Catholic Church" instead of "Catholic Rite"
The current GitHub project description is:
Generates the General Roman Calendar used in the Catholic Rite.
There is no such thing as the "Catholic Rite". All of the rites used within the Catholic Church -- the Roman Rite, a few variations on Latin Rite such as the Ambrosian and Dominican Rites, and all of eastern rites -- are Catholic rites, and many of them have their own calendars.
It should instead say something like "the Roman Rite of the Catholic Church".
What do you think @pejulian ? I don't have the right to edit this headline.
Thank you! It's now updated to:
Generates the liturgical calendar of the Catholic Church used by the Roman Rite.
And let me know if the wording could be improved, English is not my primary language 😉
@emagnier, how about to get inspired by Wiki: GRC article short description?
Liturgical calendar of Saints' Days for the ordinary form of the Roman Rite (Roman Catholic Church)
Like: Generates liturgical calendars of the Roman Rite of the Roman Catholic Church.
I think calendars should plural, as we generate different cals by country or language.
I personally wouldn’t include ordinary form. We also might place Roman Catholic Church in brackets or even remove it altogether, but for people less learned in the liturgy, it might be useful to keep it.
And in English, there is no such thing as implicit/silent subject, therefore it sounds strange and is grammatically incorrect without romcal at the the beginning. I know in code we do it that way a llt, but in user-facing docs I don’t think it is a valid reason to remove the subject. That said I let you decide on this. :wink:
What do you think?
Generates liturgical calendars of the Roman Rite of the Roman Catholic Church.
This is a good proposition. I've updated the repo description.
We still have to update the headline in the main readme.md
We still have to update the headline in the main readme.md
Should I create a PR for that?
Yes please 😄
|
2025-04-01T04:35:22.667835
| 2020-07-27T23:50:14
|
666654109
|
{
"authors": [
"ematipico",
"ljharb",
"mbrookes",
"milesj",
"sebmck"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10405",
"repo": "romefrontend/rome",
"url": "https://github.com/romefrontend/rome/issues/893"
}
|
gharchive/issue
|
Default project config location
I'm writing the docs and right now we recommend an .rjson file for the project config. I'm not totally sold on this and want us to consider the following two options, if anything so we can rule them out:
Defaulting to package.json
Force rome.r?json files to be inside of a .config directory. Many tools support this although I would propose forcing it if we decide to support it. We can pave the cowpath for the JS ecosystem.
Main motivation for this is that the root of projects has become extremely cluttered for most projects. Scanning project structure is difficult since you need to mentally filter out all the dot files.
Just chiming in with my support for placing config in .config. I did this recently for a personal project of mine (https://www.npmjs.com/package/@boost/config) and I believe it to be a good standard for JS projects going forward. Will help to reduce project root clutter.
As long as the config isn't directory-specific (in which case, a config file directly in the directory makes the most sense), that would work fine. However, it might be nice to have config able to live in two places: either directly in package.json, or, in a path that might default to .config/rome.rjson. That way users get to choose for themselves where they want it to live.
Right now the possible locations are:
rome field in package.json
rome.json and rome.rjson in the project base
I'm proposing it be changed to:
rome field in package.json
.config/rome.json and .config/rome.rjson in the project base
We automatically discover the project config so it has to be in one of these hardcoded locations. Although with extends in package.json you could put the config pretty much anywhere:
package.json
{
"rome": {
"extends": "some-other-page/config.json"
}
}
I like the idea of having .config/rome.json/.config/rome.rjson; the name of the folder tells what's that about.
I would discourage the use of the package.json because some projects would already have a big file and this would increase the size and lower the readability.
🎉
You heard it here first I even tagged @babeljs! 😉
(I wish I could actually take credit for this being added to Rome, but glad to see it happen. I hope it paves the way for more tools to follow suit!)
|
2025-04-01T04:35:22.702554
| 2020-05-07T08:23:24
|
613868657
|
{
"authors": [
"MaxDeg",
"ronaldschlenker"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10406",
"repo": "ronaldschlenker/FsHttp",
"url": "https://github.com/ronaldschlenker/FsHttp/pull/44"
}
|
gharchive/pull-request
|
Add proxy setting
This PR is related to #4
I propose a way to support proxy (with and without credentials) for request.
Thanks again for the PR! I hope to finalize this and some other things by the end of the week...
|
2025-04-01T04:35:22.758163
| 2022-11-01T22:56:39
|
1432210354
|
{
"authors": [
"roosterkid"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10407",
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/10655"
}
|
gharchive/issue
|
⚠️ TROJAN 🇺🇸 United States USO 1 has degraded performance
In c2fffbf, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) experienced degraded performance:
HTTP code: 200
Response time: 17372 ms
Resolved: TROJAN 🇺🇸 United States USO 1 performance has improved in c936cd6.
|
2025-04-01T04:35:22.829635
| 2016-07-14T16:17:27
|
165601215
|
{
"authors": [
"aviadhahami",
"nemosdo",
"redflasher",
"walidvb",
"zjjt"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10418",
"repo": "root-two/react-native-drawer",
"url": "https://github.com/root-two/react-native-drawer/issues/191"
}
|
gharchive/issue
|
Drawer not closing properly
Hey, i'm getting issues with the drawer:
with static positioning: the drawer first is displayed on top of the other views, and when closed, the content overflows:
with overlay: initially the drawer is displayed correctly, but once open, it won't close to the end.
my code is as follows:
<Drawer
type="static"
content={<CategoriesList {...this.props} />}
openDrawerOffset={160}
closedDrawerOffset={0}
onOpen={() => this.props.toggleFilters(true)}
onClose={() => this.props.toggleFilters(false)}
styles={{
drawer:{
backgroundColor: 'white',
marginTop: 60,
overflow: 'hidden'
}
}}
open={this.props.ui.filters.drawerOpen}
tweenHandler={Drawer.tweenPresets.parallax}
tapToClose={true}
>
<MenuBar {...this.props} style={{paddingTop: 20}} />
<PartnerList partners={this.state.partners} style={{
flex: 1,paddingLeft: 15, paddingRight: 15, marginBottom: sub.isValid() ? 0 : 60,
}}/>
<GetPass float/>
</Drawer>
Versions:
"react-native": "^0.29.0"
"react-native-drawer": "^2.2.6"
Link to the code
If it makes a difference, this is wrapped within a Router of react-native-router-flux
Any help/hints greatly appreciated, and thank you for the hard work!
:+1: for the overlay option
About the first case(the static one), it looks like I had to add a background color to the main component, which fixed it. Overlay is still not closing properly, though.
Hi there, i also have the same issue with overlay...It isnt closing properly any updates on that topic might help us out a great deal.
thanks and cheers
have the same issue with overlay
it works fine for me when i remove the line 'tweenHandler={Drawer.tweenPresets.parallax}' ,but i still didn't know why @walidvb @zjjt
yes, when ''tweenHandler={Drawer.tweenPresets.parallax}' was removed drawer work good.
Cause of problem may be wrong parameter for tween (forget about 'drawer.width' or something else)
i didnt check it out recently as i use the drawer component for android,if you think that is the tween handler which creates the issue ill check it out again and get back to you @nemosdo nemosdo...
|
2025-04-01T04:35:22.847384
| 2020-11-16T14:37:41
|
743888194
|
{
"authors": [
"Log1x",
"TangRufus",
"albertonieto",
"chesio",
"koraysels"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10419",
"repo": "roots/bedrock",
"url": "https://github.com/roots/bedrock/issues/559"
}
|
gharchive/issue
|
oscarotero/env v2.1.0 Class 'Env' not found
[X] I've read the guidelines for Contributing to Roots Projects
[X] This request isn't a duplicate of an existing issue
[X] I've read the docs and followed them (if applicable)
[X] This is not a personal support request that should be posted on the Roots Discourse community
Description
I have recently updated composer and got the oscarotero/env v2.1.0 package. Since then I keep getting this error:
Fatal error: Uncaught Error: Class 'Env' not found in /myproject/config/application.php:6
Steps to reproduce
Update composer to v2.1.0 on package oscarotero/env
Check your local repository
Expected behavior: [What you expect to happen]
Expected to run site normally.
Actual behavior: [What actually happens]
Throws error Fatal error: Uncaught Error: Class 'Env' not found in /myproject/config/application.php:6
Reproduces how often: [What percentage of the time does it reproduce?]
100%
Versions
oscarotero/env v2.1.0
Additional information
This is the content of the referenced file, where Env::init() corresponds to line 6:
<?php
$root_dir = dirname(__DIR__);
$webroot_dir = $root_dir . '/public';
Env::init();
// Use Dotenv to set required environment variables and load .env file in root
$dotenv = new Dotenv\Dotenv($root_dir);
if (file_exists($root_dir . '/.env')) {
$dotenv->load();
try {
$dotenv->required('DATABASE_URL')->notEmpty();
} catch (Exception $e) {
$dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']);
}
}
It seems as if the package were not loaded since the Env Class doesn't load. Ideas?
Patch Bedrock with the https://github.com/roots/bedrock/tree/master/ branch, https://github.com/roots/bedrock/pull/530/ to be specific.
Patch Bedrock with the master branch, #530 to be specific.
Sounds good! How can I do that?
Download https://patch-diff.githubusercontent.com/raw/roots/bedrock/pull/530.diff and $ git apply it.
However, depends on how much change you made and how "out of sync" your code is, it might not work.
For a small patch like #530, manually copy and paste the changes might be easier.
Props to @austinpray
Thanks, that was helpful. I have indeed manually added the missing parts. This is how the edited file looks like now:
<?php
use function Env\env;
// Directory containing all of the site's files
$root_dir = dirname(__DIR__);
$webroot_dir = $root_dir . '/public';
// Use Dotenv to set required environment variables and load .env file in root
$dotenv = Dotenv\Dotenv::createImmutable($root_dir);
if (file_exists($root_dir . '/.env')) {
$dotenv->load();
**var_dump(env('DB_USER'));exit;**
try {
$dotenv->required('DATABASE_URL')->notEmpty();
} catch (Exception $e) {
$dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']);
}
}
The problem I am facing is that if I var_dump the env variables I get empty values. Because of that, my msqli is not connecting to the database. Any idea?
@albertonieto What version of vlucas/phpdotenv are you using? I have to update the bootstrap code when I've upgraded from version 4 to 5.
Also see this issue: https://github.com/oscarotero/env/issues/8
I ended adding this to the top of the bootstrap file:
\Env\Env::$options |= \Env\Env::USE_ENV_ARRAY;
I am using the latest version "vlucas/phpdotenv": "^5.2",
Sorry, when you mention the bootstrap file, which file are you referring to? I would like to try that as well.
Sorry, I meant config/application.php - it now looks like this in my Bedrock-based project (the top part):
<?php
/**
* Expose global env() function from oscarotero/env
*/
use function Env\env;
/**
* Instruct Env\Env to read environment variables via $_ENV instead of getenv().
* The latter does not work with DotEnv 5 by default, see: https://github.com/oscarotero/env/issues/8
*/
\Env\Env::$options |= \Env\Env::USE_ENV_ARRAY;
/** @var string Directory containing all of the site's files */
$root_dir = dirname(__DIR__);
/** @var string Document Root */
$webroot_dir = $root_dir . '/web';
/**
* Use Dotenv to set required environment variables and load .env file in root
*/
$dotenv = \Dotenv\Dotenv::createImmutable($root_dir);
if (file_exists($root_dir . '/.env')) {
$dotenv->load();
$dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']);
}
...
Thanks. I have added the instruction as you referred. I am able to see the env variables when I var_dump them, however my mysqli is still not connecting:
<?php
use Roots\WPConfig\Config;
use function Env\env;
\Env\Env::$options |= \Env\Env::USE_ENV_ARRAY;
$root_dir = dirname(__DIR__);
$webroot_dir = $root_dir . '/public';
$dotenv = \Dotenv\Dotenv::createImmutable($root_dir);
if (file_exists($root_dir . '/.env')) {
$dotenv->load();
Config::define('DB_NAME', env('DB_NAME'));
Config::define('DB_USER', env('DB_USER'));
Config::define('DB_PASSWORD', env('DB_PASSWORD'));
Config::define('WP_HOME', env('WP_HOME'));
Config::define('WP_SITEURL', env('WP_SITEURL'));
var_dump(env('DB_NAME'));
var_dump(env('DB_USER'));
var_dump(env('DB_PASSWORD'));
var_dump(env('WP_HOME'));
var_dump(env('WP_SITEURL'));
$dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']);
}
I still get the:
Warning: mysqli_real_connect(): (HY000/2002): No such file or directory in /myproject/public/cms/wp-includes/wp-db.php on line 1635
No such file or directory
Error establishing a database connection
I am very confused.
@albertonieto Are you sure you have Config::apply(); at the end of the file as well? See: https://github.com/roots/bedrock/blob/master/config/application.php#L123 AFAIK, this is where the actual constants are defined. Config::define(...); calls are just part of internal setup, but do not actually define any constants.
Sorry about that, I was missing the apply at the end of the file. I have managed to actually connect via mysli now. I noticed however that our theme is using a /public/content/themes/mytheme/inc/assets.php file that reads as follows:
<?php
define('ASSETS_REVISION_FILE', dirname(dirname(ABSPATH)) . '/REVISION');
define('ASSETS_REVISION', substr(trim(@file_get_contents(ASSETS_REVISION_FILE)), 0, 12));
define('ASSETS_PATH', dirname(dirname(ABSPATH)) . DIRECTORY_SEPARATOR . env('ASSETS_COMPILE'));
define('ASSETS_DIRECTORY', str_replace(dirname(ABSPATH), '', ASSETS_PATH));
define('ASSETS_HOST', env('ASSETS_HOST') ? env('ASSETS_HOST') : '//' . $_SERVER['SERVER_NAME']);
//define('ASSETS_VERSION', ASSETS_REVISION ? ASSETS_REVISION : time('U'));
define('ASSETS_VERSION', ASSETS_REVISION ? ASSETS_REVISION : time());
function asset_path($path) {
return ASSETS_HOST . ASSETS_DIRECTORY . '/' . $path . '?v=' . ASSETS_VERSION;
}
I am getting a Call to undefined function env() for line 5. Any idea for that fix?
I am getting a Call to undefined function env() for line 5. Any idea for that fix?
You have to add use function Env\env; to this file as well - the use statements must be defined per file, see the docs.
@chesio
Nice touch about vlucas/phpdotenv v5!
Do you mind make a PR for it?
@TangRufus Sure, I'll try to get a PR ready over the coming weekend.
SO how do we fix this ? I am stuck here aswell..
SO how do we fix this ? I am stuck here aswell..
Run composer require vlucas/phpdotenv:^5.2 oscarotero/env:^2.1
Update your application.php to reflect https://github.com/roots/bedrock/blob/master/config/application.php#L31
|
2025-04-01T04:35:22.858112
| 2021-11-05T16:06:38
|
1046010598
|
{
"authors": [
"klnit",
"openoms"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10420",
"repo": "rootzoll/raspiblitz",
"url": "https://github.com/rootzoll/raspiblitz/issues/2701"
}
|
gharchive/issue
|
(UNKNOWN) [<IP_ADDRESS>] 50001 (?) : Connection refused
after upgrade to 1.7.1 I cannot connect to electrs (at first it did a new sync and I thougt that was the reason, but now after rebooot still)
STATUS ELECTRS SERVICE
configured=1
serviceInstalled=1
serviceRunning=1
syncProgress=100.00%
tipSynced=1
(UNKNOWN) [<IP_ADDRESS>] 50001 (?) : Connection refused
electrumResponding=0
initialSynced=1
localIP='<IP_ADDRESS>'
portTCP='50001'
localTCPPortActive=0
publicTCPPortAnswering=0
portSSL='50002'
localHTTPPortActive=1
publicHTTPPortAnswering=0
TorRunning=1
nginxTest=1
● electrs.service - Electrs
Loaded: loaded (/etc/systemd/system/electrs.service; enabled; vendor preset: enabled)
Active: active (running) since Fri 2021-11-05 15:45:30 GMT; 12min ago
Main PID: 6245 (electrs)
Tasks: 16 (limit: 4915)
CGroup: /system.slice/electrs.service
└─6245 /home/electrs/electrs/target/release/electrs --electrum-rpc-addr=<IP_ADDRESS>:50001
Nov 05 15:45:30 raspberrypi electrs[6245]: Config { network: Bitcoin, db_path: "/mnt/hdd/app-storage/electrs/db/bitcoin", daemon_dir: "/home/electrs/.bitcoin", daemon_auth: UserPass("raspibolt
Nov 05 15:45:30 raspberrypi electrs[6245]: [2021-11-05T15:45:30.141Z INFO electrs::metrics::metrics_impl] serving Prometheus metrics on <IP_ADDRESS>:4224
Nov 05 15:45:30 raspberrypi electrs[6245]: [2021-11-05T15:45:30.838Z INFO electrs::db] "/mnt/hdd/app-storage/electrs/db/bitcoin": 319 SST files, 39.025542469 GB, 4.027567841 Grows
Nov 05 15:45:35 raspberrypi electrs[6245]: [2021-11-05T15:45:35.493Z INFO electrs::chain] loading 708330 headers, tip=0000000000000000000b93efaa186fe13b8f96eececa5560021ea2036a42ab6c
Nov 05 15:45:39 raspberrypi electrs[6245]: [2021-11-05T15:45:39.586Z INFO electrs::chain] chain updated: tip=0000000000000000000b93efaa186fe13b8f96eececa5560021ea2036a42ab6c, height=708330
Nov 05 15:45:39 raspberrypi electrs[6245]: [2021-11-05T15:45:39.630Z INFO electrs::index] indexing 12 blocks: [708331..708342]
Nov 05 15:45:40 raspberrypi electrs[6245]: [2021-11-05T15:45:40.626Z INFO electrs::chain] chain updated: tip=0000000000000000000531ede5e077d5595576928cb1ab3cc404ff6e24a5c2db, height=708342
Nov 05 15:45:40 raspberrypi electrs[6245]: [2021-11-05T15:45:40.964Z INFO electrs::db] starting config compaction
Nov 05 15:45:40 raspberrypi electrs[6245]: [2021-11-05T15:45:40.987Z INFO electrs::db] starting headers compaction
Nov 05 15:45:45 raspberrypi electrs[6245]: [2021-11-05T15:45:45.016Z INFO electrs::db] starting txid compaction
admin@raspberrypi:~ $ sudo -u electrs lsof -i
COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME
electrs 6245 electrs 4u IPv4 1515489 0t0 TCP localhost:xtell (LISTEN)
electrs 6245 electrs 15u IPv4 1521738 0t0 TCP localhost:36664->localhost:8333 (ESTABLISHED)
electrs 6245 electrs 16u IPv4 1521738 0t0 TCP localhost:36664->localhost:8333 (ESTABLISHED)
Any idea?
starting txid compaction
Electrs is finishing reindexing after the update. Just need too wait.
|
2025-04-01T04:35:22.860350
| 2023-01-17T14:04:28
|
1536435411
|
{
"authors": [
"davebhc71",
"roovo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10421",
"repo": "roovo/obsidian-card-board",
"url": "https://github.com/roovo/obsidian-card-board/issues/131"
}
|
gharchive/issue
|
UI option to change color indicators in Dated board
Would like to change colors for Past and Current items. Orange and Red are too similar!😉
thanks btw :)
Just to follow up on this FR - Green for Today would be nice. And, I’m not asking for the User to be able to change the color, but that you consider a ‘universal’ change for Today to Green in the next update. 😉
Just realised that you should be able to use a css snippet along these lines to customise the colours:
.card-board-view .card-board-card-highlight-area.critical {
background-color: red;
}
.card-board-view .card-board-card-highlight-area.good {
background-color: green;
}
.card-board-view .card-board-card-highlight-area.important {
background-color: orange;
}
I have updated the readme with info on doing this and have just released a related feature: via snippets, you can set the color indicators based on tags too.
|
2025-04-01T04:35:23.006732
| 2020-03-10T14:03:12
|
578602622
|
{
"authors": [
"gavanderhoorn",
"jontje"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10422",
"repo": "ros-industrial/abb_librws",
"url": "https://github.com/ros-industrial/abb_librws/issues/80"
}
|
gharchive/issue
|
Python wrapper?
Being able to use RWS from a Python application would be perfect.
I've titled this issue "Python wrapper" as that could of course be one way to do this. It may however complicate things needlessly.
Searching around, I've been able to find some mentions of using requests to interface with RWS, but I haven't been able to find any proper Python package.
As RWS purports to be a REST based interface: @jontje would you happen to know whether RWS comes with an API spec (OpenAPI, Swagger, RADL, RAML, RSDL, something else) which would make it possible auto-generate (Python) wrappers or runtime code generate them?
I'm not entirely up-to-date on REST and web APIs, but there are definitely more advanced (and convenient) ways of consuming a REST interface than Python requests and composing raw URIs.
As RWS purports to be a REST based interface: @jontje would you happen to know whether RWS comes with an API spec (OpenAPI, Swagger, RADL, RAML, RSDL, something else) which would make it possible auto-generate (Python) wrappers or runtime code generate them?
RWS comes in two versions:
RWS 1.0 (RobotWare 6) doesn't have any such API specs (as far as I know).
RWS 2.0 (RobotWare 7) is compliant with OpenAPI and Swagger can be used to generate client code, but I have not worked with RobotWare 7 systems yet.
Ah, great.
Is abb_libegm RW7 compatible?
Would there be a way to make sure? :)
Of course there are ;) I can check with a colleague tomorrow, but I am quite certain there aren't any.
Is abb_libegm RW7 compatible?
Did you mean abb_libegm or abb_librws?
abb_librws: Unfortunately not compatible, since RWS 1.0 has been deprecated in RW7. And I haven't really started thinking about a good approach to solved this yet, mainly since I haven't worked with RW7 systems.
abb_libegm: Should be compatible with RW7, but I haven't had time to test this yet.
Hm.
At least for RW7 we should be able to use something like OpenAPITools/openapi-generator.
@gavanderhoorn, FYI, I got a no regarding auto-generation of client code for RWS 1.0.
Too bad, but thanks for asking.
Closing this now.
|
2025-04-01T04:35:23.018476
| 2019-01-18T20:40:06
|
400881154
|
{
"authors": [
"christian-rauch",
"jproberge"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10423",
"repo": "ros-industrial/robotiq",
"url": "https://github.com/ros-industrial/robotiq/pull/142"
}
|
gharchive/pull-request
|
Mesh colour
This PR adds colour (black) to the articulated meshes. Fixes #141
Edit: I just saw that I had other commits in my devel branch. There are two commits to fix the installation and running of the simulation launch file.
@jproberge Is there any news on an "official" colour/material definition from Robotiq?
If this is not happening in the short term, I am happy if you change the colour to something slightly lighter.
Maybe the issue is also with the material definition and not the colour per se. In Blender, the palm mesh shows more shading than in Gazebo:
hi @christian-rauch ,
Thanks for your follow-up! I will have the information this week and I'll post back as soon as possible. Also, I agree that it depends on the visualisation environment. At the end, it's just aesthetic, so I don't want to delay this PR too much, and I'll post back here this week as soon as I get the information.
Thanks again!
jproberge
@christian-rauch,
I'm accepting this PR, since I've reviewed / tested it and found that it works flawlessly and is an improvement. Also, Robotiq told me that there were no precise color definition. In fact they ask their supplier to provide black anodized parts, but then these parts are also subject to a tumbler surface finish process, which might also alter the colour a little. Thus no precise colour definition. Thanks!
|
2025-04-01T04:35:23.071624
| 2016-05-02T12:44:06
|
152545650
|
{
"authors": [
"dirk-thomas",
"harvcode",
"tfoote"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10424",
"repo": "ros-visualization/python_qt_binding",
"url": "https://github.com/ros-visualization/python_qt_binding/issues/38"
}
|
gharchive/issue
|
kinetic version breaks rqt_graph
command ran: rosrun rqt_graph rqt_graph
error:
File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui/ros_plugin_provider.py", line 83, in load raise e ImportError: cannot import name QGraphicsPathItem
error originates from here:
File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_dotgraph/edge_item.py", line 32, in <module> from python_qt_binding.QtGui import QBrush, QGraphicsPathItem, QGraphicsPolygonItem, QGraphicsSimpleTextItem, QPainterPath, QPen, QPolygonF ImportError: cannot import name QGraphicsPathItem
It seems like you are using an older version of qt_dotgraph. Version 0.3.2 imports that class from QWidgets: https://github.com/ros-visualization/qt_gui_core/blob/kinetic-devel/qt_dotgraph/src/qt_dotgraph/edge_item.py#L33
I am using the version that gets installed automatically from the kinetic binaries, so perhaps the repository binaries need to be changed to reflect this?
Please state the exact version number of qt_dotgraph as well as rqt_graph you are using.
@harvcode The versions available from binary have changed recently. That's why we need to confirm the version on your system.
ros-kinetic-qt-dotgraph/xenial,now 0.3.2-0xenial-20160428-192006-0700 amd64
ros-kinetic-rqt-graph/xenial,now 0.4.0-1xenial-20160429-165937-0700 amd64
The code in the error message in your original comment (edge_item.py, line 32) does not match the code available in these versions.
Version 0.3.0 has the line you quoted above: https://github.com/ros-visualization/qt_gui_core/blob/0.3.0/qt_dotgraph/src/qt_dotgraph/edge_item.py#L32
Version 0.3.1 as well as 0.3.2 have this instead which should work:
https://github.com/ros-visualization/qt_gui_core/blob/0.3.1/qt_dotgraph/src/qt_dotgraph/edge_item.py#L32
https://github.com/ros-visualization/qt_gui_core/blob/0.3.2/qt_dotgraph/src/qt_dotgraph/edge_item.py#L32
Just tried it again and I still get "ImportError: cannot import name QFormLayout".
Apt says I am up to date on both of those packages and lists those versions.
Since the error is different then what you reported above please post the complete error message and steps you did to reproduce.
From command line: rqt
From menu: Plugins > Robot Tools > Controller Manager
Error appears in console.
Please post the full error message.
`
PluginManager._load_plugin() could not load plugin "rqt_controller_manager/ControllerManager":
Traceback (most recent call last):
File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/plugin_handler.py", line 99, in load
self._load()
File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/plugin_handler_direct.py", line 54, in _load
self._plugin = self._plugin_provider.load(self._instance_id.plugin_id, self._context)
File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/composite_plugin_provider.py", line 71, in load
instance = plugin_provider.load(plugin_id, plugin_context)
File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/composite_plugin_provider.py", line 71, in load
instance = plugin_provider.load(plugin_id, plugin_context)
File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui_py/ros_py_plugin_provider.py", line 60, in load
return super(RosPyPluginProvider, self).load(plugin_id, plugin_context)
File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/composite_plugin_provider.py", line 71, in load
instance = plugin_provider.load(plugin_id, plugin_context)
File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui/ros_plugin_provider.py", line 83, in load
raise e
ImportError: cannot import name QFormLayout
RosPluginProvider.load(rqt_controller_manager/ControllerManager) exception raised in builtin.import(rqt_controller_manager.controller_manager, [ControllerManager]):
Traceback (most recent call last):
File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui/ros_plugin_provider.py", line 77, in load
module = builtin.import(attributes['module_name'], fromlist=[attributes['class_from_class_type']], level=0)
File "/root/catkin_ws/src/ros_control/rqt_controller_manager/src/rqt_controller_manager/controller_manager.py", line 35, in
from python_qt_binding.QtGui import QCursor, QFont, QFormLayout, QHeaderView,
ImportError: cannot import name QFormLayout`
The stack trace indicates that the problem is in the package rqt_controller_manager which is in your local workspace (I assume it hasn't been released into Kinetic yet?). You might want to fill a ticket in the repository which contains that package: https://github.com/ros-controls/ros_control/issues
|
2025-04-01T04:35:23.081395
| 2022-07-27T07:58:37
|
1319169258
|
{
"authors": [
"rhaschke",
"ysl-design"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10425",
"repo": "ros-visualization/rviz",
"url": "https://github.com/ros-visualization/rviz/issues/1758"
}
|
gharchive/issue
|
The plug-in in the display panel sometimes does not shrink as the display panel compresses.
Describe your issue here and explain how to reproduce it.
Your environment
OS Version: Ubuntu 20.04
ROS Distro: Noetic
RViz, Qt, OGRE, OpenGl version as printed by rviz:[ INFO] [1658890616.733419800]: rviz version 1.14.14
[ INFO] [1658890616.733471700]: compiled against Qt version 5.12.8
[ INFO] [1658890616.733501700]: compiled against OGRE version 1.9.0 (Ghadamon)
[ INFO] [1658890616.741053500]: Forcing OpenGl version 0.
[ INFO] [1658890616.894986100]: Stereo is NOT SUPPORTED
[ INFO] [1658890616.895090400]: OpenGL device: llvmpipe (LLVM 12.0.0, 256 bits)
[ INFO] [1658890616.895148600]: OpenGl version: 3.1 (GLSL 1.4).
System locale, zh_CN.UTF-8
When I stretch the displays panel first, then compress it, and then double-click the plug-in component in the displays panel (such as the drop-down list), the picture below appears. It seems that the components don't adapt to size as the displays panel compresses. I do not know whether the problem is caused by the QT or the displays panel.
I cannot reproduce this issue. Could you please provide more detailed instructions and a video as well?
https://user-images.githubusercontent.com/5376030/181205794-3ee24b7f-cc94-46ae-82f4-a135c89b42a6.mp4
Sorry, I can't upload the video right now. The procedure is as follows:
step1:
step2:
step3:
result:
Thanks for the clarification. Now, I can reproduce the issue.
Could you please test https://github.com/ros-visualization/rviz/pull/1760?
I tried #1760 and the problem did not recur.
Closed via #1760.
|
2025-04-01T04:35:23.083947
| 2020-12-15T08:23:43
|
767314141
|
{
"authors": [
"stertingen"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10426",
"repo": "ros/catkin",
"url": "https://github.com/ros/catkin/pull/1127"
}
|
gharchive/pull-request
|
Added --force-color flag to catkin_make and catkin_make_isolated
This can be used to enforce colored output even on non-interactive output. --force-color can still be overriden by --no-color.
This PR is for CI environments where the output is stored in a file but a colored output makes it easier for a user to read the log.
There are some alternatives to the current implementation:
--force-color overrides --no-color
Error/Warning if --force-color and --no-color are both set
|
2025-04-01T04:35:23.095033
| 2015-07-10T10:06:29
|
94271157
|
{
"authors": [
"asmodehn",
"dirk-thomas"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10427",
"repo": "ros/catkin",
"url": "https://github.com/ros/catkin/pull/747"
}
|
gharchive/pull-request
|
fixing package_dir location logic.
package_dir: {'foo':'lib'} means 'lib/init.py' and not 'lib/foo/init.py'
It seems some tests should also be changed... I d rather have some feedback about this pull request before diving into all the tests that might need to be changed to follow the same behavior as the distutils behavior ( that you can test with python setup.py install ) so that our devel/ and install/ folder endup being similar.
references :
https://docs.python.org/2/distutils/setupscript.html#listing-whole-packages
https://docs.python.org/2/distutils/examples.html#pure-python-distribution-by-package
Can you please provide some more information what the current problem is as well as step-by-step instructions how to reproduce the problem.
Current problem :
If your files structure to access your python modules is :
python-mod-name/setup.py
python-mod-name/python_module_1/__init__.py
python-mod-name-2/setup.py
python-mod-name-2/python_module_2/__init__.py
which enables you to have submodules pointing to the source repository of a package by the way, instead of the usual :
src/python_module_1/__init__.py
src/python_module_2/__init__.py
and if you have in your setup.py :
d = generate_distutils_setup(
packages=[
'python_module_1',
'python_module_2',
],
package_dir={
'python_module_1': 'python-mod-name',
'python_module_2': 'python-mod-name-2',
})
the catkin_make build works fine. The __init__.py is found in devel/lib/python2.7/dist-packages/python_module.
However catkin_make install doesn't install the package properly. instead the setup.py file ends up in install/lib/python2.7/dist-packages/python_module.
Also python setup.py install is completely broken.
You can test that on this revision : https://github.com/asmodehn/flask-ext-catkin.git @ de04b20773460d86f7d19b3970eb5728c09e901b
Using the proper distutils parameter for the catkin version also doesn't work. :
Using 'python_module_1': 'python-mod-name/python_module_1', makes catkin expect the __init__.py file in python-mod-name/python_module_1/python_module_1 which is incorrect.
The reason is that the logic to find the location implemented in catkin is different than the logic implemented in distutils ( which makes things very confusing )
Solution : fixing the logic used to determine locations, and
d = generate_distutils_setup(
packages=[
'python_module_1',
'python_module_2',
],
package_dir={
'python_module_1': 'python-mod-name/python_module_1',
'python_module_2': 'python-mod-name-2/python_module_2',
})
which matches the distutils logic and solves both problem.
catkin_make build and catkin_make install now work properly.
also python setup.py install works as expected.
Implemented here : https://github.com/asmodehn/flask-ext-catkin.git @ 41fd3f7077ec7d8cb4a54a89ed830a212c387a5d
Thank you for the patch!
I updated the condition I have commented on in #751. In that branch I have also updated the tests to match the new behavior.
Can you please check and confirm that the new PR also works for you?
Thanks ! I will check it over the next few days and let you know.
On Aug 6, 2015 8:02 AM, "Dirk Thomas"<EMAIL_ADDRESS>wrote:
Thank you for the patch!
I updated the condition I have commented on in #751
https://github.com/ros/catkin/pull/751. In that branch I have also
updated the tests to match the new behavior.
Can you please check and confirm that the new PR also works for you?
—
Reply to this email directly or view it on GitHub
https://github.com/ros/catkin/pull/747#issuecomment-128176931.
I am closing this in favor of the new PR #751.
|
2025-04-01T04:35:23.109071
| 2016-02-12T12:15:29
|
133226713
|
{
"authors": [
"dirk-thomas",
"nlyubova",
"tfoote"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10428",
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/10486"
}
|
gharchive/pull-request
|
romeo_moveit_actions: 0.0.7-2 in 'indigo/distribution.yaml' [bloom]
Increasing version of package(s) in repository romeo_moveit_actions to 0.0.7-2:
upstream repository: https://github.com/nlyubova/romeo_moveit_actions.git
release repository: https://github.com/nlyubova/romeo_moveit_actions-release.git
distro file: indigo/distribution.yaml
bloom version: 0.5.20
previous version for package: 0.0.7-0
romeo_moveit_actions
* Merge pull request #1 <https://github.com/nlyubova/romeo_moveit_actions/issues/1> from IanTheEngineer/remove_shape_tools
Convert deprecated shape_tools dependency
* Convert deprecated shape_tools dependency
shape_tools functionality was merged into geometric_shapes:
https://github.com/ros-planning/geometric_shapes/pull/32
and removed from moveit_core
https://github.com/ros-planning/moveit_core/pull/242
which caused this issue.
This commit updates the pick and place tutorial and adds
geometric_shapes to the package.xml and CMakeLists.txt to
prevent the ROS buildfarm from failing to build this package.
* Contributors: Ian McMahon, Natalia Lyubova
Please either change the patch to not change the comment line manually or create a new patch using bloom-release ... -r after updating to the latest release of python-rosdistro.
That's bloom-release with --pull-request-only or -p not -r You need python-rosdisto 0.4.4 or higher.
|
2025-04-01T04:35:23.112324
| 2022-04-29T02:44:22
|
1219652609
|
{
"authors": [
"ihadzic",
"methylDragon"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10429",
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/32944"
}
|
gharchive/pull-request
|
Two new packages to add to index: move_base_swp and lsm_localization
Please Add This Package to be indexed in the rosdistro.
melodic, noetic
The source is here:
https://github.com/ihadzic/move_base_swp
https://github.com/ihadzic/lsm_localization
Checks
[x] All packages have a declared license in the package.xml
[x] This repository has a LICENSE file
[x] This package is expected to build on the submitted rosdistro
Holding for Noetic sync
|
2025-04-01T04:35:23.113885
| 2023-03-31T18:10:56
|
1649808141
|
{
"authors": [
"ChrisThrasher",
"audrow"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10430",
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/36672"
}
|
gharchive/pull-request
|
Revert "Revert "rsl: 0.2.2-1 in 'humble/distribution.yaml' [bloom]""
Reverts ros/rosdistro#36613
@ChrisThrasher, the Humble sync just went out. We can see if this breaks any downstream packages and try to fix them before the next Humble sync in 2-3 weeks.
This release of pick_ik should fix at least one of the build farm errors were previously encountering.
|
2025-04-01T04:35:23.118498
| 2024-04-22T06:10:20
|
2255664809
|
{
"authors": [
"Timple"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10431",
"repo": "ros/rosdistro",
"url": "https://github.com/ros/rosdistro/pull/40717"
}
|
gharchive/pull-request
|
azure_iot_sdk_c: 1.13.0-3 in 'rolling/distribution.yaml' [bloom]
Increasing version of package(s) in repository azure_iot_sdk_c to 1.13.0-3:
upstream repository: https://github.com/Azure/azure-iot-sdk-c.git
release repository: https://github.com/ros2-gbp/azure_iot_sdk_c-release.git
distro file: rolling/distribution.yaml
bloom version: 0.12.0
previous version for package: null
Replaces https://github.com/ros/rosdistro/pull/40536
I might need a pointer here...
The build fails on: https://build.ros2.org/job/Rdev__azure_iot_sdk_c__ubuntu_noble_amd64/1/console
File "/usr/lib/python3/dist-packages/colcon_cmake/task/cmake/build.py", line 173, in _reconfigure
raise RuntimeError("Could not find 'cmake' executable")
However cmake seems correctly listed as build dependency: https://github.com/ros2-gbp/azure_iot_sdk_c-release/blob/master/patches/package.xml#L11C1-L12C1
|
2025-04-01T04:35:23.182620
| 2021-02-23T19:56:24
|
814784326
|
{
"authors": [
"clalancette",
"stephenstarkie"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10432",
"repo": "ros2/ros2_documentation",
"url": "https://github.com/ros2/ros2_documentation/issues/1131"
}
|
gharchive/issue
|
Install ROS2 on macOS fails with pygraphviz
I tried to install according to the instructions here: https://index.ros.org/doc/ros2/Installation/Foxy/macOS-Install-Binary/
Installation fails on the line:
python3 -m pip install pygraphviz pydot
with;
Collecting pygraphviz
Using cached pygraphviz-1.7.zip (118 kB)
Building wheels for collected packages: pygraphviz
Building wheel for pygraphviz (setup.py) ... error
ERROR: Command errored out with exit status 1:
command<EMAIL_ADDRESS>-u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-install-j0l_0s3c/pygraphviz_ac93efbdfc764377ae19b8ff1b8c3d95/setup.py'"'"'; file='"'"'/private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-install-j0l_0s3c/pygraphviz_ac93efbdfc764377ae19b8ff1b8c3d95/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(file);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, file, '"'"'exec'"'"'))' bdist_wheel -d /private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-wheel-gsy173sm
cwd: /private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-install-j0l_0s3c/pygraphviz_ac93efbdfc764377ae19b8ff1b8c3d95/
Complete output (311 lines):
running bdist_wheel
running build
running build_py
creating build
creating build/lib.macosx-11-x86_64-3.8
creating build/lib.macosx-11-x86_64-3.8/pygraphviz
copying pygraphviz/scraper.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz
copying pygraphviz/graphviz.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz
copying pygraphviz/init.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz
copying pygraphviz/agraph.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz
copying pygraphviz/testing.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz
creating build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_unicode.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_scraper.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_readwrite.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_string.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/init.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_html.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_node_attributes.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_drawing.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_subgraph.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_close.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_edge_attributes.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_clear.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_layout.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_attribute_defaults.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
copying pygraphviz/tests/test_graph.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests
running egg_info
writing pygraphviz.egg-info/PKG-INFO
writing dependency_links to pygraphviz.egg-info/dependency_links.txt
writing top-level names to pygraphviz.egg-info/top_level.txt
reading manifest file 'pygraphviz.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
warning: no files found matching '.png' under directory 'doc'
warning: no files found matching '.txt' under directory 'doc'
warning: no files found matching '.css' under directory 'doc'
warning: no previously-included files matching '~' found anywhere in distribution
warning: no previously-included files matching '*.pyc' found anywhere in distribution
warning: no previously-included files matching '.svn' found anywhere in distribution
no previously-included directories found matching 'doc/build'
writing manifest file 'pygraphviz.egg-info/SOURCES.txt'
copying pygraphviz/graphviz.i -> build/lib.macosx-11-x86_64-3.8/pygraphviz
copying pygraphviz/graphviz_wrap.c -> build/lib.macosx-11-x86_64-3.8/pygraphviz
running build_ext
building 'pygraphviz._graphviz' extension
creating build/temp.macosx-11-x86_64-3.8
creating build/temp.macosx-11-x86_64-3.8/pygraphviz
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk -I/usr/local/include<EMAIL_ADDRESS>-I/usr/local/opt/sqlite/include -I/usr/local/opt/tcl-tk/include<EMAIL_ADDRESS>-c pygraphviz/graphviz_wrap.c -o build/temp.macosx-11-x86_64-3.8/pygraphviz/graphviz_wrap.o
In file included from pygraphviz/graphviz_wrap.c:154:
In file included from /usr/local/Cellar/python@3.8/3.8.8/Frameworks/Python.framework/Versions/3.8/include/python3.8/Python.h:34:
In file included from /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/stdlib.h:66:
In file included from /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/sys/wait.h:110:
In file included from /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/sys/resource.h:72:
In file included from /usr/local/include/stdint.h:59:
In file included from /usr/local/include/stdint.h:59:
...lots of the same line
In file included from /usr/local/include/stdint.h:59:
In file included from /usr/local/include/stdint.h:59:
/usr/local/include/stdint.h:2:10: error: #include nested too deeply
macos 11.2.1
We don't officially support Big Sur with any of our platforms, so it is not entirely surprising. That being said, this looks like a problem in pygraphviz, nothing specific to ROS 2 here. So I'm going to close this out; I'll suggest that you open an issue against https://github.com/pygraphviz/pygraphviz instead.
|
2025-04-01T04:35:23.220750
| 2023-12-03T22:46:11
|
2022749262
|
{
"authors": [
"WardBrian",
"bob-carpenter",
"roualdes"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10433",
"repo": "roualdes/bridgestan",
"url": "https://github.com/roualdes/bridgestan/issues/190"
}
|
gharchive/issue
|
Python callback into BridgeStan with C types
I have an implementation of Stan that iterates and does warmup calculations in Python and leapfrog/trajectories in C++. The gradient calculations implicit in each leapfrog step happen via a Python callback which then calls BridgeStan: Python -> C++ -> Python/BridgeStan -> C/C++/BridgeStan.
Since the C++ code is calling a BridgeStan function, and Python/BridgeStan expects numpy arrays, I've got to convert a double* to a numpy array just so that BridgeStan can convert the numpy array back to double*. The conversion to a numpy array is unnecessarily costly, especially since everything is using the same underlying memory.
Consider BridgeStan's python/bridgestan/model.py lines 219 to 229. This code chunk defines the Python/BridgeStan interface to the C/BridgeStan function bs_log_density_gradient. Lines 225 and 227 specify numpy arrays as argument types from Python/BridgeStan -> C/C++/BridgeStan.
Simply adding the following code to python/bridgestan/model.py::StanModel's __init__() method allows a Python callback from C++ to re-use the original double*:
self._ldg = self.stanlib["bs_log_density_gradient"]
self._ldg.restype = ctypes.c_int
self._ldg.argtypes = [
ctypes.c_void_p,
ctypes.c_int,
ctypes.c_int,
ctypes.POINTER(ctypes.c_double), # changed from double_array
ctypes.POINTER(ctypes.c_double),
ctypes.POINTER(ctypes.c_double), # changed from double_array
star_star_char,
]
What do you think of building such functionality into BridgeStan? Just adding the above code won't break our API in anyway, but does offer a whole new world of possibility for BridgeStan. I'm not sold on the name _ldg, I'm just asking about the possibility of including such functionality.
If y'all are into this idea, would you brainstorm with me what a better design might look like that allow callbacks, from the various higher level languages, into BridgeStan using plain C types?
What do you think of building such functionality into BridgeStan?
I couldn't quite follow all the callbacks, so I'm not entirely sure what you are suggesting adding to BridgeStan and what that would enable for clients that they can't achieve now.
My understanding of Edward's request is that a user may be working with other code that uses ctypes, and as a result have a double array which is represented as a ctypes.POINTER(ctypes.c_double), rather than a numpy array object.
There is nothing at the C level of BridgeStan to differentiate between these, and in principle either is fine, but for convenience BridgeStan only speaks the language of numpy ndarrays, not 'raw' ctypes pointers. You can convert from a ctypes pointer to an ndarray, but since all BridgeStan really does is some bounds checking and then convert it back to a pointer internally, this comes with an extra cost at no extra benefit.
The concrete code Ed provided would give you a different function which accepts ctypes objects as opposed to numpy ones, but I also have a minimal patch prepared which would make it so either can be accepted anywhere we currently accept numpy arrays, meaning no extra functions, just slightly broader types being accepted.
Got it. Thanks!
Right, this doesn't seem to be an issue outside of Python, for the reasons you mention. Good call.
Brian proposed some code, in branch python/allow-ctypes-double-pointers, that allows direct access to BridgeStan's log_density_gradient() for both numpy arrays and ctypes. The solution involves an extra if statement, relative to what lives in main branch. To help us determine whether or not we want to include such functionality in BridgeStan, I ran some experiments and report the results below.
I ran some numerical simulations to better understand the computational cost trade-offs associated with this extra if statement. From a numpy array user perspective, we expect a slight slow down (one extra if statement). From a ctypes user perspective, we expect a major gain (no unnecessary numpy array creation). The code used for the simulations lives in the github repository bridgestan-speed. The simulations were across three models: gaussian with data (2 parameters), logistic regression (25 parameters), and standard normal with no data (1_000 parameters).
Numpy array perspective. 500 runs of 1_000 evaluations of log_density_gradient
across two branches. Times reported are means per run plus/minus standard
deviations.
dimensions / branch main python/allow-ctypes-double-pointers
2 (normal w/ data) 13.9 ms +/- 40.9 μs 14.6 ms +/- 493 μs
25 (logistic regression) 96.2 ms +/- 8.42 ms 82.7 ms +/- 1.13 ms
1_000 (std_normal no data) 27.5 ms +/- 1.48 ms 27.1 ms +/- 721 μs
Ctypes perspective. 20 runs of Stan sampling with 1_000 warmup and 1_000
iterations across two branches. Times reported are means per Stan run
plus/minus standard deviations. RNG seeds were coordinated to ensure no branch
lucked into a worse part of the model parameter space than the other branch.
dimensions / branch main python/allow-ctypes-double-pointers
2 (normal w/ data) 251 ms +/- 7.36 ms 89.9 ms +/- 5.3 ms
25 (logistic regression) 2.56 s +/- 132 ms 2 s +/- 74.9 ms
1_000 (std_normal no data) 1.51 s +/- 30.3 ms 908 ms +/- 92 ms
What do you all think, should we add this functionality to BridgeStan?
Thanks for the report @roualdes. I couldn't quite tell what was being compared. I see two tables of results, but don't know what they mean. What does "across two branches" mean for the "Numpy array perspective"? Is it that the first one is passing in an numpy array and the second passing in some raw memory?
What does "across two branches" mean for the "Numpy array perspective"?
The first table is meant to show costs associated with the proposed solution to this issue for the most common Python/BridgeStan user, a user who is only dealing with numpy arrays. The table presents mean run times for calling BridgeStan's log_density_gradient with numpy arrays. The first column sets the baseline for run times in the branch main. The second column displays run times for the proposed solution from branch python/allow-ctypes-double-pointers.
Is it that the first one is passing in an numpy array and the second passing in some raw memory?
Correct. The second table is meant to show benefits associated with the proposed solution when a user has a double* to raw memory.
Let me know if I can offer any other words to help clarify what my results are about.
Thanks---that's what I suspected and this looks close enough to be worth doing.
If the results weren't clear enough here, I'd be tempted to do about 10 times as many evals over several batches because the inconsistent sd in the tests you list (e.g., "13.9 ms +/- 40.9 μs 14.6 ms +/- 493 μs" in the first line of the first table, where the new method has ten times the standard deviation). This kind of micro-benchmarking is notoriously tricky, so probably not worth it.
They align with my expectation - for numpy arrays it is a wash and for raw pointers the overhead is much less.
If it had ended up being a significant increase for numpy arrays I would be more hesitant toward adding it. However, it's really just adding one more if into a part of the code that already has a bunch of ifs for validation logic, so I'm not shocked the impact is essentially nothing.
|
2025-04-01T04:35:23.222438
| 2019-09-22T00:56:03
|
496718354
|
{
"authors": [
"roughike",
"srihamat"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10434",
"repo": "roughike/flutter_facebook_login",
"url": "https://github.com/roughike/flutter_facebook_login/pull/182"
}
|
gharchive/pull-request
|
Update README
I spent 2 days looking for solution then eventually found this is the solution, look like someone also face the same issues as well. Update solution here should help save time for more people
Thanks! I'll modify this a tiny bit, but LGTM.
|
2025-04-01T04:35:23.293456
| 2022-08-03T14:02:06
|
1327270684
|
{
"authors": [
"remorses",
"rpearce"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10435",
"repo": "rpearce/react-medium-image-zoom",
"url": "https://github.com/rpearce/react-medium-image-zoom/issues/333"
}
|
gharchive/issue
|
Document how to change background color
In v4 you could use the overlayBgColorStart props to change background color, now you can use the following css (in this case i am also changing background when in dark mode)
[data-rmiz-modal-overlay='hidden'] {˙
background-color: rgba(255, 255, 255, 0);
}
[data-rmiz-modal-overlay='visible'] {
background-color: rgba(255, 255, 255, 0.6);
backdrop-filter: saturate(180%) blur(60px);
-webkit-backdrop-filter: saturate(180%) blur(60px);
}
.dark [data-rmiz-modal-overlay='hidden'] {
background-color: rgba(0, 0, 0, 0);
}
.dark [data-rmiz-modal-overlay='visible'] {
background-color: rgba(0, 0, 0, 0.6);
}
@remorses Thanks for calling that out. Yes, nearly all style-related things (including transition times) have moved to being CSS controlled now, and you're right — it would be helpful to document that somehow.
@remorses Thank you for bringing this up and, in doing so, contributing to this project!
I added info to the Styles section in the readme, including a link to an example of how to do custom modal styling.
Thank you, amazing project btw!
|
2025-04-01T04:35:23.298500
| 2015-02-11T20:08:14
|
57369930
|
{
"authors": [
"rpetz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10436",
"repo": "rpetz/SharpNetSH",
"url": "https://github.com/rpetz/SharpNetSH/issues/1"
}
|
gharchive/issue
|
Maintenance nightmare
Currently responses to each method call are parsed by hand on a method by method basis. This creates a maintenance nightmare that I do not want to personally support. Considering we are following convention over configuration for the method calls themselves (no concrete classes, all interface proxying) we should do the same thing for the response objects.
Rather than using a hand coded response object class for each method call, response objects should be standardized with the following code:
public int ExitCode { get; }
public Boolean IsNonStandardExitCode { get { return ExitCode != 0; } }
public IEnumerable<String> Response { get; }
public dynamic ResponseObject { get; }
The dynamic response object should follow a set of standardized parsing conventions which should be configured using an attribute on the method (since different methods have different conventions).
This has been resolved in the latest code for the master branch
|
2025-04-01T04:35:23.552588
| 2022-03-30T06:42:56
|
1185965249
|
{
"authors": [
"ened",
"rrousselGit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10438",
"repo": "rrousselGit/riverpod",
"url": "https://github.com/rrousselGit/riverpod/issues/1326"
}
|
gharchive/issue
|
examples/stackoverflow: Fast Scroll/Small Page Size will cause App to hang
Describe the bug
I try to understand the stackoverflow example, specifically how the pagination is done.
The forked version of the repo is here:
https://github.com/ened/river_pod/tree/examples/stackoverflow/pageSize5
The page size was changed to 5 and I started scrolling quickly.
This caused a few errors and eventually, the App is not loading any new data.
Expected behavior
Open requests to missing pages should load appropriately and backfill what is missing.
App should never hang and stop loading
This caused a few errors and eventually, the App is not loading any new data.
What error?
Errors like:
Exception has occurred.
DioError (DioError [DioErrorType.cancel]:
Source stack:
#0 DioMixin.fetch (package:dio/src/dio_mixin.dart:473:35)
#1 DioMixin.request (package:dio/src/dio_mixin.dart:468:12)
#2 DioMixin.requestUri (package:dio/src/dio_mixin.dart:422:12)
#3 DioMixin.getUri (package:dio/src/dio_mixin.dart:72:12)
#4 paginatedQuestionsProvider.<anonymous closure> (package:stackoverflow/home.dart:79:8)
#5 paginatedQuestionsProvider.<anonymous closure> (package:stackoverflow/home.dart:54:37)
#6 AutoDisposeFutureProviderFamily.create.<anonymous closure> (package:riverpod/src/future_provider/auto_dispose.dart:132:23)
#7 AutoDisposeFutureProvider.create.<anonymous closure> (package:riverpod/src/future_provider/auto_dispose.dart:68:43)
#8 _FutureProviderElementMixin._listenFuture (package:riverpod/src/future_provider.dart:89:27)
#9 AutoDisposeFutureProvider.create (package:riverpod/src/future_provider/auto_dispose.dart:68:16)
#10 ProviderElementBase._buildState (package:riverpod/src/framework/provider_base.dart:476:26)
#11 AutoDisposeProviderElementBase._buildState (package:riverpod/src/framework/auto_dispose.dart:152:11)
#12 ProviderElementBase.mount (package:riverpod/src/framework/provider_base.dart:348:5)
#13 _StateReader._create (package:riverpod/src/framework/container.dart:104:11)
#14 _StateReader.getElement (package:riverpod/src/framework/container.dart:92:52)
#15 ProviderContainer.readProviderElement.<anonymous closure> (package:riverpod/src/framework/container.dart:490:38)
#16 ProviderContainer.readProviderElement (package:riverpod/src/framework/container.dart:523:6)
════════ Exception caught by image resource service ════════════════════════════
The following HandshakeException was thrown resolving an image codec:
Connection terminated during handshake
When the exception was thrown, this was the stack
Image provider: NetworkImage("[https://lh3.googleusercontent.com/a/AATXAJwfy-25HkvYx5KfDm8sxqtm0yMCLIoLgxTofHrJ=k-s256]()", scale: 1.0)
Image key: NetworkImage("[https://lh3.googleusercontent.com/a/AATXAJwfy-25HkvYx5KfDm8sxqtm0yMCLIoLgxTofHrJ=k-s256]()", scale: 1.0)
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by image resource service ════════════════════════════
Connection terminated during handshake
════════════════════════════════════════════════════════════════════════════════
════════ Exception caught by image resource service ════════════════════════════
If waited for long enough (or release build) the errors seem to have no impact yet interrupt the development workflow.
If waited for long enough (or release build) the errors seem to have no impact yet interrupt the development workflow.
Maybe too many requests were made at the same time due to the small page side.
Closing that the Riverpod error shouldn't happen anymore. (well the exception should be thrown, but that's on purpose. On the other hand, it shouldn't be considered "uncaught")
|
2025-04-01T04:35:23.554780
| 2023-12-05T01:26:56
|
2025101624
|
{
"authors": [
"rruckley"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10439",
"repo": "rruckley/tmflib",
"url": "https://github.com/rruckley/tmflib/pull/15"
}
|
gharchive/pull-request
|
Fix quote item
Added missing fields from QuoteItem
Added missing ProductRefOrValue struct
Added missing fields from Quote
Changed visibility to public for most fields in QuoteItem
Again merging forward.
|
2025-04-01T04:35:23.561273
| 2020-06-11T07:10:06
|
636776248
|
{
"authors": [
"dadepo",
"koivunej"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10440",
"repo": "rs-ipfs/rust-ipfs",
"url": "https://github.com/rs-ipfs/rust-ipfs/issues/186"
}
|
gharchive/issue
|
Using rust-ipfs as a library
This is not a feature request but a request for clarification on the usage of rust-ipfs as a library
Currently I am interacting with IPFS by calling out to an installed binary.
So far, here are the commands I need for my interactions:
//1
// Ensure ipfs is started with --enable-pubsub-experiment and --enable-namesys-pubsub
//2
ipfs add -r /path_to_dir
//3
ipfs name publish --key=hash-public-key cid_to_publish
//4
ipfs get "ipns/path"--output=destination_path
My questions now are:
I know rust-ipfs is pre-alpha, but does it currently support --enable-pubsub-experiment and --enable-namesys-pubsub?
Does it also currently support adding a directory, publishing to ipns and requesting from ipns?
Would there be a theoretical/real performance benefit of using rust-ipfs as a library compared with interacting with an installed binary?
Thanks!
Thanks for reaching out!
I know rust-ipfs is pre-alpha, but does it currently support --enable-pubsub-experiment and --enable-namesys-pubsub?
The current build has --enable-pubsub-experiment always on, and in fact, it cannot be turned off via configuration. The inability to configure it off with the #132 makes it so that you cannot connect to go-ipfs 0.5 which is running with --enable-pubsub-experiment.
We currently do not support IPNS at all, so no support for IPNS over pubsub either.
Does it also currently support adding a directory, publishing to ipns and requesting from ipns?
We do not currently have an unixfs importer or the add but we have planned implementing it rather soon. Currently I am working over at the unixfs reading support #172 and /cat is looking ok with #184 which will leave the main library or crate missing out the IpfsPath resolving, which has had some iterations over it. The IpfsPath enhancements should be following rather soon.
I don't think we have any concrete plans for IPNS. On my mind the path has been to first fix the dns story #83 to enable dnslink, then implement keystore management followed by initial work on IPNS over DHT. At that point I expect the libp2p-gossipsub to be compatible with go-ipfs and go-ipfs defaulting to it as well, so the IPNS over pubsub would be next logical step.
We are currently gathering Phase 3 ideas over at #181 but since this issue was already created: Could you open a bit your use case up? Any details would be much appreciated.
Would there be a theoretical/real performance benefit of using rust-ipfs as a library compared with interacting with an installed binary?
At least the theoretical benefit is there but at least for now, it might be non-existent or only caused by the fact that we implement so few features compared to {go,js}-ipfs. In the long run we aim to provide a predictable and trustworthy experience as crate and of course rust and LLVMs optimization capabilities will shine compared to communicating over HTTP when we have the features working.
Thanks for the response. I have enough information to make my decision regarding usage of the current version rust-ipfs. Tnanks!
|
2025-04-01T04:35:23.574035
| 2018-07-17T08:47:26
|
341822948
|
{
"authors": [
"bpoplauschi",
"codecov-io",
"dreampiggy"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10441",
"repo": "rs/SDWebImage",
"url": "https://github.com/rs/SDWebImage/pull/2388"
}
|
gharchive/pull-request
|
Update the migration guide with the latest version.
Update one changelog
New Pull Request Checklist
[x] I have read and understood the CONTRIBUTING guide
[x] I have read the Documentation
[x] I have searched for a similar pull request in the project and found none
[x] I have updated this branch with the latest master to avoid conflicts (via merge from master or rebase)
[x] I have added the required tests to prove the fix/feature I am adding
[x] I have updated the documentation (if necessary)
[x] I have run the tests and they pass
[x] I have run the lint and it passes (pod lib lint)
This merge request fixes / reffers to the following issues: ...
Pull Request Description
This PR contains any update for the wiki/documentation/changelog. You can just put anything to change for these into single PR.
Everyone who need to change documentations can directlly add commit here. So we don't need to merge again and again.
Until final update is done, we can merge and release the 5.0 version (Or another beta version if we find something important to change and test)
Good idea @dreampiggy
Codecov Report
Merging #2388 into 5.x will decrease coverage by 0.06%.
The diff coverage is n/a.
@@ Coverage Diff @@
## 5.x #2388 +/- ##
==========================================
- Coverage 68.77% 68.71% -0.07%
==========================================
Files 47 47
Lines 6607 6607
==========================================
- Hits 4544 4540 -4
- Misses 2063 2067 +4
Flag
Coverage Δ
#iOS
69.72% <ø> (-0.02%)
:arrow_down:
#macOS
68.11% <ø> (-0.02%)
:arrow_down:
Impacted Files
Coverage Δ
SDWebImage/SDWebImageManager.m
81.3% <0%> (-0.82%)
:arrow_down:
SDWebImage/SDWebImageDownloaderOperation.m
87.62% <0%> (-0.69%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 957de6d...ffad516. Read the comment docs.
@bpoplauschi This one now been updated to the latest 5.x. Maybe it's time to merge.
|
2025-04-01T04:35:23.598133
| 2021-11-05T12:47:11
|
1045813029
|
{
"authors": [
"TheTams",
"rsh249"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10442",
"repo": "rsh249/cRacle",
"url": "https://github.com/rsh249/cRacle/issues/4"
}
|
gharchive/issue
|
Define the 7 outputs from get_optim() in documentation
Currently the titles of the lists in the output from get_optim() do not make it clear to the user, which output corresponds which method as discussed in Harbert and Baryiames 2019. I have been unable to find any documentation that defines them elsewhere. The categories are:
conintkde
conintgauss
origk
origg
dircontint
means
sds
I will work on this documentation as well. These names are a holdover from how I initially set up the models and output collection, but we worked on the terms in the last publication a bit more to better align with the modeling that was actually taking place.
For a quick reference:
The "*kde" and "*k" methods are non-parametric models described in the Harbert & Baryiames, 2019 paper.
The "*gauss" and "*g" methods are the parametric models.
"conint" returns 95% confidence intervals of both methods
"orig" returns an estimate of the maximum likelihood value (the exact values aren't calculated but a narrow interval around the maximum is easy to get.
dirconint is functionally equivalent to conintgauss but uses the weighted means method instead of the ML method.
means is equivalent to origg but uses the weighted means method instead of the ML method.
sds is a calculation to go with means of the weighted standard deviation (something you don't get with the ML method but I haven't looked into much).
I hope this helps for now.
Thank you for that, I thought I had figured some of it out...
So is the weighted means the equivalent of the MCRM from Thompson et al. (2012)? And none of these are equivalent to the Coexistence Approach?
The weighted means is equivalent to the parametric CRACLE. There is a separate function for MCR a la Thompson et al. 2012 that does the unweighted MCR which (I think) is the same as the coexistence approach.
Sorry, my bad, I see the MCR one there. Great!
|
2025-04-01T04:35:23.603608
| 2020-03-03T21:40:21
|
574965784
|
{
"authors": [
"StriveForBest",
"WoLpH",
"bharling",
"pablolmedorado"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10443",
"repo": "rsinger86/drf-flex-fields",
"url": "https://github.com/rsinger86/drf-flex-fields/issues/46"
}
|
gharchive/issue
|
Adding FlexFieldsFilterBackend immediately breaks the app
I am utilizing drf-flex-fields heavily and everything works fine until the moment i add:
REST_FRAMEWORK = {
...
'DEFAULT_FILTER_BACKENDS': (
'rest_flex_fields.filter_backends.FlexFieldsFilterBackend',
),
...
}
It crashes the app with the following traceback:
INFO 2020-03-03 21:38:35,864 autoreload 90410<PHONE_NUMBER> Watching for file changes with StatReloader
Performing system checks...
Exception in thread django-main-thread:
Traceback (most recent call last):
File<EMAIL_ADDRESS>line 932, in _bootstrap_inner
self.run()
File<EMAIL_ADDRESS>line 870, in run
self._target(*self._args, **self._kwargs)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/autoreload.py", line 53, in wrapper
fn(*args, **kwargs)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/management/commands/runserver.py", line 117, in inner_run
self.check(display_num_errors=True)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/management/base.py", line 392, in check
all_issues = self._run_checks(
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/management/base.py", line 382, in _run_checks
return checks.run_checks(**kwargs)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/checks/registry.py", line 72, in run_checks
new_errors = check(app_configs=app_configs)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/checks/urls.py", line 13, in check_url_config
return check_resolver(resolver)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/checks/urls.py", line 23, in check_resolver
return check_method()
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/resolvers.py", line 407, in check
for pattern in self.url_patterns:
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/functional.py", line 48, in __get__
res = instance.__dict__[self.name] = self.func(instance)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/resolvers.py", line 588, in url_patterns
patterns = getattr(self.urlconf_module, "urlpatterns", self.urlconf_module)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/functional.py", line 48, in __get__
res = instance.__dict__[self.name] = self.func(instance)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/resolvers.py", line 581, in urlconf_module
return import_module(self.urlconf_name)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1014, in _gcd_import
File "<frozen importlib._bootstrap>", line 991, in _find_and_load
File "<frozen importlib._bootstrap>", line 975, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 671, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 783, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/urls.py", line 6, in <module>
path('api/v1/', include('gagosian.api.v1.urls', namespace='api'))
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/conf.py", line 34, in include
urlconf_module = import_module(urlconf_module)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1014, in _gcd_import
File "<frozen importlib._bootstrap>", line 991, in _find_and_load
File "<frozen importlib._bootstrap>", line 975, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 671, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 783, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/api/v1/urls.py", line 4, in <module>
from .contacts import urls as contacts_urls
File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/api/v1/contacts/urls.py", line 3, in <module>
from .viewsets import (
File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/api/v1/contacts/viewsets.py", line 1, in <module>
from rest_framework.viewsets import ModelViewSet
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/viewsets.py", line 27, in <module>
from rest_framework import generics, mixins, views
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/generics.py", line 24, in <module>
class GenericAPIView(views.APIView):
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/generics.py", line 43, in GenericAPIView
filter_backends = api_settings.DEFAULT_FILTER_BACKENDS
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 220, in __getattr__
val = perform_import(val, attr)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 168, in perform_import
return [import_from_string(item, setting_name) for item in val]
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 168, in <listcomp>
return [import_from_string(item, setting_name) for item in val]
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 177, in import_from_string
return import_string(val)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/module_loading.py", line 17, in import_string
module = import_module(module_path)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_flex_fields/__init__.py", line 3, in <module>
from .views import FlexFieldsModelViewSet
File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_flex_fields/views.py", line 21, in <module>
class FlexFieldsModelViewSet(FlexFieldsMixin, viewsets.ModelViewSet):
AttributeError: partially initialized module 'rest_framework.viewsets' has no attribute 'ModelViewSet' (most likely due to a circular import)
Adding the backend directly to the view works fine.
Yeah seeing this too, but I get:
ImportError: Could not import 'rest_flex_fields.filter_backends.FlexFieldsFilterBackend' for API setting 'DEFAULT_FILTER_BACKENDS'. ImportError: cannot import name 'GenericViewSet'.
looks like a circular import also
As a workaround (I also posted this at #59 which appears to be a duplicate) you can set the filter backend at the view:
import rest_flex_fields.filter_backends as flex_filters
from rest_framework import viewsets
class ModelViewSet(viewsets.ModelViewSet):
filter_backends = viewsets.ModelViewSet.filter_backends + [
flex_filters.FlexFieldsFilterBackend,
]
Starting from @WoLpH idea, I have created a mixin to replace the "FlexFieldsMixin" that includes the FilterBackend and also respects the user settings.
I'll leave the code here just in case there was anyone interested.
from rest_flex_fields.filter_backends import FlexFieldsFilterBackend
from rest_flex_fields.views import FlexFieldsMixin
class FlexFieldsOptimizedMixin(FlexFieldsMixin):
filter_backends = [FlexFieldsFilterBackend] + api_settings.DEFAULT_FILTER_BACKENDS
|
2025-04-01T04:35:23.607048
| 2024-09-25T01:00:14
|
2546665427
|
{
"authors": [
"jeremy-then"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10444",
"repo": "rsksmart/rootstock-integration-tests",
"url": "https://github.com/rsksmart/rootstock-integration-tests/pull/98"
}
|
gharchive/pull-request
|
Adds 'should reject a basic pegin v1 with value exactly below minimum…
Adds 'should reject a basic pegin v1 with value exactly below minimum' test.
This test sends a pegin v1 with a value exactly below minimum (the minimum pegin value minus 1 satoshis), to assert that it should be rejected with rejection reason 5 (invalid amount).
Asserts that the sender funds are lost and that the sender's derived rsk address nor the pegin v1 rsk recipient address get any funds, but the federation balance is increased by the pegin amount while the Bridge doesn't mark that pegin tx as processed.
https://github.com/rsksmart/rootstock-integration-tests/pull/108
|
2025-04-01T04:35:23.610263
| 2023-10-03T12:56:54
|
1924095814
|
{
"authors": [
"marcgarba",
"otterdahl"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10445",
"repo": "rsmp-nordic/rsmp_core",
"url": "https://github.com/rsmp-nordic/rsmp_core/issues/129"
}
|
gharchive/issue
|
Clarify 4.5.3. Unimplemented statuses or commands
This sentence can be written more clearly:
If a status (sCI) or command (cCI) is unimplemented, the site answers with CommandResponse/Response where q or age is set according to the table below. v should be set to null.
To understand better the subtlety compared with the other case of "unknown" status/command (SXL mismatch),
we could have this written instead :
If a status (sCI) or command (cCI) is recognized in relation to its SXL but not implemented the site answers with StatusResponseOrUpdate/CommandResponse where q or age is set according to the table below. v should be set to null.
For the 4.5.1 (unknown component object), we could have the same list "StatusResponseOrUpdate/CommandResponse" to be consistent.
Opened a PR here: https://github.com/rsmp-nordic/rsmp_core/pull/136
The PR has been merged. Closing this issue
|
2025-04-01T04:35:23.792024
| 2017-02-15T08:50:55
|
207741311
|
{
"authors": [
"FinScience",
"javierluraschi"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10446",
"repo": "rstudio/sparklyr",
"url": "https://github.com/rstudio/sparklyr/pull/492"
}
|
gharchive/pull-request
|
Normalize jar paths for windows under spark 1.6
Normalize jar paths for windows under spark 1.6 to fix #491 #490 #414 #410 #398 #305
Planning to backport this one to CRAN, since this is causing a bunch of issues. The problem is scoped to only Windows and Spark 1.6.x; however, a lot of users use this configuration since our documentation defaults to 1.6.2.
The fix is straightforward, we need to normalize the paths to jars we specify for the embedded CSV parser. Unfortunately, this affects almost any operation since other non-csv operations also cause Spark to enumerate jar dependencies which trigger an java.io.IOException: No FileSystem for scheme exception.
@jjallaire @kevinushey I'm planning to push this as a patch to CRAN and release as sparkly 0.5.2, see branch: https://github.com/rstudio/sparklyr/tree/bugfix/hotfix-0.5.2 for complete changes.
This problem makes 1.6.x unusable in Windows, a workaround in the config is possible but very inconvenient since all our docs are based on 1.6.
I've validated that normalizePath has no effect in OS X and that this correctly resolves the windows issue. I tested the fix in Windows with 2.0, 1.6 and OS X with 2.0 and 1.6 as well.
I'm also backporting a Livy fix which makes spark_connect fail intermittently when using Livy, also really safe fix.
@kevinushey if there is something very safe worth backporting, let me know.
@javierluraschi . With respect to issue #491 , the errors still persist. I used devtools::install_github("rstudio/sparklyr") to get the latest version. Please let me know in case anything else needs to be done.
@FinScience since this PR was not merged at the time, you could have used:
devtools::install_github("rstudio/sparklyr", ref = "bugfix/copy-to-windows-spark-1-6")
However, now that is merged you can use:
devtools::install_github("rstudio/sparklyr")
@javierluraschi I tried it now. For some reason, I am still getting the same errors.
sc <- spark_connect(master = "local")
Warning messages:
1: In value[3L] :
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522)
at org.apache.spark.sql.hive.client.ClientWrapper.(ClientWrapper.scala:204)
at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:238)
at org.apache.spark.sql.hive.HiveContext.executionHive$lzycompute(HiveContext.scala:218)
at org.apache.spark.sql.hive.HiveContext.executionHive(HiveContext.scala:208)
at org.apache.spark.sql.hive.HiveContext.setConf(HiveContext.scala:440)
at org.apache.spark.sql.SQLContext$$anonfun$4.apply(SQLContext.scala:272)
at org.apache.spark.sql.SQLContext$$anonfun$4.apply(SQLContext.scala:271)
at scala.collection.Iterator$class.foreach(Iterator.scala:727)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
at [... truncated]
2: In create_hive_context_v1(sc) :
Failed to create Hive context, falling back to SQL. Some operations, like window-functions, will not work
@FinScience I have a couple suggestions for you, please see: https://github.com/rstudio/sparklyr/issues/491 and reply on that thread to continue this investigation. Thank you!
|
2025-04-01T04:35:23.835826
| 2018-03-28T10:30:31
|
309309361
|
{
"authors": [
"GeekRishabh",
"morenoh149"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10447",
"repo": "rt2zz/react-native-contacts",
"url": "https://github.com/rt2zz/react-native-contacts/issues/278"
}
|
gharchive/issue
|
Docs: Using it with Pod
How to use this package when we are using Cocoapods. I am also using react-native-firebase . Please provide docs how to use it.
Sharing My Podfile
platform :ios, '8.0'
target 'test' do
# Uncomment the next line if you're using Swift or would like to use dynamic frameworks
# use_frameworks!
# Required by RNFirebase
pod 'Firebase/Core'
pod 'Firebase/Auth'
pod 'Firebase/Firestore'
pod 'Firebase/Database'
pod 'Firebase/Messaging'
pod 'RNVectorIcons', :path => '../node_modules/react-native-vector-icons'
pod 'react-native-contacts', :path => '../node_modules/react-native-contacts'
end
Have upgrade to platform :ios, '9.0' but this breaks the firebase linking modules
I don't use podfiles, perhaps borrow inspiration from https://github.com/msand/SVGPodTest/blob/master/ios/Podfile#L12
|
2025-04-01T04:35:23.845654
| 2021-01-07T05:20:40
|
781037038
|
{
"authors": [
"klorel",
"teamxSxoptimiz"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10448",
"repo": "rte-france/challenge-roadef-2020",
"url": "https://github.com/rte-france/challenge-roadef-2020/issues/25"
}
|
gharchive/issue
|
Semi-final phase
The semi final phase will end on March 8 2021 and we will be competing on set B of instances. Are we allowed/requested to send a new version of the program before that deadline?
Thanks
Dear competitors,
Please refer to the dedicated web page : https://www.roadef.org/challenge/2020/en/calendrier.php
New contribution are to be sent before March 16th 2021.
Great, thanks. What is the protocol for the submission? The same protocol used for the qualification phase but changing the subject to "ROADEF/EURO challenge 2020 semi final material from XX"?
|
2025-04-01T04:35:23.854807
| 2018-12-11T02:28:40
|
389570077
|
{
"authors": [
"Jacob-Stevens-Haas",
"SilleBille",
"humitos"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10449",
"repo": "rtfd/readthedocs.org",
"url": "https://github.com/rtfd/readthedocs.org/issues/4987"
}
|
gharchive/issue
|
Pandoc support
Details
Read the Docs project URL:
Build URL (if applicable):
Read the Docs username (if applicable):
Expected Result
This request may be turned down. But, giving my shot
I'm pretty new to readthedocs and I was impressed by the features. Thanks for this awesome project!
While researching, I came across pandoc which isn't supported by readthedocs (or at least, I don't find it in the tutorial).
I am trying to achieve the following:
A set of markdown files that:
a. is generated in the form of linux man pages
b. is generated into HTML
With the help of pandoc, I'm able to achieve the above 2. However, I am not able to host it on readthedocs. Is there a possiblity to add support for pandoc?
Actual Result
No way to build with pandoc and deploy to readthedocs
Hi @SilleBille! Read the Docs currently achieve 2 -- generate HTML from Markdown.
Also, Sphinx -- the builder that RTD uses -- has the ability to generate man pages but we are not doing that currently. Although, there is an issue already open for that #4458
Would it be enough for you to enable man page generation for Sphinx builder?
@humitos thanks for the reply! To summarize my needs with a simplified diagram:
javadoc is generated using the official javadoc package
man pages are currently processed using pandoc
Documentation can also be converted to HTML by pandoc. Using sphinx/mkdocs for just this workflow makes the pipeline complicated.
I want to keep the pipeline as simplified as possible. Though sphinx supports all the above, as per your previous comment, it seems rtfd doesn't provide support for man pages?
Thanks for the explanation.
I think this is a very fair feature request considering that we support sphinx and mkdocs. Why not pandoc also?
I don't have too much experience with pandoc (I've only used it to convert files between formats and it's great) and I don't know how much complicated it would be, but I'm sure that it's not an easy feature to add to our platform. On the other hand, considering our current roadmap, I don't think that this will happen soon, unfortunately.
I marked this issue as Needed: design decision so other core developers can write their thoughts here.
To answer your question, though sphinx supports all the above, as per your previous comment, it seems rtfd doesn't provide support for man pages? Yes, I can switch over to sphinx if man pages can be generated in both roff and HTML formats
Currently, we are not generating man pages. So, I'd say that it's not supported. The discussion for this will happen in #4458
@humitos Thanks for an optimistic reply. The main concern for me is that I don't want my team to follow different formats to write down the documentation. I want them to follow consistent formatting across all the docs they write.
I'm going to close this issue since it's not going to happen soon, unfortunately. We can revisit later if we consider valuable to add another builder. Thanks!
Adding a link here because I found this issue before the relevant, older one, but it looks like readthedocs has pandoc support already: https://github.com/readthedocs/readthedocs.org/issues/579#issuecomment-31547081
|
2025-04-01T04:35:23.857670
| 2018-09-05T22:55:31
|
357433159
|
{
"authors": [
"agjohnson",
"ericholscher",
"humitos"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10450",
"repo": "rtfd/readthedocs.org",
"url": "https://github.com/rtfd/readthedocs.org/pull/4608"
}
|
gharchive/pull-request
|
Add a contrib Dockerfile for local build image on Linux
This is necessary as permissions are all incorrect on the paths that are
shared between the host system and the Docker build container.
Closes #2692
Anything we can do to move this along? Seems like a useful improvement.
I rebased this branch and push a new commit to always use latest to build the image and also guide the user to re-tag the Docker image after building it so Read the Docs uses it without changing any setting that could break in other places.
I've updated this pr to use a configurable label in the docker script. The last piece is how to handle DOCKER_IMAGE_SETTINGS, and I think the answer is maybe alter this setting conditionally to point to the dev images.
I added a quick hack to settings.dev to allow for manual triggering of the docker image name replacement. One last review and I think this is ready.
Let's see what new and interesting problems this will likely create for us eventually! :crossed_fingers:
|
2025-04-01T04:35:23.860919
| 2015-12-05T01:32:35
|
120519150
|
{
"authors": [
"agjohnson",
"danroth27"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10451",
"repo": "rtfd/sphinx-autoapi",
"url": "https://github.com/rtfd/sphinx-autoapi/issues/45"
}
|
gharchive/issue
|
Default output folder for docfx changed to _site
Looks like the .NET mapper assumes that the default output folder will be _api_, but it recently go changed to _site. The .NET mapper should probably explicitly specify the output folder when running docfx to make sure it can find the output metadata.
This can be moved to a mapper-specific configuration option, though we should also specify the path with the docfx option -o or --output.
It appears the -o option only prepends to the output path, it never overrides. Forcing -o _api creates an output path -o _api/_api. If there is a dest setting of 'docs/apiin thedocfx.jsonfile, forcing-o _apicreates an output path of_api/docs/api`. So, seems we can't use this to set a path quite yet.
Not sure if this is a docfx-dev vs docfx mainline issue, but I get the output path of _api. Addressing this in a PR, if the path does change to _site we can update again.
|
2025-04-01T04:35:23.871774
| 2021-08-25T01:39:47
|
978619249
|
{
"authors": [
"korken89",
"pdgilbert"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10452",
"repo": "rtic-rs/cortex-m-rtic",
"url": "https://github.com/rtic-rs/cortex-m-rtic/issues/517"
}
|
gharchive/issue
|
trait implementation fails on task shared object
(Using 0.6-alpha from current github #22ec841.) I am trying to implement a trait but the implementation is not recognized for the shared object. The following code compiles with the cx.shared.led.on() line commented out
Click to expand
#![deny(unsafe_code)]
#![no_std]
#![no_main]
#[cfg(debug_assertions)]
use panic_semihosting as _;
#[cfg(not(debug_assertions))]
use panic_halt as _;
use rtic::app;
#[cfg_attr(feature = "stm32l4xx", app(device = stm32l4xx_hal::pac, peripherals = true, dispatchers = [USART1, USART2]))]
mod app {
use dwt_systick_monotonic::DwtSystick;
use rtic::time::duration::{Seconds, Milliseconds};
const PULSE: u32 = 1; // 1 second
const PERIOD: u32 = 10 * PULSE; // 10 seconds
use stm32l4xx_hal::{
gpio::{gpioc::PC13, Output, PushPull},
pac::Peripherals,
prelude::*,
};
use embedded_hal::digital::v2::OutputPin;
type LedType = PC13<Output<PushPull>>;
fn setup(dp: Peripherals) -> LedType {
let mut rcc = dp.RCC.constrain();
let mut gpioc = dp.GPIOC.split(&mut rcc.ahb2);
let led = gpioc.pc13.into_push_pull_output(&mut gpioc.moder, &mut gpioc.otyper);
impl LED for PC13<Output<PushPull>> {
fn on(&mut self) -> () { self.set_low().unwrap() }
fn off(&mut self) -> () { self.set_high().unwrap()}
}
led
}
pub trait LED {
fn on(&mut self) -> ();
fn off(&mut self) -> ();
}
#[monotonic(binds = SysTick, default = true)]
type DwtMono = DwtSystick<80_000_000>;
#[local]
struct Local {}
#[init]
fn init(cx: init::Context) -> (Shared, Local, init::Monotonics) {
let mut dcb = cx.core.DCB;
let dwt = cx.core.DWT;
let systick = cx.core.SYST;
let mono = DwtSystick::new(&mut dcb, dwt, systick, 8_000_000);
let device = cx.device;
let mut led = setup(device);
led.off(); //WORKS HERE
led.on();
led.off();
crate::app::one::spawn_after(Milliseconds(PULSE * 1000 - 700)).ok();
(Shared {led}, Local {}, init::Monotonics(mono))
}
#[shared]
struct Shared {
led: LedType,
}
#[task(shared = [led])]
fn one(cx: one::Context) {
//cx.shared.led.on(); //FAILS HERE compiles with this line commented out
crate::app::one::spawn_after(Seconds(PULSE)).ok();
}
}
but with the line uncommented I get
3 | #[cfg_attr(feature = "stm32l4xx", app(device = stm32l4xx_hal::pac, peripherals = true, dispatchers = [USART1, USART2]))]
| -------------------------------------------------------------------------------------- method `on` not found for this
...
85 | cx.shared.led.on(); //FAILS HERE compiles with this line commented out
| ^^ method not found in `led<'_>`
|
= help: items from traits can only be used if the trait is implemented and in scope
note: `LED` defines an item `on`, perhaps you need to implement it
--> examples/zzled.rs:46:5
|
46 | pub trait LED {
| ^^^^^^^^^^^^^
The implementation works in the #[init] section, it is only in the #[task(shared = [led])] section that I get the error.
I've been testing mostly with stm32l4xx but it also occurs with stm32f1xx.
Hi,
You need to lock the led fits.
A more complete working and run tested example, from which the above was extracted, is now at #490.
|
2025-04-01T04:35:23.892863
| 2016-12-07T21:50:09
|
194180657
|
{
"authors": [
"IljaDaderko",
"Stringsaeed",
"ads1018",
"digitalmaster",
"faddee",
"jachiike-madubuko",
"louisgv",
"rtsao",
"tajo"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10453",
"repo": "rtsao/styletron",
"url": "https://github.com/rtsao/styletron/issues/26"
}
|
gharchive/issue
|
React Native Support?
The state of managing css in React-Native is still kinda clunky. This would be a huge improvement..
Do you know if it's supported?
As someone who hasn't actually had a chance to use React Native myself, I'm interested to learn more about which parts of Styletron could help! I don't know much about it, to be honest!
In my mind, the coolest thing about Styletron is the under-the-hood mapping of declarations to atomic classes, with the result of producing tiny stylesheets for server-rendering for browsers -- I wouldn't be able to say how that transfers into the context of React Native.
React Native support would be awesome! I know styled-components support its -> https://github.com/styled-components/styled-components#react-native
Haven't got a chance to work with React Native so I'm not that familiar of what's happening under the hood but shouldn't this work?
function styled(Base, style) {
return (props) => (
<Base
{...props}
style={{
...(typeof style === 'function' ? style(props) : style),
...props.style
}}
/>
);
}
With of course some minor tweaks so you're able to compose styled components.
Any updates on react-native support / implementation examples?
With the new v4 architecture, all that needs to be done here is a rendering engine for react native (which uses RN-shaped style objects). There's no assumptions about atomic rendering or even the shape of style objects.
Is this issue resolved? If not I can take a whack at it this week
Is this issue resolved?
It's not. We don't use RN.
If not I can take a whack at it this week
That would be great!
Very excited to see this extension of styletron. Is there any news for 2020 on progress. Thanks for being badass!
@jachiike-madubuko We still don't have any plans for RN engine since we don't use RN.
anyone work on this or should work on it?
@Stringsaeed No. The status is unchanged.
|
2025-04-01T04:35:23.906979
| 2024-04-03T18:15:04
|
2223603297
|
{
"authors": [
"Blad3forc3",
"ruanformigoni"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10454",
"repo": "ruanformigoni/gameimage",
"url": "https://github.com/ruanformigoni/gameimage/issues/52"
}
|
gharchive/issue
|
Wineprefix not created error
When I get to the configure wine options during setup (install dxvk etc) i get this error:
Writing image to "/media/ade/Temp/Temp/build/Age-Of-Empires/icon/icon.png"
I :: Writing grayscale image to "/media/ade/Temp/Temp/build/Age-Of-Empires/icon/icon.grayscale.png"
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/Age-Of-Empires/gameimage.json"' as UPDATE
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ
I :: application: Age-Of-Empires
I :: image: "/media/ade/Temp/Temp/build/wine.flatimage"
I :: prefix: "/media/ade/Temp/Temp/build/Age-Of-Empires/wine"
I :: [e] :: bwrap: Can't create file at /tmp/fim/instance/502568f_20240217183447/mount/KIdoON.mount.dwarfs.usr//lib/libnvidia-gtk2.so.510.47.03: Function not implemented
I :: [e] :: Killed
I :: Finished Command: '/media/ade/Temp/Temp/build/wine.flatimage fim-exec wine.sh regedit'
Wine prefix does not exist
Thanks for the report @Blad3forc3 ! I found the issue and fixed it, it was on the wine runner when used with nvidia, erase the build folder and re-download the wine packages. That should do it :smile:
Fixed in latest wine runner, thanks! The issue is that the HOME was set I :: [o] :: [wine.sh] HOME : /home/ade, and it was building from an external drive.
Just resolved other issues with winetricks, please wait for this pipeline to finish.
I'll assume this is fixed, feel free to re-open if it is not. :smile:
This was fixed yes :)
|
2025-04-01T04:35:23.914896
| 2020-02-07T11:09:21
|
561572254
|
{
"authors": [
"michae1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10455",
"repo": "rubicon-project/prebid-server-java",
"url": "https://github.com/rubicon-project/prebid-server-java/pull/612"
}
|
gharchive/pull-request
|
Add influxDb tags
Sometimes influxDb tags can come in handy to split data by region, hostname etc.
I find there is no tagging in current prebid-server-java.
As dropwizard-metrics-influxdb already has tags support, I suggest this simple implementation.
Fixed empty tags exception
|
2025-04-01T04:35:23.920061
| 2018-07-04T03:45:37
|
338120737
|
{
"authors": [
"composerinteralia",
"koic"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10456",
"repo": "rubocop-hq/rubocop-performance",
"url": "https://github.com/rubocop-hq/rubocop-performance/issues/11"
}
|
gharchive/issue
|
Add RuboCop rake task
And run RuboCop in CI (see #10)
I'm preparing for minimal rake tasks now. I will prepare #27 and some rake tasks.
Basic rake tasks have been prepared with #27, #28, and #30. I will close the issue.
|
2025-04-01T04:35:23.922547
| 2018-07-02T18:39:28
|
337622162
|
{
"authors": [
"egze",
"jfelchner"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10457",
"repo": "rubocop-hq/rubocop",
"url": "https://github.com/rubocop-hq/rubocop/issues/6069"
}
|
gharchive/issue
|
ForceEqualSignAlignment should only align block content
The current alignment makes it not obvious what variables are set inside and outside the block. I think rubocop should align only what is inside the block.
Here is an example:
my_secret_values = [1,2,3,4].map do |i|
x = i * 2
secret = x*x
end
Expected behavior
my_secret_values = [1, 2, 3, 4].map do |i|
x = i * 2
secret = x * x
end
Actual behavior
my_secret_values = [1, 2, 3, 4].map do |i|
x = i * 2
secret = x * x
end
Steps to reproduce the problem
Run rubocop with autocorrect.
RuboCop version
Include the output of rubocop -V. Here's an example:
$ rubocop -V
0.54.0 (using Parser <IP_ADDRESS>, running on ruby 2.4.2 x86_64-darwin17)
This should be fixed in #7211
|
2025-04-01T04:35:23.924434
| 2019-10-03T13:16:33
|
502060316
|
{
"authors": [
"bbatsov",
"mvz",
"scottmatthewman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10458",
"repo": "rubocop-hq/rubocop",
"url": "https://github.com/rubocop-hq/rubocop/pull/7394"
}
|
gharchive/pull-request
|
[Fix #6759] Reference configuration in Basic Usage
Within the Basic Usage page, make a brief mention of the use of .rubocop.yml to change RuboCop's behaviour. The additional content links to the Configuration page to go into more detail.
Fixes #6759.
Looks good. Thanks!
Thanks, @scottmatthewman!
|
2025-04-01T04:35:23.925681
| 2020-08-10T18:39:16
|
676336307
|
{
"authors": [
"fatkodima",
"marcandre"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10459",
"repo": "rubocop-hq/rubocop",
"url": "https://github.com/rubocop-hq/rubocop/pull/8511"
}
|
gharchive/pull-request
|
Fix a false positive for Style/CaseLikeIf when conditional contains comparison with a class
Closes #8508
Marked it as unsafe.
I'll merge after the merge conflict is resolved... 😢
Thanks!
|
2025-04-01T04:35:23.932296
| 2024-02-22T05:53:21
|
2148264711
|
{
"authors": [
"TJNII",
"jonas054"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10460",
"repo": "rubocop/rubocop",
"url": "https://github.com/rubocop/rubocop/issues/12706"
}
|
gharchive/issue
|
Add ability to globally disable caching
Is your feature request related to a problem? Please describe.
While backing up my system today I noticed Rubocop has begun writing into ~/.cache. I find this behavior undesirable as I try to keep my code and it's dependencies tightly contained in working directories, it's never pleasant to find when a tool escapes where I put it and begins writing to other paths on my filesystem. I found the documentation at https://docs.rubocop.org/rubocop/usage/caching.html, but I was disappointed to see that this option, which writes outside the per-project tree, is currently only settable in per-project config. I'm not looking forward to having to remember to turn this off on every project I work on.
Describe the solution you'd like
As Rubocop has begun writing outside the bounds of the project tree, please add a config file that reads from outside the project tree (Say, ~/.config/rubocop.yml) where environmental options like this can be globally specified. For this case I'm not seeking any deep configs so a simple read and merge should be sufficient.
Describe alternatives you've considered
Per https://docs.rubocop.org/rubocop/usage/caching.html there aren't any. chmod 000 ~/.cache/rubocop_cache I guess?
Per https://docs.rubocop.org/rubocop/configuration.html reading from ~/.rubocop.yml is supported, but if $(pwd)/.rubocop.yml exists it will not be loaded. The behavior added in https://github.com/rubocop/rubocop/pull/8314 appears to be at odds with with the behavior I want.
I maintain my assertion that if Rubocop is going to write outside the project tree, then it must accept config from outside the project tree as well. Perhaps an ENV var? Per https://github.com/rubocop/rubocop/blob/master/lib/rubocop/cache_config.rb#L8 setting RUBOCOP_CACHE_ROOT to an empty string will not disable it.
What about the --cache false command line option? Have you considered that?
I have, but the spirit of the request is for when I forget about this behavior and don't remember I need to add a flag or tune the project config file. I'm looking for something I can set in my home directory or bashrc to configure Rubocop to stay within the project tree. The problem with a command line argument is I often run Rubocop from Make or bundler, and bundler (apparently) doesn't use bash aliases so I can't alias rubocop="rubocop --cache-false" in my .bashrc.
I actually just realized that there is a simple, supported solution to this: My grouse isn't with the caching itself, it's with the default cache directory being outside the project tree. RUBOCOP_CACHE_ROOT supports relative paths. RUBOCOP_CACHE_ROOT='./tmp/rubocop/cache' works.
As this caching feature has been around for a while and nobody else is asking for a global disable option I think we can close this. I found a solution to my concern.
And for the record chmod 000 ~/.cache/rubocop_cache does work, but that's only good for making sure ~/.cache/rubocop_cache is never written and reminding that the project config needs updated via the error. It's heavy handed, but workable.
|
2025-04-01T04:35:23.937781
| 2023-01-03T05:03:30
|
1516933340
|
{
"authors": [
"koic",
"sambostock"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10461",
"repo": "rubocop/rubocop",
"url": "https://github.com/rubocop/rubocop/pull/11378"
}
|
gharchive/pull-request
|
Fix OperatorMethodCall when forwarding arguments
... cannot be used as an argument to an operator, so it should not be an offense. For example,
def m(...) = obj.==(...)
would be autocorrected as
def m(...) = obj ==(...)
which is a syntax error, as would be
def m(...) = obj == (...)
and
def m(...) = obj == ...
Before submitting the PR make sure the following are checked:
[x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences.
[x] Wrote good commit messages.
[x] Commit message starts with [Fix #issue-number] (if the related issue exists).
[x] Feature branch is up-to-date with master (if not - rebase it).
[x] Squashed related commits together.
[x] Added tests.
[x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code.
[x] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details.
Closes #11377
Thanks!
|
2025-04-01T04:35:23.958215
| 2023-08-25T18:14:09
|
1867498391
|
{
"authors": [
"jcouball",
"srem1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10462",
"repo": "ruby-git/ruby-git",
"url": "https://github.com/ruby-git/ruby-git/issues/666"
}
|
gharchive/issue
|
--allow-empty
how add oprions --allow-empty to commit
or --allow-empty-message
# assuming your current directory is in a git working directory
g = Git.open('.')
message = ''
g.commit(message, allow_empty: true, allow_empty_message: true)
why not g.config() ?
this is not commit, I need setup for whole repo
This library does not provide that capability.
|
2025-04-01T04:35:24.010198
| 2023-09-13T11:43:14
|
1894349442
|
{
"authors": [
"eregon",
"seven1m"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10463",
"repo": "ruby/yarp",
"url": "https://github.com/ruby/yarp/issues/1461"
}
|
gharchive/issue
|
How to differentiate WhileNode in modifier form
What is the way to determine if a WhileNode was in modifier form?
irb(main):001:0> n = YARP.parse('foo while bar').value.statements.body.first
=>
@ WhileNode (location: (0...13))
...
irb(main):002:0> ls n
YARP::WhileNode#methods: accept begin_modifier? child_nodes closing closing_loc comment_targets copy deconstruct deconstruct_keys inspect keyword keyword_loc predicate set_newline_flag statements
YARP::Node#methods: location newline? pretty_print slice
instance variables: @closing_loc @flags @keyword_loc @location @predicate @statements
=> nil
irb(main):003:0> n.closing
=> nil
The only difference I can see is that WhileNode#closing is nil (vs "end"), but that doesn't seem appropriate to use. Should there be (or is there already) a method to determine if the modifier form was used?
I suppose the same question goes for UntilNode.
Via LoopFlags
Ah sorry I thought you meant checking whether it's a do/while vs a while loop (there is LoopFlags.BEGIN_MODIFIER for that).
$ bin/parse -e 'foo while bar'
@ ProgramNode (location: (0...13))
├── locals: []
└── statements:
@ StatementsNode (location: (0...13))
└── body: (length: 1)
└── @ WhileNode (location: (0...13))
├── keyword_loc: (4...9) = "while"
├── closing_loc: ∅
├── predicate:
│ @ CallNode (location: (10...13))
│ ├── receiver: ∅
│ ├── call_operator_loc: ∅
│ ├── message_loc: (10...13) = "bar"
│ ├── opening_loc: ∅
│ ├── arguments: ∅
│ ├── closing_loc: ∅
│ ├── block: ∅
│ ├── flags: variable_call
│ └── name: "bar"
├── statements:
│ @ StatementsNode (location: (0...3))
│ └── body: (length: 1)
│ └── @ CallNode (location: (0...3))
│ ├── receiver: ∅
│ ├── call_operator_loc: ∅
│ ├── message_loc: (0...3) = "foo"
│ ├── opening_loc: ∅
│ ├── arguments: ∅
│ ├── closing_loc: ∅
│ ├── block: ∅
│ ├── flags: variable_call
│ └── name: "foo"
└── flags:
I think you could check via while_node.statements.location.start_offset < while_node.predicate.location.start_offset.
And the same for UntilNode and also IfNode and UnlessNode.
Note that semantically there is 0 difference between foo while bar and while bar; foo; end.
Yes, I was confused myself. My question actually should have been: how do you know if it's a begin while vs a regular while. Yep, begin_modifier? is actually what I need. Thank you @eregon! ❤️
|
2025-04-01T04:35:24.019100
| 2022-09-09T20:32:07
|
1368293949
|
{
"authors": [
"cielf",
"dorner",
"edwinthinks"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10464",
"repo": "rubyforgood/human-essentials",
"url": "https://github.com/rubyforgood/human-essentials/pull/3143"
}
|
gharchive/pull-request
|
3088 distribution pdf
Resolves #3088
Description
This updates the output of the distribution PDF when there is an associated request.
I've done a minor refactor of the LineItem class to accommodate the requested changes.
Please see screenshots. The "request" version of it is missing a border on the bottom right of the data table - I couldn't figure out how to put it back and I ran out of time. We can decide if that's a blocker or if we can merge as is.
Type of change
New feature (non-breaking change which adds functionality)
How Has This Been Tested?
Local/unit tests
Screenshots
Thanks @dorner -- I probably won't get to this until at least Wednesday, because life.
Still pending some fixes and after that I think we mergey merge
Ack, forgot to push my commit! 😮
@edwinthinks I think this should be good now.
@edwinthinks thanks for the QA - it should be fixed now!
Awesome! Thanks @dorner just QA'ed this and it works.
A bit of a side thought for a future issue, perhaps we can use something like https://adminlte.io/themes/v3/pages/examples/invoice.html from the theme which allows us to consolidate the printing and the show page of distributions. I figure this data could be helpful outside of the PDF. Lemme know what you think.
That looks interesting! Definitely worth looking into next time we do something like this.
|
2025-04-01T04:35:24.046625
| 2017-07-02T15:06:06
|
240015532
|
{
"authors": [
"tyranja",
"zaziemo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10465",
"repo": "rubymonsters/speakerinnen_liste",
"url": "https://github.com/rubymonsters/speakerinnen_liste/pull/588"
}
|
gharchive/pull-request
|
Update all the gems
https://github.com/rubymonsters/speakerinnen_liste/issues/470
Too many gems :(
Travis checks are failing. You have to update the ruby version in the travis.yml as well I think.
I am not yet ready, just half way through.
|
2025-04-01T04:35:24.071138
| 2023-03-02T12:50:07
|
1606752734
|
{
"authors": [
"maany"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10466",
"repo": "rucio/webui",
"url": "https://github.com/rucio/webui/issues/134"
}
|
gharchive/issue
|
Fix storybook deployment
Description
The deploy-to-gh-pages script is broken since #117. It expects the storybook build to be stored in a docs directory, which we do not do anymore.
The solution would be to modify the package.json to build storybook into the docs directory. The would result in successful deployments of storybook to rucio.cern.ch/webui
Steps to reproduce
No response
Rucio WebUI Version
No response
Additional Information
No response
@ThePhisch could you work on this while I am away?
Already fixed by @ThePhisch
|
2025-04-01T04:35:24.124191
| 2014-12-16T11:17:29
|
52100770
|
{
"authors": [
"andrewchambers",
"nickzuck"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10467",
"repo": "rui314/8cc",
"url": "https://github.com/rui314/8cc/issues/36"
}
|
gharchive/issue
|
segfault
Put this in test.c
long long a = 0LLU;
run 8cc -S ./test.c.
-> segfault.
Hey there can you please give the link to the file of which you are talking about
What file? the contents of the file are what I showed. Simply that single line.
"What file? the contents of /*the file */ are what I showed. Simply that single line."
According to your answer ...the file in the /* */ ......The file which you are talking about
|
2025-04-01T04:35:24.128680
| 2021-01-20T14:22:37
|
790021317
|
{
"authors": [
"ruimarinho",
"torkelrogstad"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10468",
"repo": "ruimarinho/docker-bitcoin-core",
"url": "https://github.com/ruimarinho/docker-bitcoin-core/pull/108"
}
|
gharchive/pull-request
|
CI: remove deprecated set-env, use build outputs
set-env was removed a while ago from GitHub actions, because it was a
securit risk. We replace this with the recommended workaround to get CI
working again.
@ruimarinho a bit unsure of what's going on here. Is the docker password empty because I'm submitting a PR from a fork, or because you've removed it?
@ruimarinho a bit unsure of what's going on here. Is the docker password empty because I'm submitting a PR from a fork, or because you've removed it?
Hmm, I see DOCKER_HUB_PASSWORD is correctly set so I need to look at what's going on.
It might if you open a PR from this repo, with the same changes, I know that there are some restrictions on reading secrets from forks
It might if you open a PR from this repo, with the same changes, I know that there are some restrictions on reading secrets from forks
Thank you @torkelrogstad 🙏 fixed in https://github.com/ruimarinho/docker-bitcoin-core/pull/112.
|
2025-04-01T04:35:24.136677
| 2024-10-01T13:08:59
|
2559181366
|
{
"authors": [
"LawrenceGB",
"marcusschiesser"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10469",
"repo": "run-llama/LlamaIndexTS",
"url": "https://github.com/run-llama/LlamaIndexTS/pull/1282"
}
|
gharchive/pull-request
|
fix: ensure id_ is correctly passed during creation
Fixes: #1278
This PR addresses issue #1278 where the id_ field was not being correctly passed as the id when creating a TextNode. This caused the upsert operation to the vector database to rely on the idGenerator instead of using the provided document ID, if available.
Changes:
Updated the logic for creating TextNode to ensure the id_ is set as the document ID when available.
This ensures that when upserting to the vector database, the correct document ID is used, avoiding the generation of unnecessary random IDs.
This fix improves consistency in ID handling, ensuring that document IDs are properly retained in the vector database.
@LawrenceGB and @himself65, this PR was causing us a little headache:
A document is parsed into several nodes. This PR causes all these nodes to have the same ID as the parent. So we get duplicate IDs, and these IDs are no longer UUIDs (which is required by some vector DBs).
I, therefore, reverted this PR in https://github.com/run-llama/LlamaIndexTS/pull/1311
To get the ID of the source document from a split node, please check node.relationships[NodeRelationship.SOURCE]
@marcusschiesser Could you please help clarify how I can get the actual ID that is being set as I can't seem to get it working.
I am creating a unique document like so, it's not using that ID but generates the random one when upserting to Pinecone and also the doc returned relationships are empty.
new Document({
id_: `${fileId}-${index}`,
text: page.text,
metadata: { page: page.index }
})
Thanks!
|
2025-04-01T04:35:24.139092
| 2024-05-16T07:11:20
|
2299542010
|
{
"authors": [
"monoboard1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10470",
"repo": "run-llama/llama_index",
"url": "https://github.com/run-llama/llama_index/issues/13530"
}
|
gharchive/issue
|
[Question]: How can I create a tool that enables querying a specific pandas DataFrame and then using the query result to perform calculations through a custom function?
Question Validation
[X] I have searched both the documentation and discord for an answer.
Question
I need to develop a tool that allows me to query a specific pandas DataFrame and then apply a custom function to the query result. How can I achieve this functionality in an efficient manner?
@dosu thanks but could you please give me an example how to build the agent that first query pandas dataframe with specific LLM using the "pandas_query_engine" and then use the query result into LLM custom calling function?
@dosu Ok now please give me an example of FunctionTool that can query data from pandas dataframe and then use the query result in a custom function
|
2025-04-01T04:35:24.147436
| 2024-01-02T20:16:31
|
2062896571
|
{
"authors": [
"ji21"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10471",
"repo": "run-llama/llama_index",
"url": "https://github.com/run-llama/llama_index/pull/9801"
}
|
gharchive/pull-request
|
Integrations/deepeval observability
Description
Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.
Fixes #9800
Requires deepeval (v0.20.43)
Type of Change
Please delete options that are not relevant.
[x] New feature (non-breaking change which adds functionality)
[x] This change requires a documentation update
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Provide instructions so we can reproduce. Please also list any relevant details for your test configuration
[x] Added new notebook (that tests end-to-end) (on DeepEval repo: https://github.com/confident-ai/deepeval)
Suggested Checklist:
[x] I have performed a self-review of my own code
[x] I have commented my code, particularly in hard-to-understand areas
[x] I have made corresponding changes to the documentation
[ ] I have added Google Colab support for the newly added notebooks.
[x] My changes generate no new warnings
[x] I have added tests that prove my fix is effective or that my feature works
[x] New and existing unit tests pass locally with my changes
[x] I ran make format; make lint to appease the lint gods
@logan-markewich thanks for the prompt review! happy new year
|
2025-04-01T04:35:24.150270
| 2017-12-28T22:44:35
|
285003059
|
{
"authors": [
"xiaods"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10472",
"repo": "runconduit/conduit",
"url": "https://github.com/runconduit/conduit/issues/98"
}
|
gharchive/issue
|
cargo test came across failed.
$ cargo test --all
Compiling abstract-ns v0.4.2
Compiling env_logger v0.4.3
error[E0277]: the trait bound `std::sync::Arc<str>: std::convert::From<&str>` is not satisfied
--> /Users/xiaods/.cargo/registry/src/github.com-1ecc6299db9ec823/abstract-ns-0.4.2/src/name.rs:67:23
|
67 | Ok(Name(value.into()))
| ^^^^ the trait `std::convert::From<&str>` is not implemented for `std::sync::Arc<str>`
|
= help: the following implementations were found:
<std::sync::Arc<T> as std::convert::From<T>>
= note: required because of the requirements on the impl of `std::convert::Into<std::sync::Arc<str>>` for `&str`
error: aborting due to previous error
error: Could not compile `abstract-ns`.
warning: build failed, waiting for other jobs to finish...
error: build failed
$ rustc -V
rustc 1.20.0 (f3d6973f4 2017-08-27)
the rust version should be 1.21+
|
2025-04-01T04:35:24.222508
| 2024-04-26T06:26:16
|
2265051362
|
{
"authors": [
"Ian-cmd-ops",
"runeharlyk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10473",
"repo": "runeharlyk/SpotMicroESP32-Leika",
"url": "https://github.com/runeharlyk/SpotMicroESP32-Leika/issues/37"
}
|
gharchive/issue
|
ESP32 Cam not working
I can't get the my ESP32 cam to stream or view after I got my ESP32 Uploaded with via PIO.
do I need the other modules attached first before it runs
#define USE_PSRAM true
#define USE_WIFI true
#define WAIT_FOR_WIFI false
#define USE_WEBSERVER true
#define USE_WEBSERVER_SSL false
#define USE_WEBSOCKET true
#define USE_OAT false
#define USE_NTP false
#define USE_MDNS true
#define USE_DNS_SERVER false
#define USE_REMOTE_SERIAL false
#define USE_LOW_POWER false
#define USE_CAMERA true
#define USE_MPU true
#define USE_POWER_BUTTON true
#define USE_USS true
The camera stream should work out of the box. Is there anything in the browser console log?
Bad ESP32 Cam
|
2025-04-01T04:35:24.224624
| 2024-06-04T00:08:46
|
2332272166
|
{
"authors": [
"Ben10164",
"Unmoon",
"aHooder",
"capslock13"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10474",
"repo": "runelite/plugin-hub",
"url": "https://github.com/runelite/plugin-hub/pull/6125"
}
|
gharchive/pull-request
|
Update advanced-raid-tracker
Add TOA, Inferno, Colosseum
Improved UI using icons and handling the new raids
Various chart improvements and expansion
adds Chart Creator
Apologies for the size of this update; any future changes will not be nearly this large or as comprehensive in the future. No expectations of this being reviewed anytime soon.
Are you sure you need this many changes? I don't have time to review a 20k+ line diff.
Closing & re-opening to re-run checks.
Any updates on the review process?
|
2025-04-01T04:35:24.225819
| 2020-03-28T14:25:59
|
589587143
|
{
"authors": [
"Pingvinszar2",
"raiyni"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10475",
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/11117"
}
|
gharchive/issue
|
Miscellaneous: Inventory/Bank Slot binder.
Self-explanatory.
Locks your items/objects in your inventory/bank, or other interfaces.
The option will prevent you from accidentally dragging/switching items while doing click-intensive tasks like stringing bows, crafting jewelleries, making arrows, etc.
Anti drag works in bank now.
|
2025-04-01T04:35:24.228360
| 2022-08-03T14:18:44
|
1327294686
|
{
"authors": [
"Adam-",
"ghostoo666"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10476",
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/15436"
}
|
gharchive/issue
|
Max hit hitsplat not interacting with Special Attack counter plugin
Type
Incorrect behavior
Operating System
Windows
OS Version
Win10
Bug description
The special attack counter records either number of successful hits per weapon (dwh) or total damage in special attacks per weapon (bgs). With the new max hit hitsplats, if you hit your max hit during the special attack, it won't be recorded to the counters.
Screenshots or videos
No response
RuneLite version
No response
Logs
No response
fixed in d81463e4ad60eceab47470519ac49c91eea53139
|
2025-04-01T04:35:24.230320
| 2018-07-30T09:08:26
|
345672580
|
{
"authors": [
"generiskk",
"memementor"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10477",
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/4563"
}
|
gharchive/issue
|
Achievement diary typo
In the hard lumbridge/draynor diary...
says i need 68 farming but really i know i need 63
Since its under the same, I will add this too.
|
2025-04-01T04:35:24.233415
| 2019-03-21T13:14:45
|
423723001
|
{
"authors": [
"Abextm",
"deathbeam",
"yyruneliter"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10478",
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/issues/8278"
}
|
gharchive/issue
|
Camera Zoom Settings Not Working
Outer zoom limit, inner zoom limit, require control down in "Camera Zoom" settings do not work after Runescape update for March 21, 2019. Seems to be at default zoom levels. Vertical camera is functioning properly.
Steps to reproduce the behavior:
Go to 'Runelite'
Login
Click on 'Configuration'
Scroll down to 'Camera Zoom'
Click on 'Edit Plugin Configuration'
See that settings do not work when changed (except 'Vertical camera')
OS: Windows 10
RuneLite version: 1.5.16
Launcher version: 1.6
This was fixed in 776d27cc4772d9ad3c779a2799ecef9a12aa0904
Restart your client
Restart your client
|
2025-04-01T04:35:24.246204
| 2020-12-29T01:39:48
|
775654028
|
{
"authors": [
"Adam-",
"Nightfirecat",
"chaticon"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10479",
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/pull/12990"
}
|
gharchive/pull-request
|
opponentinfo: make health bar more active
Made the health bar continue to display while the opponent attacks the player. Useful for keeping track of enemy hp while recoiling.
This looks like it would break whenever you are fighting something in a multi combat area and then had something else attack you.
This looks like it would break whenever you are fighting something in a multi combat area and then had something else attack you.
My bad, I had tested it in multi but clearly not well enough. Should correctly prioritise the NPC you're attacking now.
The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent?
This change very much would benefit from some tests being added, given its complexity.
The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent?
This change very much would benefit from some tests being added, given its complexity.
The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent?
This change very much would benefit from some tests being added, given its complexity.
I added some tests to clarify behaviour in some ambiguous scenarios.
To address the specific scenarios you brought up, you were partially right that I goofed the second scenario. It wouldn't start tracking the hellhound if it wasn't already active opponent, but if it was, it would stay the active opponent until that NPC died or some other NPC became active opponent. I've fixed that now.
As for the first scenario you mentioned, I don't think it's as weird as you say. If the player is attacking an NPC, then that NPC will always be the active opponent regardless of what happens in multi. The only time the most recent attacker will become the active opponent is if the player is not attacking any NPC at all. That might sound odd, but if you're, say, cannoning while idle, that NPC is probably the most recent enemy your cannon hit, in which case it's behaving similarly to the current version of the plugin.
Regardless, I hope the tests are to your liking and can help illustrate my explanation. Cheers.
The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent?
This change very much would benefit from some tests being added, given its complexity.
I added some tests to clarify behaviour in some ambiguous scenarios.
To address the specific scenarios you brought up, you were partially right that I goofed the second scenario. It wouldn't start tracking the hellhound if it wasn't already active opponent, but if it was, it would stay the active opponent until that NPC died or some other NPC became active opponent. I've fixed that now.
As for the first scenario you mentioned, I don't think it's as weird as you say. If the player is attacking an NPC, then that NPC will always be the active opponent regardless of what happens in multi. The only time the most recent attacker will become the active opponent is if the player is not attacking any NPC at all. That might sound odd, but if you're, say, cannoning while idle, that NPC is probably the most recent enemy your cannon hit, in which case it's behaving similarly to the current version of the plugin.
Regardless, I hope the tests are to your liking and can help illustrate my explanation. Cheers.
While reviewing this, I've thought of another case which causes this code to fail, and I've written a brief test case which demonstrates it:
@Test
public void singleCombatIssue()
{
Player localPlayer = mock(Player.class), two = mock(Player.class), three = mock(Player.class);
when(client.getLocalPlayer()).thenReturn(localPlayer);
// TODO: find some unique var/condition representing a player being in multi-combat and mock it to be false here
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(localPlayer, two));
assertSame(two, opponentInfoPlugin.getLastOpponent());
// Another player tries to attack or cast magic on the player which causes interaction target changed from null -> player -> null in succession
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null));
assertSame(two, opponentInfoPlugin.getLastOpponent());
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, localPlayer));
assertSame(two, opponentInfoPlugin.getLastOpponent()); // this line fails
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null));
assertSame(two, opponentInfoPlugin.getLastOpponent());
}
Keep in mind this is, as far as I'm aware, the only case that would be affected by the player being in a multi-combat zone (or not, in this case)
If I'm understanding correctly, the reason this appears to change the value of lastOpponent is because you didn't also mock the localPlayer.getInteracting() method to return 'two'. It uses that to check if the player is still attacking a different actor when an actor attacks them, to prevent lastOpponent from taking on the value of the actor they aren't attacking. I did a test for this from lines 73-85.
While reviewing this, I've thought of another case which causes this code to fail, and I've written a brief test case which demonstrates it:
@Test
public void singleCombatIssue()
{
Player localPlayer = mock(Player.class), two = mock(Player.class), three = mock(Player.class);
when(client.getLocalPlayer()).thenReturn(localPlayer);
// TODO: find some unique var/condition representing a player being in multi-combat and mock it to be false here
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(localPlayer, two));
assertSame(two, opponentInfoPlugin.getLastOpponent());
// Another player tries to attack or cast magic on the player which causes interaction target changed from null -> player -> null in succession
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null));
assertSame(two, opponentInfoPlugin.getLastOpponent());
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, localPlayer));
assertSame(two, opponentInfoPlugin.getLastOpponent()); // this line fails
opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null));
assertSame(two, opponentInfoPlugin.getLastOpponent());
}
Keep in mind this is, as far as I'm aware, the only case that would be affected by the player being in a multi-combat zone (or not, in this case)
If I'm understanding correctly, the reason this appears to change the value of lastOpponent is because you didn't also mock the localPlayer.getInteracting() method to return 'two'. It uses that to check if the player is still attacking a different actor when an actor attacks them, to prevent lastOpponent from taking on the value of the actor they aren't attacking. I did a test for this from lines 73-85.
The tests do add full coverage of the changed code, though I'm still not 100% sure I understand this code inside and out. I'll review this more in depth at a later time.
I've added some further tests and cleaned up the test class a bit. @chaticon, does this still look good to you? From some quick ingame testing (in addition to the full test coverage provided by this PR's tests) it seems to still work just fine.
I've added some further tests and cleaned up the test class a bit. @chaticon, does this still look good to you? From some quick ingame testing (in addition to the full test coverage provided by this PR's tests) it seems to still work just fine.
Sorry for the late response. Looks good to me. And I think I may have commented on something outdated that you'd already fixed, sorry about that.
This throws an exception on login
java.lang.NullPointerException: null
at net.runelite.client.plugins.opponentinfo.OpponentInfoPlugin.onInteractingChanged(OpponentInfoPlugin.java:161)
at net.runelite.client.eventbus.EventBus$Subscriber.invoke(EventBus.java:67)
at net.runelite.client.eventbus.EventBus.post(EventBus.java:238)
at net.runelite.client.callback.Hooks.post(Hooks.java:169)
Additionally we should probably add comments to all of this logic since it looks complicated and I don't follow all of it.
We may want to keep an eye on this since there is no despawn event hooks for the opponent actor, and it is still invoking getInteracting() on it. But I think it is probably okay.
|
2025-04-01T04:35:24.248557
| 2019-04-08T01:05:29
|
430215513
|
{
"authors": [
"xDemoN"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10480",
"repo": "runelite/runelite",
"url": "https://github.com/runelite/runelite/pull/8442"
}
|
gharchive/pull-request
|
World Map: Identify Both Shield of Arrav Quest Start Points
Closes #8437
Let me know if you'd like the syntax to be different.
Enum names updated @Nightfirecat
|
2025-04-01T04:35:24.253347
| 2023-08-24T20:33:41
|
1865844411
|
{
"authors": [
"austinvazquez",
"vsiravar"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10481",
"repo": "runfinch/finch-core",
"url": "https://github.com/runfinch/finch-core/pull/142"
}
|
gharchive/pull-request
|
fix: Change rootfs compression to gzip
Issue #, if available:
Description of changes:
Change compression of rootfs bundle to gzip since windows does not come with zstd installed.
Testing done:
Yes, locally.
[X] I've reviewed the guidance in CONTRIBUTING.md
License Acceptance
By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Do we need to update Makefile and bin/update-rootfs.sh now as well? See https://github.com/search?q=org%3Arunfinch .tar.zst&type=code
Do we need to update Makefile and bin/update-rootfs.sh now as well? See https://github.com/search?q=org%3Arunfinch .tar.zst&type=code
This is a good point. However the Makefile is updated when https://github.com/runfinch/finch-core/blob/main/.github/workflows/update-dependencies.yaml#L32 action runs. For the next run we still need https://github.com/runfinch/finch-core/blob/main/bin/update-rootfs.sh#L5 as the Makefile has a zst compressed tar. Once the change is merged, I will update the update-rootfs.sh to
AARCH64_FILENAME_PATTERN="common/aarch64/finch-rootfs-production-arm64-[0-9].*\.tar.gz$"
AMD64_FILENAME_PATTERN="common/x86-64/finch-rootfs-production-amd64-[0-9].*\.tar.gz$"
|
2025-04-01T04:35:24.257866
| 2023-10-03T23:35:23
|
1925108567
|
{
"authors": [
"pendo324",
"vsiravar"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10482",
"repo": "runfinch/finch",
"url": "https://github.com/runfinch/finch/pull/594"
}
|
gharchive/pull-request
|
feat: windows persistent disk and other misc fixes/refactors
Issue #, if available:
Description of changes:
Adds persistent disk support to Windows
This took more work than anticipated, in order to deal with non-Admin users
pkg/disk/dpgo/ is brand new code, and should be a focus of the review
pkg/winutil/run_windows.go is also new and requires careful review. This is what allows Finch to run as the regular user, except for when it needs Admin access to call diskpart (to create the persistent disk)
Fix paths in nerdctl_config_applier to make the post-boot/init shelling work
Added winres to allow the finch.exe to have metadata attached to it. This is WIP, need final icons and descriptions etc.
Large (in terms of lines changed) refactor of pkg/path/finch.go, but it should have no impact on functionality (needs careful review)
Fixed the Makefile's clean target for Windows
Most of the other changes are just noise from refactoring (like, literally renaming things). Nothing major, but take a look if possible. Sorry the diff is so large
Testing done:
[x] I've reviewed the guidance in CONTRIBUTING.md
License Acceptance
By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Really cool PR, I left a few minor comments from the first pass, especially some files missing licenses.
|
2025-04-01T04:35:24.259242
| 2023-12-04T22:16:20
|
2024860212
|
{
"authors": [
"elboy3"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10483",
"repo": "rungalileo/dataquality",
"url": "https://github.com/rungalileo/dataquality/pull/814"
}
|
gharchive/pull-request
|
fix: disallow emb cols as metadata
https://app.shortcut.com/galileo/story/9579/dq-adding-x-y-data-x-or-data-y-as-metadata-succeeds-but-should-not
Customer ran into issues when logging x as metadata
@setu4993 yes we were getting quite an odd error on the runners side!
|
2025-04-01T04:35:24.272651
| 2023-12-19T21:54:56
|
2049495631
|
{
"authors": [
"Sirri69",
"alpayariyak",
"christopherwolfsaal",
"martinkallstrom"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10484",
"repo": "runpod-workers/worker-vllm",
"url": "https://github.com/runpod-workers/worker-vllm/issues/24"
}
|
gharchive/issue
|
Runpod
I tried to load the docker image and use in runpod. But i got the following message.
023-12-19T21:52:21.224549804Z Traceback (most recent call last):
2023-12-19T21:52:21.224594641Z File "/handler.py", line 3, in
2023-12-19T21:52:21.224600084Z import runpod
2023-12-19T21:52:21.224605461Z File "/usr/local/lib/python3.11/dist-packages/runpod/init.py", line 6, in
2023-12-19T21:52:21.224610387Z from . import serverless
2023-12-19T21:52:21.224616241Z File "/usr/local/lib/python3.11/dist-packages/runpod/serverless/init.py", line 16, in
2023-12-19T21:52:21.224627070Z from .modules import rp_fastapi
2023-12-19T21:52:21.224631744Z File "/usr/local/lib/python3.11/dist-packages/runpod/serverless/modules/rp_fastapi.py", line 10, in
2023-12-19T21:52:21.224636270Z from fastapi import FastAPI, APIRouter
2023-12-19T21:52:21.224640611Z File "/usr/local/lib/python3.11/dist-packages/fastapi/init.py", line 7, in
2023-12-19T21:52:21.224644971Z from .applications import FastAPI as FastAPI
2023-12-19T21:52:21.224649164Z File "/usr/local/lib/python3.11/dist-packages/fastapi/applications.py", line 16, in
2023-12-19T21:52:21.224653711Z from fastapi import routing
2023-12-19T21:52:21.224691454Z File "/usr/local/lib/python3.11/dist-packages/fastapi/routing.py", line 22, in
2023-12-19T21:52:21.224716007Z from fastapi import params
2023-12-19T21:52:21.224724461Z File "/usr/local/lib/python3.11/dist-packages/fastapi/params.py", line 5, in
2023-12-19T21:52:21.224816202Z from fastapi.openapi.models import Example
2023-12-19T21:52:21.224821635Z File "/usr/local/lib/python3.11/dist-packages/fastapi/openapi/models.py", line 4, in
2023-12-19T21:52:21.224825700Z from fastapi._compat import (
2023-12-19T21:52:21.224829829Z File "/usr/local/lib/python3.11/dist-packages/fastapi/_compat.py", line 20, in
2023-12-19T21:52:21.224913810Z from fastapi.exceptions import RequestErrorModel
2023-12-19T21:52:21.224923693Z File "/usr/local/lib/python3.11/dist-packages/fastapi/exceptions.py", line 3, in
2023-12-19T21:52:21.224950980Z from pydantic import BaseModel, create_model
2023-12-19T21:52:21.224958616Z File "/usr/local/lib/python3.11/dist-packages/pydantic/init.py", line 372, in getattr
2023-12-19T21:52:21.225128094Z module = import_module(module_name, package=package)
2023-12-19T21:52:21.225144639Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
2023-12-19T21:52:21.225149112Z File "/usr/lib/python3.11/importlib/init.py", line 126, in import_module
2023-12-19T21:52:21.225160188Z return _bootstrap._gcd_import(name[level:], package, level)
2023-12-19T21:52:21.225196585Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
2023-12-19T21:52:21.225211329Z File "/usr/local/lib/python3.11/dist-packages/pydantic/main.py", line 11, in
2023-12-19T21:52:21.225294129Z import pydantic_core
2023-12-19T21:52:21.225303857Z File "/usr/local/lib/python3.11/dist-packages/pydantic_core/init.py", line 30, in
2023-12-19T21:52:21.225314032Z from .core_schema import CoreConfig, CoreSchema, CoreSchemaType, ErrorType
2023-12-19T21:52:21.225320699Z File "/usr/local/lib/python3.11/dist-packages/pydantic_core/core_schema.py", line 15, in
2023-12-19T21:52:21.225353524Z from typing_extensions import deprecated
2023-12-19T21:52:21.225360153Z ImportError: cannot import name 'deprecated' from 'typing_extensions' (/usr/local/lib/python3.11/dist-packages/typing_extensions.py)
Yep, known issue because of locking dependencies
runpod==1.4.2
Absolutely silly behaviour. Pliz fix. ✨
Finding same
Should be fixed now
Will there be a new docker image or where it is fixed?
|
2025-04-01T04:35:24.275770
| 2017-06-30T22:11:35
|
239900189
|
{
"authors": [
"d3zd3z",
"utzig"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10485",
"repo": "runtimeco/mcuboot",
"url": "https://github.com/runtimeco/mcuboot/pull/77"
}
|
gharchive/pull-request
|
Always validate slot 0 when requested
The MCUBOOT_VALIDATE_SLOT0 feature only verifies the signature when
there is no swapping happening. The assumption was that if there is a
swap being done, the code will verify the signature of slot 1 before
doing the slot.
However, either due to bugs, or intentional trickery, it may be possible
to confuse the code into continuing a swap operation. If the data is
modified before this, the bootloader can be tricked into booting the
resulting image in slot 0 without having verified the signature.
Fix this by always verifying slot 0's signature before booting it.
JIRA: MCUB-64
Signed-off-by: David Brown<EMAIL_ADDRESS>
Don't merge yet, until we can create some enhancements to the sim that will allow us to test this.
@d3zd3z I'm doing some "final" changes for #81 which affect some of the same lines as this patch. Not sure there's something remaining to be done, otherwise would be glad if it is merged soon!
|
2025-04-01T04:35:24.286458
| 2023-10-29T17:47:15
|
1967090341
|
{
"authors": [
"Anmol-Baranwal",
"aslezar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10486",
"repo": "rupali-codes/LinksHub",
"url": "https://github.com/rupali-codes/LinksHub/pull/2145"
}
|
gharchive/pull-request
|
Added link to material tailwind
Fixes Issue
Fixes #2144
Changes proposed
New UI Library Material Tailwind added to the database of frontend ui- libraries
@rupali-codes
The issue of not showing in subcategories. In this case, the resource is limited.
|
2025-04-01T04:35:24.288144
| 2024-06-19T17:13:48
|
2362831320
|
{
"authors": [
"Riyazul555"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10487",
"repo": "rupali-codes/LinksHub",
"url": "https://github.com/rupali-codes/LinksHub/pull/2397"
}
|
gharchive/pull-request
|
Yaak added in Backend Testing
Fixes Issue #2380
Changes proposed
Screenshots
Note to reviewers
@rupali-codes please review this pR
Thanks
|
2025-04-01T04:35:24.289527
| 2019-06-30T18:23:23
|
462432380
|
{
"authors": [
"bernardesrodrigo",
"ramezanifar"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10488",
"repo": "ruscito/pycomm",
"url": "https://github.com/ruscito/pycomm/issues/60"
}
|
gharchive/issue
|
Should it work with AB Emulator 5000?
Hi,
Greats for the library, I´m trying to make it work with RSLogixEmulator 5000 running in remote machine over ethernet.
Should it work like that or only with real hardware?
When reading a tag I´m getting the error:
"pycomm.cip.cip_base.CommError: can only concatenate str (not "bytes") to str"
RSLogixEmulator 5000 does not let read and write. Instead use SoftLogix5800. It makes your PC like a PLC.
|
2025-04-01T04:35:24.303470
| 2016-11-12T03:09:49
|
188888785
|
{
"authors": [
"edtanous",
"ruslo"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10489",
"repo": "ruslo/hunter",
"url": "https://github.com/ruslo/hunter/pull/580"
}
|
gharchive/pull-request
|
Fix an issue in OpenSSL build when cross compiling
The commit 44c681e1a551e0599a8ca170cb0cfcc19c4a5d54 seems to have
broken cross compiling builds, as openssl ./config seems to require
both the CC and MACHINE variables to be set, otherwise it seems to
default to build host compiler and machine variables. Someone appears
to have already seen this issue when cross compiling for android, and
fixed it in commit bb657ff41c874ca3bba1ca394ff1ea09ee8049b3. This
patch moves the MACHINE argument and adds it to ALL build types, not
just android.
I'm unclear on the architecture of the PACKAGE_INTERNAL_DEPS_ID, but I saw it being incremented for other similar changes, so I assume this will require an update as well. If I'm wrong, let me know and I will update the PR.
I'm unclear on the architecture of the PACKAGE_INTERNAL_DEPS_ID
I've added some details here:
https://docs.hunter.sh/en/latest/reference/user-modules/hunter_download.html
The commit 44c681e seems to have broken cross compiling builds
So what platforms do we fix by this change?
Testing:
https://ci.appveyor.com/project/ingenue/hunter/build/1.0.1165
https://travis-ci.org/ingenue/hunter/builds/175285201
This fixes cross compiling from Linux x86_64 to Armv6 32 bit. My CMAKE_TOOLCHAIN_FILE looks like the following
SET(CMAKE_SYSTEM_NAME Linux)
SET(CMAKE_SYSTEM_VERSION 1)
set(CMAKE_SYSTEM_PROCESSOR armv6)
# specify the cross compiler
SET(CMAKE_C_COMPILER arm-linux-gnueabi-gcc)
SET(CMAKE_CXX_COMPILER arm-linux-gnueabi-g++)
# where is the target environment
SET(CMAKE_FIND_ROOT_PATH /usr/arm-linux-gnueabi)
# search for programs in the build host directories
SET(CMAKE_FIND_ROOT_PATH_MODE_PROGRAM NEVER)
# for libraries and headers in the target directories
SET(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY)
SET(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY)
This is on an Ubuntu 16.04 machine with a 5.4 gcc toolchain standard, although it's not clear if that matters.
it's not clear if that matters
I will merge it because it doesn't break old behavior and seems to fix the issue for you. However I can't guarantee that there will be no problem with your toolchain in future. If you want to be sure you have to extend testing.
Uploading:
https://ci.appveyor.com/project/ingenue/hunter/build/1.0.1171
https://travis-ci.org/ingenue/hunter/builds/175429374
Released: https://github.com/ruslo/hunter/releases/tag/v0.16.27
|
2025-04-01T04:35:24.313403
| 2021-07-12T13:52:53
|
942088120
|
{
"authors": [
"0x2b00b1e5",
"gwenn",
"thomcc"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10490",
"repo": "rusqlite/rusqlite",
"url": "https://github.com/rusqlite/rusqlite/issues/992"
}
|
gharchive/issue
|
Parsing keys for SQLCipher databases
Hi, I am probably just missing something but I tried to connect to a sqlcipher database but it didn't work.
I first did conn.execute_batch("PRAGMA key = 'test123';") and then conn.execute_batch("SELECT count(*) FROM sqlite_master;") as a test, but this did not decrypt the database correctly for some reason (returned message: file is not a database). I have tested the decryption using the sqlcipher CLI tool, which just worked fine.
I am using the sqlcipher-bundled feature.
execute_batch should be fine but you should use:
conn.pragma_update(None, "key", &passphrase)?;
And do you know which kind of error is expected when the passphrase is incorrect ?
Page https://www.zetetic.net/sqlcipher/sqlcipher-api/#PRAGMA_key doesn't help.
https://discuss.zetetic.net/t/file-is-not-a-database/3320
https://discuss.zetetic.net/t/error-file-is-not-a-database/4434
let cipher_version: String = db.pragma_query_value(None, "cipher_version", |row| row.get(0))?;
And do you know which kind of error is expected when the passphrase is incorrect ?
It gives the error "not a database" when you try to use it (e.g. on the first query, not on the PRAGMA key). It is somewhat weird...
This sounds like your key is wrong to me. This wouldn't explain why it works on the CLI though.
Perhaps the page size doesn't match though (it must match for sqlcipher, and perhaps it has a different default on the CLI)...
execute_batch should be fine but you should use:
conn.pragma_update(None, "key", &passphrase)?;
Tried that too, same result.
Just to clarify that the
conn.execute_batch("PRAGMA key = 'test123;');
... does succeed. The problems start when trying to touch/read/alter the database.
Perhaps the page size doesn't match though (it must match for sqlcipher, and perhaps it has a different default on the CLI)...
Maybe? How would I change the page size?
Update: used the rusqlite api to create a db instead. This does work. So I guess there is indeed a pagesize/encoding issue there.
|
2025-04-01T04:35:24.315481
| 2020-11-11T03:50:50
|
740435956
|
{
"authors": [
"Veykril",
"jrmuizel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10491",
"repo": "rust-analyzer/rust-analyzer",
"url": "https://github.com/rust-analyzer/rust-analyzer/issues/6527"
}
|
gharchive/issue
|
Renaming type renames variables
Given the following code:
struct tag {
x: i32
}
fn main() {
let tag = Some(4);
if let Some(tag) = tag {
dbg!(tag);
}
}
renaming struct tag to struct Tag incorrectly produces:
struct Tag {
x: i32
}
fn main() {
let tag = Some(4);
if let Some(tag) = Tag {
dbg!(Tag);
}
}
This seems to have been fixed.
|
2025-04-01T04:35:24.317671
| 2022-02-03T12:20:06
|
1123019532
|
{
"authors": [
"flodiebold"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10492",
"repo": "rust-analyzer/rust-analyzer",
"url": "https://github.com/rust-analyzer/rust-analyzer/pull/11399"
}
|
gharchive/pull-request
|
Fix assoc type shorthand from method bounds
In code like this:
impl<T> Option<T> {
fn as_deref(&self) -> T::Target where T: Deref {}
}
when trying to resolve the associated type T::Target, we were only
looking at the bounds on the impl (where the type parameter is defined),
but the method can add additional bounds that can also be used to refer
to associated types. Hence, when resolving such an associated type, it's
not enough to just know the type parameter T, we also need to know
exactly where we are currently.
This fixes #11364 (beta apparently switched some bounds around).
bors r+
|
2025-04-01T04:35:24.318861
| 2020-02-18T11:26:40
|
566828109
|
{
"authors": [
"flodiebold",
"lnicola"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:10493",
"repo": "rust-analyzer/rust-analyzer",
"url": "https://github.com/rust-analyzer/rust-analyzer/pull/3215"
}
|
gharchive/pull-request
|
Exclude methods from non-parameter types introduced by generic constraints
Fixes #3184.
r? @flodiebold
bors r+
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.