added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:35:22.104557
2016-09-26T09:28:39
179179595
{ "authors": [ "challiwill", "ktogo", "robfig" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10382", "repo": "robfig/cron", "url": "https://github.com/robfig/cron/pull/76" }
gharchive/pull-request
Fix: Cron.run() was resetting the timezone time.Timer() does not return a timezone within time.Time, since Timer is unaware of the custom timezone as it only accepts time.Duration when it's initialized. Thus the receiver needs to take care of the timezone. // NewTimer only receives time.Duration func NewTimer(d Duration) *Timer // And time.Duration only holds int64 but no timezone specifier type Duration int64 Interesting.. your change to the tests causes them to fail without your change to the code? Great find, thanks +1 I was about to try and track down what was happening here also. Good find.
2025-04-01T04:35:22.131692
2017-08-23T09:47:35
252222908
{ "authors": [ "AndiLeni", "robiso" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10383", "repo": "robiso/wondercms", "url": "https://github.com/robiso/wondercms/issues/39" }
gharchive/issue
Persistant update notification | Summernote button size Hello, I found several issues which I could not resolve. The first issue is the notification which says "New WonderCMS update available." When I click on "Update WonderCMS" the success message shoows up but the notification itself won't disappear. The second one has to deal with the summernote plugin. The buttons of the editor don't have the same height. In the developer console I could not find any css class which overrides the default values for the buttons. Hopefully someone can help me :) Regards AndiLeni Hello AndiLeni. Please answer a couple of question to make the assistance faster. Have you ever successfully used the update before? When you open the settings panel, what version of WonderCMS does it show you? Are you coming from a clean install of WonderCMS? Do you know what version of PHP you're running and on what server (Apache, NGINX, IIS)? I'm sure more question will arise such as (if you're coming from an older version): Did you update the summernote plugin manually, or is this a clean install? Were the summernote buttons fine before the update (if you are in fact coming from an update)? No, it is the first time I tried WonderCMS 2.3.0 see 1. I installed Wondercraft on my local webserver using MAMP. Apache, PHP 7.0.9 Everything was installed freshly. Somebody had a similar issue before on a local installation, because of a missing certificate on the local machine: https://github.com/robiso/wondercms/issues/33 @halojoy solved this with I downloaded a certificate from the cURL website. cacert.pem So now it works alright. You could turn of the cerificate check for cURL, but it's bad practice and considered unsafe as it enables MITM (Man In The Middle) attacks, so please use the above solution to solve your problem. Just to add, this can be fixed on your local machine by install the certificate as posted above. So it would work with your MAMP installation without any problems, thus reducing the need to test it on a "real" host. I skipped the prozesss to install a certificate and simply added the line curl_setopt($ch, CURLOPT_SSL_VERIFYPEER, 0); to the _getExternalFile function. The update notification is now working as expected but the Summernote plugin still does not work. UPDATE: For further reference and if someone might has this problem too here is the solution. In my php.ini file display_errors was set to on which forces the browser to render a warning, notice, error, etc. before the <!DOCTYPE html> tag. As seen in your link to stackoverflow the problem was caused by a wrong doctype. Example: <br /> <b>Notice</b>: Undefined index: token in <b>C:\MAMP\htdocs\wCMS\index.php</b> on line <b>408</b><br /> <br /> <b>Notice</b>: Undefined index: token in <b>C:\MAMP\htdocs\wCMS\index.php</b> on line <b>365</b><br /> <!DOCTYPE html> As I stated before: turning of the certificate check (with curl_setopt($ch, CURLOPT_SSL_VERIFYPEER, 0);) instead of actually installing the certificate is considered bad practice, unsafe - it enables MITM (Man In The Middle) attacks. Regarding the Summernote plugin, we have nothing to do with the actual styling of the plugin except for turning off the background to transparent. I tried Googling for your problem and the only thing I came up with was: https://stackoverflow.com/questions/24594828/summernote-buttons-varying-height. You can open a seperate issue about this as this is not a part of WonderCMS. Summernote is a seperate project, and a WonderCMS plugin was made by Prakai Nadee, I only patched up the vulnerabilities and turned off the Air mode. When you open a new issue in the plugins section, make sure to post a screenshot or a demo example on your server. The cause for those errors is because we've missed to check if the REQUEST[TOKEN] exists before putting into an if statement. Closing this issue and issuing a patch soon. Note: the patch won't fix the cURL certificate error you're having. Updating WonderCMS to a newer version overwrites any changes you've made to index.php. If you need to make changes, you can do so with the functions.php file or a plugin, check our wiki (https://github.com/robiso/wondercms/wiki) for more info. // Edit, here's the fixed index.php, you should not be getting any more errors. Please confirm. index.php.zip Fatal error: Cannot use isset() on the result of an expression (you can use "null !== expression" instead) in C:\MAMP\htdocs\wCMSold\index.php on line 406 Found the type, please try again. index.php.zip If everything goes well, let me know. Can you please post your first name / last name and link to website so we can link to it on https://wondercms.com/whatsnew I guess you posted the wrong file. The first one you gave me and the second one are the same :) My name is Andreas Lenhardt. Here comes the right one: index.php.zip We'll list you just by name, any website links, or maybe Twitter? Parse error: syntax error, unexpected ')' in C:\MAMP\htdocs\wCMSold\index.php on line 406 A ) too much. If I remove it it's fine. You can link to: https://twitter.com/AndiLenhardt Sorry for the hassle, one last time attaching the corrected ZIP. index.php.zip If you don't mind running it for a test one more time. This will be dispatched as version 2.3.1 within the next 24 hours. Thanks @AndiLeni. Patch for the two "is sets" deployed with 2.3.1. @AndiLeni, we have thanked/linked to you on the following locations: https://wondercms.com/whatsnew https://github.com/robiso/wondercms/releases/tag/2.3.1 https://wondercms.com/forum/viewtopic.php?f=8&p=1833#p1833 We have additionally added you to our special contributors list https://www.wondercms.com/special-contributors Also on the WonderCMS download page https://www.wondercms.com/latest Thank you. :) There are still some warnings which didn't occur before. Are you sure you didn't mix up some files accidentally? Notice: Undefined index: token in C:\MAMP\htdocs\wCMS\index.php on line 407 Notice: Undefined index: token in C:\MAMP\htdocs\wCMS\index.php on line 364 Both of them can be fixed using isset() Are you sure this happens with the latest 2.3.1 version? Here's the official index.php: https://github.com/robiso/wondercms/blob/master/index.php Both of those lines you're reporting have an isset check in the official repository. Well, obviously this time I messed stuff up. My fault, sorry! We're only human. :)
2025-04-01T04:35:22.170058
2015-10-30T18:52:58
114332405
{ "authors": [ "emanchado" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10384", "repo": "robohydra/robohydra", "url": "https://github.com/robohydra/robohydra/issues/17" }
gharchive/issue
Implement plugin configuration defaults Many plugins might share configuration keys, and it's tedious to repeat their values. Thus, there should be some kind of plugin configuration defaults in the configuration file. See https://groups.google.com/forum/?hl=es&fromgroups#!topic/robohydra/Bsf8Wo0efvE for background. This should do the trick. I'm not sure when I'll make a new release (I'd like to try to get the proxy support in there, too), but for now you can use master.
2025-04-01T04:35:22.177225
2018-05-01T05:22:48
319119509
{ "authors": [ "cmeury", "gtaylor" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10385", "repo": "roboll/helmfile", "url": "https://github.com/roboll/helmfile/issues/131" }
gharchive/issue
KubeCon Europe 2018 I am not sure whether an issue is the right place to discuss this, but I wouldn't know of any other communication channel (Chat, ML). Is any maintainer, contributor or user at KubeCon Europe in Copenhagen this week? If yes, care to have a coffee? Drop me a mail. This issue might be changed to: establish and/or communicate discussion channel for helmfile devs and users. Could be worth seeing if we could get a helmfile channel created on the Kubernetes Slack org. Yeah, I asked the admins already, will check once in the hotel :) On Tue, May 1, 2018, 17:48 Greg Taylor<EMAIL_ADDRESS>wrote: Could be worth seeing if we could get a helmfile channel created on the Kubernetes Slack org. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/roboll/helmfile/issues/131#issuecomment-385705351, or mute the thread https://github.com/notifications/unsubscribe-auth/ABQgSmLxYyp7Z4q8zXj4qP-P4u7-u0x9ks5tuIPfgaJpZM4TtlsX . Didn't have a response from the admins - maybe try again in the future, closing for now.
2025-04-01T04:35:22.182029
2022-11-07T08:01:30
1437924411
{ "authors": [ "Danfoa", "stephane-caron" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10386", "repo": "robot-descriptions/robot_descriptions.py", "url": "https://github.com/robot-descriptions/robot_descriptions.py/issues/12" }
gharchive/issue
Missing dependency package yourdfpy Good Day, This issue is just to call attention to a missing dependency during package build and install. Specifically, yourdfpy package is missing as a description and throws an error when using the CLI. I will submit a pull request with the change soon if it is not easily fixed. Thanks for your feedback :slightly_smiling_face: If everything is implemented correctly yourdfpy should be an optional dependency (e.g. pip install robot_descriptions[cli] installs it): There shouldn't be any yourdfpy-related import error when doing import robot_descriptions If an example or sub-module uses it, it should raise a user-friendly instruction such as "Importing this module requires yourdfpy, you can install it by..." I see that point 2 was lacking, so I started a PR for it: https://github.com/robot-descriptions/robot_descriptions.py/pull/13 @Danfoa feel free to review and let me know if that would fix it for you.
2025-04-01T04:35:22.183017
2017-12-13T20:09:44
281879695
{ "authors": [ "jmauss" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10387", "repo": "robotattackorg/robot-detect", "url": "https://github.com/robotattackorg/robot-detect/issues/9" }
gharchive/issue
Scanning subnets I would love to have the ability to scan entire subnets with this tool, unless I am missing something on how this is done? (Using .* and 0/24 both do not work) I have closed this as I just used a bash script to fix the problem. Reopen if wanting to implement.
2025-04-01T04:35:22.187995
2016-08-24T00:57:58
172841721
{ "authors": [ "arunku2x", "pekkaklarck" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10388", "repo": "robotframework/robotframework", "url": "https://github.com/robotframework/robotframework/issues/2429" }
gharchive/issue
Values of Robot Framework Automatic Variables (${TEST STATUS} and ${SUITE STATUS}) not correct in certain scenarios. Hi, I have found that the values of Robot Framework Automatic Variables (${TEST STATUS} and ${SUITE STATUS}) not correct in the following scenarios. Added Test Timeout parameter under Test Suite settings so that it becomes applicable for all test cases under the test suite. Observed that in case a test case is timed out, its status ${TEST STATUS} still return PASS in Test Case Teardown section. This happens if a keyword in Test Case Teardown is timed out. There was a failure in Test Suite Teardown section but its status ${SUITE STATUS} returned PASS. I want to perform some actions in Test Case and Test Suite teardown sections on the basis of test case and test suite status but due to inconsistent status of these variable, not able to do so. Could you please look into these issues. This is a variation of #2078. Test timeout occurring during the teardown is a bit more problematic than a normal failure, though, becuase test timeouts are, by design, disabled during the teardown to allow teardowns to run fully. Robot only marks the test failed after the test ended if it in the end took too much time. I'll update #2078 accordingly and close this issue. Notice that the required changes to actually fix this are pretty large. They cannot be done in RF 3.0.1 release but need to wait for a major version. Unfortunately I cannot give any promises when RF 3.1 would be release nor would these fixes make into that release. Hi Pekka, As an alternative, I switched to listeners. I have added listener for end_test and end_suite and taking decision on the basis of attrs['status'] field inside these listener functions. Here as well, I have observed one weird issue. I had a test suite containing 7 test cases. Now all the test cases inside this test suite failed and I got correct test case status inside "end_test" listener. The issue which I have observed is that test suite status inside "end_suite" listener shows up as PASS. Wouldn't it be FAIL even if 1 test case inside the test suite fails. Regards, Arun Kumar Saini Suite status is PASS if all critical tests pass. Most likely cause for your issue is that tests are non-critical. If that's not the case, please submit a separate issue with instructions how to reproduce the problem. An executable example would be best.
2025-04-01T04:35:22.201267
2017-03-24T19:54:59
216889288
{ "authors": [ "m3d", "mdspk", "zwn" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10389", "repo": "robotika/simple-sfm", "url": "https://github.com/robotika/simple-sfm/pull/1" }
gharchive/pull-request
Calculate reprojection error. This is the "simplest thing that could possibly work" regarding reprojection error. Calculates gradient with respect to 3D points positions in preparation for bundle adjustment. Open question is about the best representation of orientation in 3D wrt optimization by gradient descent. Quaternions seem somewhat straight forward. Euler angles could be also used but one is never quite sure about which rotation is which. Could you please change README.md to direct links for installation? (I suppose that requirements.txt would not help in this case). Thanks https://conda.io/docs/install/quick.html d:\>conda install autograd Fetching package metadata ........... PackageNotFoundError: Package not found: '' Package missing in current win-64 channels: - autograd You can search for packages on anaconda.org with anaconda search -t conda autograd You may need to install the anaconda-client command line client with conda install anaconda-client Done. In the future I intend to add travis and appveyor unit tests, so that will always contain up-to-date install instructions. What I am most interested in the review is the readability of the code. I'd like the code to be easy to understand with minimum comments. Thanks, now it works :). I would already split it into two files (bundle.py and test_bundle.py) and somehow I am used to 2 empty lines from PEP8 (sorry for stupid comments). Maybe test_project() and test_rotate() into unittest and test_reprojection_error() keep with prints in the bundle.py Open question is about the best representation of orientation in 3D wrt optimization by gradient descent. Quaternions seem somewhat straight forward. Euler angles could be also used but one is never quite sure about which rotation is which. Any opinions about this? I would start with Quaternions ? delete branch?
2025-04-01T04:35:22.206145
2018-01-15T12:17:34
288575475
{ "authors": [ "jgvictores", "nigno17" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10390", "repo": "robotology/QA", "url": "https://github.com/robotology/QA/issues/270" }
gharchive/issue
How to change control mode in python Hello everyone, How can I change back the joints control mode in position after using the cartesian control in python? I need to first control the robot in position mode and then use the cartesian controller to control in task space. This process needs to be repeated in a loop. Thanks in advance, Nino Hi @nigno17 @pattacini !! Using the old IControlMode interface, you should still be able to use: props = yarp.Property() props.put("device","remote_controlboard") props.put("local","/client/right_arm") props.put("remote","/icubSim/right_arm") armDriver = yarp.PolyDriver(props) iPos = armDriver.viewIPositionControl() jnts=iPos.getAxes() ... ... iMode = armDriver.viewIControlMode() for j in range(1,jnts): iMode.setPositionMode(j) Bindings for IControlMode2 are still not in master, but in devel as of https://github.com/robotology/yarp/commit/554b7f64418a1949a2cce053d099457a11899807. I'll brew up an example if I have time and remember. :-) Thank a lot. The old IControlMode interface works perfectly.
2025-04-01T04:35:22.217998
2021-11-09T10:03:47
1048411930
{ "authors": [ "GiulioRomualdi", "YuRan-W", "tmacattank" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10391", "repo": "robotology/osqp-eigen", "url": "https://github.com/robotology/osqp-eigen/issues/107" }
gharchive/issue
Why the solver returns "the problem is non-convex" or " solve successfull" I am now using quadratic programming to solve a trajectory planning problem, and the test scenario is to overtake a vehicle. I try OSQP diresctly and OSQP eigen. OSQP works well, while when i use osqp eigen, sometimes it works well, sometimes it suddenly return "the problem is non-convex" or "solve unccessful". I check the error code from OSQP, and only when the Hessian matrix is not positive definite, can the problem becomes non-convex. I have follow the OSQP eigen MPC example grammar rules to use this tool, is there any suggestions on the randomly occured crash? Hi @tmacattank. Could you please give us a minimum example that we can use to better debug the problem? I am using an online simulation platform based on ROS to develop the algorithm, I will try to find an easy way so that you can run my code Hello, I also encountered this problem, did you solve it?
2025-04-01T04:35:22.306725
2023-10-17T04:25:29
1946514913
{ "authors": [ "CLAassistant", "michMartineau" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10392", "repo": "robusta-dev/robusta", "url": "https://github.com/robusta-dev/robusta/pull/1129" }
gharchive/pull-request
Add custom labels on pods We would like to add custom labels on forwarder and runner pods. We will use them in order to show which team is responsible for this pods. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T04:35:22.335441
2018-04-19T08:39:26
315779087
{ "authors": [ "coveralls", "fanc999", "rockdaboot" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10393", "repo": "rockdaboot/libpsl", "url": "https://github.com/rockdaboot/libpsl/pull/98" }
gharchive/pull-request
Enable autotools builds to build Windows binaries Hi, Apparently I was working on my branch at the same time when it was being merged, sorry! The other changes that was done in libpsl.h.in shouldn't affect things at this point. This updates the autotools builds (which I am not too good at) for enabling builds with MinGW. I will also work on Visual Studio NMake Makefiles shortly afterwards. With blessings, thank you! Coverage remained the same at 93.648% when pulling 917be0ffa82090006fe5f8dec575f60827ea2fe5 on fanc999:master into 44256b1a3aa1309a2fe60af942a8c7ace17dbcea on rockdaboot:master. The commit is too windows-centric. So here are some questions: We link to -lws2_32, which is necessary for the networking functions on Windows. Why don't you use AC_SEARCH_LIBS(inet_ntop, ws2_32) or something like this ? We default to using libicu on Windows... That's what we have configure flags for. Please describe the problem you have using them. ...not have fmemopen() So please add fmemopen to AC_CHECK_FUNCS in configure.ac, and modify the fuzzer code to exit(77) if HAVE_FMEMOPEN is not defined. That displays a 'SKIP' when running the test suite. Hi Tim, I updated the check for inet_pton(), which was what was being used, as you suggested. For the libicu part, I am leaving that intact for now, as basically the issue is where we set the default runtime and builtin as libidn2. The thing is code that uses libidn2 (and libidn as well) make use of items in langinfo.h, which is a header that is not available on native Windows builds (be it MinGW or Visual Studio). This means that --enable-runtime=libicu and --enable-builtin=libicu is required for the code to build on Windows at the current state (or we disable the runtime annd builtins). Another way is to update README.md to tell people that currently it is required to pass in --enable-runtime=libicu and --enable-builtin=libicu (or use --disable-runtime and --disable-builtin) to build libpsl on Windows. Please let me know which is your preferred way to this issue. Also, I am doing another commit here for the fuzz tests part. With blessings, thank you! make use of items in langinfo.h Right, there also should be a guard around the cal to nl_langinfo() and in the else case call the appropriate Windows function. Do you have it at hand ? I have it somewhere, must search... Another way is to update README.md to tell people that currently it is required to pass in --enable-runtime=libicu and --enable-builtin=libicu (or use --disable-runtime and --disable-builtin) to build libpsl on Windows. That sounds good for now. On the other hand, if libidn2 development files aren't installed, ./configure should fallback to libidn and then to libicu automatically. I thought it was working that way, but will have a look. Maybe it works if you fix the nl_langinfo() thing ? Hi Tim, I will look into nl_langinfo() with the appropriate Windows replacement later. For the default runtime/builtin, unfortunately configure errors out, saying something like requested libidn2 but could not find it, so no automatic fallback is being done there. I think for the Visual Studio builds, getting libidn2/libidn and libunistring is going to be tedious, due to their use of gnulib (big autotools-using stuff there)... So, it is likely that for the Visual Studio builds, I will focus on using ICU and the case where the runtime/builtin is disabled. With blessings, thank you! OK, just let me know if you are interested in a MinGW build script for libunistring, libidn2 and libidn. I use one for wget2 CI builds. Might help before you start from scratch. @fanc999 Pushed two of your commits and added another one implementing the fallback thing I mentioned. Indeed, it wasn't implemented (must have been a dream ;-)). Thanks for your work. Hi, I posted a new patch to enable DLL builds on Windows, as we need the -no-undefined linker flag on Windows for building DLLs, and enabled linking to libintl and libiconv when we are using the libidn/libidn2 code, otherwise the build will fail to link since APIs from these two libraries are being used (libintl and libiconv are things that are not in the MinGW/Windows CRT library). Let me know if these two changes could be done in a better way, since I am by no means good in autotools. With blessings, thank you! Can you remove the OS specific code and try this instead, please: diff --git a/configure.ac b/configure.ac index f0e627b..501d83e 100644 --- a/configure.ac +++ b/configure.ac @@ -10,7 +10,7 @@ AC_CONFIG_HEADERS([config.h]) AC_PROG_CC m4_ifdef([AM_PROG_AR], [AM_PROG_AR]) #LT_INIT([disable-static]) -LT_INIT +LT_INIT([win32-dll]) AC_CONFIG_MACRO_DIR([m4]) m4_ifdef([AM_SILENT_RULES], [AM_SILENT_RULES([yes])]) diff --git a/src/Makefile.am b/src/Makefile.am index e18d754..c645b70 100644 --- a/src/Makefile.am +++ b/src/Makefile.am @@ -11,7 +11,7 @@ libpsl_la_CPPFLAGS = -I$(top_srcdir)/include -I$(top_builddir)/include -DPSL_DIS $(CFLAG_VISIBILITY) -DBUILDING_PSL # include ABI version information -libpsl_la_LDFLAGS = -version-info $(LIBPSL_SO_VERSION) +libpsl_la_LDFLAGS = -no-undefined -version-info $(LIBPSL_SO_VERSION) if WITH_LIBICU libpsl_la_LDFLAGS += $(LIBICU_LIBS) endif It's from https://www.gnu.org/software/libtool/manual/html_node/LT_005fINIT.html, scroll down to 'win32-ddl'. The code works here. If it works for you we should use it. The second issue, libintl and libiconv, should also be done automatically by libtool. We just have to add $(LTLIBICONV) $(LTLIBINTL) @INTL_MACOSX_LIBS@ in Makefile.am at the right place(s). I'll try to find a machine/OS for testing. Added two commits that address the above. Please check and give me feedback. Hi Tim, The two commits do work. With blessings, thank you!
2025-04-01T04:35:22.359168
2023-01-29T13:23:57
1561287706
{ "authors": [ "aidenlangley", "tversteeg" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10395", "repo": "rockerBOO/awesome-neovim", "url": "https://github.com/rockerBOO/awesome-neovim/pull/673" }
gharchive/pull-request
Add ~nedia/auto-save.nvim Checklist: [x] The plugin is specifically built for Neovim, or if it's a colorscheme, it supports treesitter syntax. [x] The lines end with a .. This is to conform to awesome-list linting and requirements. [x] It's not already on the list. [x] The title of the pull request is Add/Update/Remove `username/repo` when adding a new plugin. [x] The description doesn't start with A Neovim plugin for... or A plugin for..., and doesn't end with ... for Neovim. [x] The description doesn't mention that it's a Neovim plugin, it's obvious from the rest of the document. [x] Neovim is spelled as Neovim (not nvim, NeoVim or neovim), Vim is spelled as Vim (capitalized), Lua is spelled as Lua (capitalized). Thanks for the PR!
2025-04-01T04:35:22.478877
2023-02-01T16:33:58
1566433949
{ "authors": [ "daniel-ciocirlan", "niklasuhrberg" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10396", "repo": "rockthejvm/rockthejvm.github.io", "url": "https://github.com/rockthejvm/rockthejvm.github.io/pull/25" }
gharchive/pull-request
Article on Akka Streams to actors interoperability Here is the article on Akka Streams and actors interoperability. Please note that I used html for the image of the creek and specified its width. It was rendered too big without this I think. Also, will make sure the final picture (whether it's changed or not) is guaranteed to be freely used. When I added a Title below the fron matter , the title appeared twice so I only have the title specified by the front matter. Thanks!
2025-04-01T04:35:22.495297
2024-09-30T20:25:49
2557574316
{ "authors": [ "Piotrk39", "gulfaraz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10397", "repo": "rodekruis/IBF-system", "url": "https://github.com/rodekruis/IBF-system/issues/1599" }
gharchive/issue
Create issues for E2E tests in trigger scenarios Who requested this task? @gulfaraz What needs to be done [x] Identify differences between trigger and no-trigger state on the UI (start with Uganda flood) [x] Update test plan [x] Create follow-up items Acceptance Criteria [x] Follow-up items created Relevant links Use prioritization done for the no-trigger state for reference. All the above test were added in "Qase" as test scenarios and separate items were created in "New" tab
2025-04-01T04:35:22.499490
2020-03-29T01:35:51
589695683
{ "authors": [ "rodrigo-brito", "rrfaria" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10398", "repo": "rodrigo-brito/backtrader-binance-bot", "url": "https://github.com/rodrigo-brito/backtrader-binance-bot/issues/6" }
gharchive/issue
how to use it in a different broker I was taking a look on your code and it is pretty good but it is to binance I need use it in iqoption as a broker there is a community behind iqoption trying to use iqoption api but no one know how to put it on backtrader do you know how to create a broker on backtrader Is there any video on internet about it ? Hi @rrfaria, You can create a custom class with iqoption operations like this: https://gist.github.com/rodrigo-brito/8c82020f04e946e3f0c39c7243cfe1ee. In the code, you can create an instance of this custom class in the constructor of your strategy and execute the buy or sell function instead of backtrader built-in functions. The main problem is the data feed. You should create a class wich provides some fields: https://www.backtrader.com/docu/datafeed-develop-general/datafeed-develop-general/ thanks @rodrigo-brito but if I use this class on strategy I'll not lose boker functionalities ? I need mensure if it is getting profit or not ? iqoption uses spread on forex and crypto I would like to use a strategy where use it as a important property to do buy or sell because some time when spread is too high doesn't compensate do any thing and it is hold time I found oanda code: https://github.com/ftomassetti/backtrader-oandav20/tree/master/btoandav20 he implements store, broker, sizer and feed I tried to read backtrader docs but it is not so clear
2025-04-01T04:35:22.506651
2017-02-08T09:43:48
206143873
{ "authors": [ "rofrischmann", "tiagojsalmeida" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10399", "repo": "rofrischmann/fela", "url": "https://github.com/rofrischmann/fela/issues/195" }
gharchive/issue
Monolithic mode? Version: 3.0.x Environment: Both Type: Feature Description Atm you guys only support atomic classnames. Would it be possible to implement an monolithic mode? Proposal Code Example: createRenderer({ mode: 'monolithic' }) Not sure about that. I don't see any advantages actually. But it might be possible to just have an alternative renderer. The React Native renderer e.g. is also a totally different renderer. @rofrischmann In our case it would be ideal to use the monotithic version on development as it's just easy to change styles. On our implementation we made our own render method as we don't want to use sheet.insertRule for multiple reasons (specially because of third-party applications). I would totally help. Could you point me on the right direction? You could basically take the existing renderer and simply replace renderRule. I may give you a list of "things" which you need to do in order to achieve correct classnames etc. Added with #200
2025-04-01T04:35:22.575279
2014-12-06T10:58:53
51179036
{ "authors": [ "bhh", "dgilperez", "guzart", "ianheggie", "jondeandres", "rokob" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10400", "repo": "rollbar/rollbar-gem", "url": "https://github.com/rollbar/rollbar-gem/issues/193" }
gharchive/issue
Configuration option to disable Rails support i know its just lame to still have 2.3 applications but i got this "new" old application i have to take care of and first thing is i want to log errors nicely. is there a way to disable the default rails stuff per config? the only problem i have is that it wants to autoload railtie and active_record_extension if i could disable it per config everything would work fine and i could just have to rescue the errors and stuff i want to and send it with the gem (this part seems to work perfectly because there is not rails integration) if you think wtf he is crazy wanting rails 2.3 support just delete it ;) i dont really need support for it just a method to disable rails support would it be a problem if i try to get it into the config with a PR? ok new approach. what do u think of the idea of spliting the rails features. so we have a rollbar-gem and rollbar-rails with rollbar dependecy of course. like rspec-rails and stuff. additionally we could implement the javascript client side integration also in the rails gem I think environment flag is probably the easiest way to go, and best practices ATM suggest your app gets the configuration from environment variables (running context). easy definitly but i like the idea of extracting the rails part way more. and if the owners approve, i have no problem of doing a bit more work to keep everything clean. for me it feels nicer to have a core gem with just the rollbar api in it @bhh feel free to open a PR with your solution, we'll be looking at this at soon as possible I have just added rollbar into a rails 2.3.18 LTS project (ruby 1.9.7, capistranio v3) - suggestions on how to improve it welcome. I disable the require in Gemfile so I can be sure it is initialised if it has been loaded. Gemfile: gem 'rollbar', require: false config/initializers/rollbar.rb: require 'rollbar' rollbar_enabled = (Rails.env.test? || Rails.env.development?) Rollbar.configure do |config| require 'rollbar' rollbar_enabled = (Rails.env.test? || Rails.env.development?) Rollbar.configure do |config| # Without configuration, Rollbar is enabled in all environments. # To disable in specific environments, set config.enabled=false. config.enabled = rollbar_enabled # ... etc etc ... end if rollbar_enabled require 'rollbar/rails23/notifier' class ActionController::Base include Rollbar::Rails23::Notifier def rescue_action_with_rollbar(exception) notify_rollbar_from_within_controller(exception) rescue_action_without_rollbar exception end alias_method_chain :rescue_action, :rollbar end end lib/rollbar/rails23/notifier.rb: module Rollbar module Rails23 module Notifier include Rollbar::ExceptionReporter include Rollbar::RequestDataExtractor def notify_rollbar_from_within_controller(exception) env = @_request.env request_data = extract_request_data_from_rack(env) request_data[:route] = { controller: params[:controller], action: params[:action], format: params[:format] } scope = { request: request_data, #context: { vertical: @vertical } # FIXME: Work out how to pass context } if current_user scope[:person]= {:id => current_user.id, :username => current_user.screenname, :email => current_user.email} scope[:person][:username] = current_user.screenname if !current_user.screenname.blank? && current_user.screenname != 'anonymous' end Rollbar.scoped(scope) do Rollbar.error(exception) end rescue Exception => exception2 # report if we stuffed up reporting with details Rollbar.error(exception) Rollbar.error(exception2) end end end end i'm so sorry i totally forgot about that topic. i think will look into it again upcoming week and maybe start a rollbar-rails project Two of my clients are on https://railslts.com/ - rails 2.3 LTS (and 3.2) is worthwhile supporting. Hey @ianheggie, have you taken a look at https://github.com/rollbar/rollbar-gem/blob/master/docs/plugins.md? It's the architecture for the plugins system. We are wondering if that rails 2.3 support could be rewriten using the new plugins interface. BTW, nice work so far on that. Thanks for the input @jondeandres - I will have a look at the plugin support - I have a few further things I need to do (missing action is being caught earlier in the process and thus is not reporting controller). Also in my project I am disabling router errors unless a user is logged in - I get way to many bots submitting mangled urls. Superb job @ianheggie! Did you eventually get a running solution? I also have a new old rails 2 app to maintain :) Yes, it works well enough (A few little niggles, but I am happy to share the code) - and re new/old rails 2 - I have two clients on 2.3.18 LTS - and I can see them staying on it because the cost to upgrade and keep upgrading would be too much. I'd love to see that final code if I may 😄 wondering if people are still interested in this. had some time to kill today and removed everything rails related as i suggested. as far as i can tell everything is still working. tests + some examples i tried. https://github.com/rubberandglue/rollbar-gem i copied every piece of removed code and tests into another repository (i guess this one will be a bit more tricky) i still missed out on the documentation for now. and i'm not sure if i did everything right for travis because rails tests aren't necessary anymore for this repository if the solution is accepted. there are still some "rails" strings but they should all be just string matching (not really rails related as far as i can see) i really like the solution so far. splitting everything up really contributes to nice replaceable modules. we wont be able to just turn rails support on/off with the new gem we also could make multiple versions for rails 2,3,4,5 hope you guys like it too I like the idea, but I'm still not sure about the benefit of doing a full split into two repos versus trying to do this via the plugin mechanism. I need to look at the code a bit more. I think the split into two repos is actually probably easier, just not sure if it is better from a maintenance perspective going forward. as i discovered the chronlogy is seems to be weird. thats why i came up with a solution i like most railtie loading getting config deciding to disable rails how not to load rails!? everything loaded/crashed already with the plugin mechanism and railtie you always have to load everything first then get the config and the decide to disable!? but the crash already happens during loading. why load it in the first place when i want to disable a specific part(not efficient anyway) and thats fine because i dont think it should be there anyway imo maintaining the rollbar part seperate from the rails part means you dont have to handle multiple rails versions in the rollbar plugin. no more RAILS_VERSION compare stuff and its even easier to handle different rails versions because you even can remove all the rails version if/switch statements and just use different rollbar-rails versions.
2025-04-01T04:35:22.617762
2017-01-03T17:47:35
198529245
{ "authors": [ "mrship", "solnic" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10401", "repo": "rom-rb/rom-sql", "url": "https://github.com/rom-rb/rom-sql/issues/123" }
gharchive/issue
Composing relations through many-to-many join We have a setup where we have a many-to-many relationship via a join table, e.g. module Relations class Questions < ROM::Relation[:sql] dataset :questions schema(infer: true) do associations do has_many :charts, through: :questions_charts end end end end module Relations class QuestionsCharts < ROM::Relation[:sql] dataset :questions_charts schema(infer: true) do associations do belongs_to :questions belongs_to :charts end end end end I can query this in my questions repository using a combine_children call to pull out all the charts and that works well, e.g. questions .combine_children( many: { charts: charts, } ) However, what I'd like to do is add an order attribute to the chart which is actually a column on the questions_tables join table, not on the chart table, alongside the attributes that come from the chart table itself, e.g. [{ id: 1, name: 'my chart', order: 1 }, { id: 2, name: 'my other chart', order: 2 }] ...where id and name are attributes of the chart table. At the moment the associations definition is happily pulling the charts back through that has_many through call, so I'm expecting to have to write a custom view in the questions_charts relation that does the same thing, but also includes the order attribute from the questions_tables table. However, I can't work out how to construct this view so that the join still works as it currently does automagically with the association setup. Could you please offer some advice on how to achieve that? Thanks in advance for your help. In rom-sql 1.0.0 you will be able to extend an association's relation with your own query logic, but for now you can define a custom view. To make it simpler for you, you can re-use association relation: class Charts < ROM::Relation[:sql] schema(infer: true) do associations do has_many :questions_charts has_many :questions, through: :questions_charts end end view(:for_questions, %[list of attributes in resulting relation]) do |questions| # not pretty, I know, there will be a nice API for that in 1.0.0 rel = associations[:questions].call(__registry__) rel.order(:questions__that_column_you_want_to_use) end end Combine will automatically use this view by convention, so it should just work. It'll be much simpler in 1.0.0 though. I'm releasing first beta in the upcoming days so stay tuned :) ps. just updated the example to be a bit nicer Thanks for the help; I probably confused things a little using order as my example attribute, but that is actually a column on the questions_tables join table, so I want to be able to select all the columns from chart as well as the order column from the join table. I'll see if I can figure that out from your example, but I'm still not 100% sure how to use the join_table as part of the for_combine. @mrship the association used under the hood joins that table for you automatically, so you can just do select_append(:questions_tables__some_col_1, :questions_tables__some_col_2). ps. this will be so much nicer in rom-sql 1.0.0 :) Thanks for the help so far and apologies for the delay in coming back to you but I'm only just coming back around to this. When I'm looking at the custom view that you wrote above, it appears to be selecting the question attributes, not chart attributes; e.g: if I strip this back to make Charts#for_questions just replicate the existing association (before adding in any other attributes) view(:for_questions, %i(id type created_at updated_at snapshot insight)) do |questions| q = for_combine(associations[:questions]) .where(questions__id: questions.pluck(:id)) require 'pry'; binding.pry q end q (above) produces this relation SQL: "SELECT `questions`.`id`, `questions`.`topic_id`, `questions`.`title`, `questions`.`short_title`, `questions`.`methodology`, `questions`.`order`, `questions`.`created_at`, `questions`.`updated_at`, `chart_id` FROM `questions` INNER JOIN `questions_charts` ON (`questions_charts`.`question_id` = `questions`.`id`) INNER JOIN `charts` ON (`charts`.`id` = `questions_charts`.`chart_id`) ORDER BY `questions`.`id`" implemented like that, I then end up with this error: ArgumentError: Relations::Questions#preload arity is 3 (2 args given) when calling the combine_children (note I also have to call the view directly as it doesn't work automatically as you implied) questions .combine_children( many: { charts: charts.for_questions, } ) So, sorry the news isn't better but I think I'm missing something fundamental here. Unfortunately, the view doesn't get called automatically if I combine_children(many: charts) only if I am explicit about it with combine_children(many: questions.for_charts). @mrship oh right I forgot that it needs keys to passed explicitly, damn :) see docs for #combine where it's explained Ok, you mean something like, this? questions.combine(many: { charts: [questions.for_charts, id: :chart_id] }) unfortunately, that gives a similar error but on charts. ArgumentError: Relations::Charts#preload arity is 3 (2 args given) Apologies if I'm being dense here; I feel like I'm so close. I thought you wanted questions.combine(many: { charts: [charts.for_questions, id: question_id])? I've created a gist - https://gist.github.com/mrship/3f59b5ab6c40e7cacc0a03eba92742d0 - that shows the problem and hopefully makes it a little clearer what is happening. The ultimate aim is to have a Question with many Charts, and the Charts have their order from the join table. If you run the gist, you'll get the same arity issue that I reported above. Thanks again for the extensive help. That's very helpful. I'll check it out and get back to you soon Cheers Piotr On 13 Jan 2017, 7:59 PM +0100, Andy Shipman<EMAIL_ADDRESS>wrote: I've created a gist - https://gist.github.com/mrship/3f59b5ab6c40e7cacc0a03eba92742d0 - that shows the problem and hopefully makes it a little clearer what is happening. The ultimate aim is to have a Question with many Charts, and the Charts have their order from the join table. If you run the gist, you'll get the same arity issue that I reported above. Thanks again for the extensive help. — You are receiving this because you commented. Reply to this email directly, view it on GitHub (https://github.com/rom-rb/rom-sql/issues/123#issuecomment-272518667), or mute the thread (https://github.com/notifications/unsubscribe-auth/AAAEKtKYsI_P2qBepmyX8AW1vqjXEqgYks5rR8mcgaJpZM4LZ2sT). @mrship pheeew, there we go: https://gist.github.com/46584d53fb1f6901b3d29381cec1ccd1 Wow! That works a treat. There's no way I'd ever have got to that point without your help though, so (a) thanks and (b) I'm looking forward to seeing how that becomes easier in v1.0.0! The good news is that this approach is going to get used a lot throughout my current project, and simplify a lot of features, so thanks again for the help. @mrship cool! I'll let you know once there's a nicer way, in 1.0.0.beta there's already support for defining associations extended with custom views (ie has_many :questions, view: :ordered) and the only missing part is having ability to extend views's schema with attributes from other relations (this order col in your case). @mrship hey man, I added a couple of features to make this easier for you. Here's an updated gist: https://gist.github.com/solnic/46584d53fb1f6901b3d29381cec1ccd1 This will be available in rom 3.0.0 and rom-sql 1.0.0 That looks really nice. Thanks for following up. I hope to try 1.0.0 when it comes out of beta.
2025-04-01T04:35:22.626921
2023-04-01T19:16:58
1650608775
{ "authors": [ "Abashinos", "mike-pisman", "roman-right" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10402", "repo": "roman-right/beanie", "url": "https://github.com/roman-right/beanie/issues/522" }
gharchive/issue
Design the back-ref interfaces Interfaces: back refs attributes and ways to map fetching Attributes Design: from beanie import Document, Link, BackLink class Door(Document): height: int width: int house: BackLink["House"] # Back-ref class House(Document): door: Link[Door] # Original reference Seems useful for model ops decoupling Hi, thank you for this feature! Could you please clarify the use of it? Am I correct to understand that the BackLink will be populated automatically after creating the link? I am trying to create a BackLink, but there is no data in the mongo document after creation(besides name, etc). I assume, there should be some metadata like id and document type, right? Here is a simplified example of what I am trying to do: class Resource(Document): id: ResourceID = Field(default_factory=ResourceID, alias="_id") resource_type = "" name: str = Field(title="Name", description="Name of the resource", min_length=3, max_length=16) groups: list[Link["Group"]] = [] class Workspace(Resource): resource_type = "workspace" class Group(Resource): resource_type = "group" # workspace: Link[Workspace] # I was using Link before, but decided to try BackLink workspace: BackLink[Workspace] = Field(original_field="groups") # Assuming I already created a workspace def create_group(workspace: Workspace): new_group = Group(name=input_data.name) workspace.groups.append(new_group) await Group.insert(new_group, link_rule=WriteRules.WRITE) await Workspace.save(workspace, link_rule=WriteRules.WRITE) Hi @mike-pisman , No, it doesn't store any additional data in the Mongo document. I use the lookup aggregation pipeline to find documents in the collection where the links were created and check if the original_field there has the id or ids of the documents in the collection with documents that have backlinks. So everything is happening on the db layer but without additional data to be stored there. The feature was implemented
2025-04-01T04:35:22.643754
2015-04-10T19:58:58
67670283
{ "authors": [ "andrewcstewart", "romansanchez" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10403", "repo": "romansanchez/Calaca", "url": "https://github.com/romansanchez/Calaca/issues/13" }
gharchive/issue
Unexpected CSS token: } I keep getting the following in the debugger. Dunno it's a big deal but thought I'd report. Unexpected CSS token: } calaca.css:115 What browser/os are you using? On Friday, April 10, 2015, Andrew Stewart<EMAIL_ADDRESS>wrote: I keep getting the following in the debugger. Dunno it's a big deal but thought I'd report. Unexpected CSS token: } calaca.css:115 — Reply to this email directly or view it on GitHub https://github.com/romansanchez/Calaca/issues/13. Safari 8.0.4 If there’s anything in my session info that’d be helpful let me know. On Fri, Apr 10, 2015 at 1:22 PM, Roman Sanchez<EMAIL_ADDRESS>wrote: What browser/os are you using? On Friday, April 10, 2015, Andrew Stewart<EMAIL_ADDRESS>wrote: I keep getting the following in the debugger. Dunno it's a big deal but thought I'd report. Unexpected CSS token: } calaca.css:115 — Reply to this email directly or view it on GitHub https://github.com/romansanchez/Calaca/issues/13. Reply to this email directly or view it on GitHub: https://github.com/romansanchez/Calaca/issues/13#issuecomment-91675903 Yeah line 114 had some extra chars. Fixed it. Thanks.
2025-04-01T04:35:22.658447
2020-06-16T18:49:38
639897708
{ "authors": [ "electrodude", "emagnier", "tukusejssirs" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10404", "repo": "romcal/romcal", "url": "https://github.com/romcal/romcal/issues/193" }
gharchive/issue
Make GitHub description say e.g. "Roman Rite of the Catholic Church" instead of "Catholic Rite" The current GitHub project description is: Generates the General Roman Calendar used in the Catholic Rite. There is no such thing as the "Catholic Rite". All of the rites used within the Catholic Church -- the Roman Rite, a few variations on Latin Rite such as the Ambrosian and Dominican Rites, and all of eastern rites -- are Catholic rites, and many of them have their own calendars. It should instead say something like "the Roman Rite of the Catholic Church". What do you think @pejulian ? I don't have the right to edit this headline. Thank you! It's now updated to: Generates the liturgical calendar of the Catholic Church used by the Roman Rite. And let me know if the wording could be improved, English is not my primary language 😉 @emagnier, how about to get inspired by Wiki: GRC article short description? Liturgical calendar of Saints' Days for the ordinary form of the Roman Rite (Roman Catholic Church) Like: Generates liturgical calendars of the Roman Rite of the Roman Catholic Church. I think calendars should plural, as we generate different cals by country or language. I personally wouldn’t include ordinary form. We also might place Roman Catholic Church in brackets or even remove it altogether, but for people less learned in the liturgy, it might be useful to keep it. And in English, there is no such thing as implicit/silent subject, therefore it sounds strange and is grammatically incorrect without romcal at the the beginning. I know in code we do it that way a llt, but in user-facing docs I don’t think it is a valid reason to remove the subject. That said I let you decide on this. :wink: What do you think? Generates liturgical calendars of the Roman Rite of the Roman Catholic Church. This is a good proposition. I've updated the repo description. We still have to update the headline in the main readme.md We still have to update the headline in the main readme.md Should I create a PR for that? Yes please 😄
2025-04-01T04:35:22.667835
2020-07-27T23:50:14
666654109
{ "authors": [ "ematipico", "ljharb", "mbrookes", "milesj", "sebmck" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10405", "repo": "romefrontend/rome", "url": "https://github.com/romefrontend/rome/issues/893" }
gharchive/issue
Default project config location I'm writing the docs and right now we recommend an .rjson file for the project config. I'm not totally sold on this and want us to consider the following two options, if anything so we can rule them out: Defaulting to package.json Force rome.r?json files to be inside of a .config directory. Many tools support this although I would propose forcing it if we decide to support it. We can pave the cowpath for the JS ecosystem. Main motivation for this is that the root of projects has become extremely cluttered for most projects. Scanning project structure is difficult since you need to mentally filter out all the dot files. Just chiming in with my support for placing config in .config. I did this recently for a personal project of mine (https://www.npmjs.com/package/@boost/config) and I believe it to be a good standard for JS projects going forward. Will help to reduce project root clutter. As long as the config isn't directory-specific (in which case, a config file directly in the directory makes the most sense), that would work fine. However, it might be nice to have config able to live in two places: either directly in package.json, or, in a path that might default to .config/rome.rjson. That way users get to choose for themselves where they want it to live. Right now the possible locations are: rome field in package.json rome.json and rome.rjson in the project base I'm proposing it be changed to: rome field in package.json .config/rome.json and .config/rome.rjson in the project base We automatically discover the project config so it has to be in one of these hardcoded locations. Although with extends in package.json you could put the config pretty much anywhere: package.json { "rome": { "extends": "some-other-page/config.json" } } I like the idea of having .config/rome.json/.config/rome.rjson; the name of the folder tells what's that about. I would discourage the use of the package.json because some projects would already have a big file and this would increase the size and lower the readability. 🎉 You heard it here first I even tagged @babeljs! 😉 (I wish I could actually take credit for this being added to Rome, but glad to see it happen. I hope it paves the way for more tools to follow suit!)
2025-04-01T04:35:22.702554
2020-05-07T08:23:24
613868657
{ "authors": [ "MaxDeg", "ronaldschlenker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10406", "repo": "ronaldschlenker/FsHttp", "url": "https://github.com/ronaldschlenker/FsHttp/pull/44" }
gharchive/pull-request
Add proxy setting This PR is related to #4 I propose a way to support proxy (with and without credentials) for request. Thanks again for the PR! I hope to finalize this and some other things by the end of the week...
2025-04-01T04:35:22.758163
2022-11-01T22:56:39
1432210354
{ "authors": [ "roosterkid" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10407", "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/10655" }
gharchive/issue
⚠️ TROJAN 🇺🇸 United States USO 1 has degraded performance In c2fffbf, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) experienced degraded performance: HTTP code: 200 Response time: 17372 ms Resolved: TROJAN 🇺🇸 United States USO 1 performance has improved in c936cd6.
2025-04-01T04:35:22.829635
2016-07-14T16:17:27
165601215
{ "authors": [ "aviadhahami", "nemosdo", "redflasher", "walidvb", "zjjt" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10418", "repo": "root-two/react-native-drawer", "url": "https://github.com/root-two/react-native-drawer/issues/191" }
gharchive/issue
Drawer not closing properly Hey, i'm getting issues with the drawer: with static positioning: the drawer first is displayed on top of the other views, and when closed, the content overflows: with overlay: initially the drawer is displayed correctly, but once open, it won't close to the end. my code is as follows: <Drawer type="static" content={<CategoriesList {...this.props} />} openDrawerOffset={160} closedDrawerOffset={0} onOpen={() => this.props.toggleFilters(true)} onClose={() => this.props.toggleFilters(false)} styles={{ drawer:{ backgroundColor: 'white', marginTop: 60, overflow: 'hidden' } }} open={this.props.ui.filters.drawerOpen} tweenHandler={Drawer.tweenPresets.parallax} tapToClose={true} > <MenuBar {...this.props} style={{paddingTop: 20}} /> <PartnerList partners={this.state.partners} style={{ flex: 1,paddingLeft: 15, paddingRight: 15, marginBottom: sub.isValid() ? 0 : 60, }}/> <GetPass float/> </Drawer> Versions: "react-native": "^0.29.0" "react-native-drawer": "^2.2.6" Link to the code If it makes a difference, this is wrapped within a Router of react-native-router-flux Any help/hints greatly appreciated, and thank you for the hard work! :+1: for the overlay option About the first case(the static one), it looks like I had to add a background color to the main component, which fixed it. Overlay is still not closing properly, though. Hi there, i also have the same issue with overlay...It isnt closing properly any updates on that topic might help us out a great deal. thanks and cheers have the same issue with overlay it works fine for me when i remove the line 'tweenHandler={Drawer.tweenPresets.parallax}' ,but i still didn't know why @walidvb @zjjt yes, when ''tweenHandler={Drawer.tweenPresets.parallax}' was removed drawer work good. Cause of problem may be wrong parameter for tween (forget about 'drawer.width' or something else) i didnt check it out recently as i use the drawer component for android,if you think that is the tween handler which creates the issue ill check it out again and get back to you @nemosdo nemosdo...
2025-04-01T04:35:22.847384
2020-11-16T14:37:41
743888194
{ "authors": [ "Log1x", "TangRufus", "albertonieto", "chesio", "koraysels" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10419", "repo": "roots/bedrock", "url": "https://github.com/roots/bedrock/issues/559" }
gharchive/issue
oscarotero/env v2.1.0 Class 'Env' not found [X] I've read the guidelines for Contributing to Roots Projects [X] This request isn't a duplicate of an existing issue [X] I've read the docs and followed them (if applicable) [X] This is not a personal support request that should be posted on the Roots Discourse community Description I have recently updated composer and got the oscarotero/env v2.1.0 package. Since then I keep getting this error: Fatal error: Uncaught Error: Class 'Env' not found in /myproject/config/application.php:6 Steps to reproduce Update composer to v2.1.0 on package oscarotero/env Check your local repository Expected behavior: [What you expect to happen] Expected to run site normally. Actual behavior: [What actually happens] Throws error Fatal error: Uncaught Error: Class 'Env' not found in /myproject/config/application.php:6 Reproduces how often: [What percentage of the time does it reproduce?] 100% Versions oscarotero/env v2.1.0 Additional information This is the content of the referenced file, where Env::init() corresponds to line 6: <?php $root_dir = dirname(__DIR__); $webroot_dir = $root_dir . '/public'; Env::init(); // Use Dotenv to set required environment variables and load .env file in root $dotenv = new Dotenv\Dotenv($root_dir); if (file_exists($root_dir . '/.env')) { $dotenv->load(); try { $dotenv->required('DATABASE_URL')->notEmpty(); } catch (Exception $e) { $dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']); } } It seems as if the package were not loaded since the Env Class doesn't load. Ideas? Patch Bedrock with the https://github.com/roots/bedrock/tree/master/ branch, https://github.com/roots/bedrock/pull/530/ to be specific. Patch Bedrock with the master branch, #530 to be specific. Sounds good! How can I do that? Download https://patch-diff.githubusercontent.com/raw/roots/bedrock/pull/530.diff and $ git apply it. However, depends on how much change you made and how "out of sync" your code is, it might not work. For a small patch like #530, manually copy and paste the changes might be easier. Props to @austinpray Thanks, that was helpful. I have indeed manually added the missing parts. This is how the edited file looks like now: <?php use function Env\env; // Directory containing all of the site's files $root_dir = dirname(__DIR__); $webroot_dir = $root_dir . '/public'; // Use Dotenv to set required environment variables and load .env file in root $dotenv = Dotenv\Dotenv::createImmutable($root_dir); if (file_exists($root_dir . '/.env')) { $dotenv->load(); **var_dump(env('DB_USER'));exit;** try { $dotenv->required('DATABASE_URL')->notEmpty(); } catch (Exception $e) { $dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']); } } The problem I am facing is that if I var_dump the env variables I get empty values. Because of that, my msqli is not connecting to the database. Any idea? @albertonieto What version of vlucas/phpdotenv are you using? I have to update the bootstrap code when I've upgraded from version 4 to 5. Also see this issue: https://github.com/oscarotero/env/issues/8 I ended adding this to the top of the bootstrap file: \Env\Env::$options |= \Env\Env::USE_ENV_ARRAY; I am using the latest version "vlucas/phpdotenv": "^5.2", Sorry, when you mention the bootstrap file, which file are you referring to? I would like to try that as well. Sorry, I meant config/application.php - it now looks like this in my Bedrock-based project (the top part): <?php /** * Expose global env() function from oscarotero/env */ use function Env\env; /** * Instruct Env\Env to read environment variables via $_ENV instead of getenv(). * The latter does not work with DotEnv 5 by default, see: https://github.com/oscarotero/env/issues/8 */ \Env\Env::$options |= \Env\Env::USE_ENV_ARRAY; /** @var string Directory containing all of the site's files */ $root_dir = dirname(__DIR__); /** @var string Document Root */ $webroot_dir = $root_dir . '/web'; /** * Use Dotenv to set required environment variables and load .env file in root */ $dotenv = \Dotenv\Dotenv::createImmutable($root_dir); if (file_exists($root_dir . '/.env')) { $dotenv->load(); $dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']); } ... Thanks. I have added the instruction as you referred. I am able to see the env variables when I var_dump them, however my mysqli is still not connecting: <?php use Roots\WPConfig\Config; use function Env\env; \Env\Env::$options |= \Env\Env::USE_ENV_ARRAY; $root_dir = dirname(__DIR__); $webroot_dir = $root_dir . '/public'; $dotenv = \Dotenv\Dotenv::createImmutable($root_dir); if (file_exists($root_dir . '/.env')) { $dotenv->load(); Config::define('DB_NAME', env('DB_NAME')); Config::define('DB_USER', env('DB_USER')); Config::define('DB_PASSWORD', env('DB_PASSWORD')); Config::define('WP_HOME', env('WP_HOME')); Config::define('WP_SITEURL', env('WP_SITEURL')); var_dump(env('DB_NAME')); var_dump(env('DB_USER')); var_dump(env('DB_PASSWORD')); var_dump(env('WP_HOME')); var_dump(env('WP_SITEURL')); $dotenv->required(['DB_NAME', 'DB_USER', 'DB_PASSWORD', 'WP_HOME', 'WP_SITEURL']); } I still get the: Warning: mysqli_real_connect(): (HY000/2002): No such file or directory in /myproject/public/cms/wp-includes/wp-db.php on line 1635 No such file or directory Error establishing a database connection I am very confused. @albertonieto Are you sure you have Config::apply(); at the end of the file as well? See: https://github.com/roots/bedrock/blob/master/config/application.php#L123 AFAIK, this is where the actual constants are defined. Config::define(...); calls are just part of internal setup, but do not actually define any constants. Sorry about that, I was missing the apply at the end of the file. I have managed to actually connect via mysli now. I noticed however that our theme is using a /public/content/themes/mytheme/inc/assets.php file that reads as follows: <?php define('ASSETS_REVISION_FILE', dirname(dirname(ABSPATH)) . '/REVISION'); define('ASSETS_REVISION', substr(trim(@file_get_contents(ASSETS_REVISION_FILE)), 0, 12)); define('ASSETS_PATH', dirname(dirname(ABSPATH)) . DIRECTORY_SEPARATOR . env('ASSETS_COMPILE')); define('ASSETS_DIRECTORY', str_replace(dirname(ABSPATH), '', ASSETS_PATH)); define('ASSETS_HOST', env('ASSETS_HOST') ? env('ASSETS_HOST') : '//' . $_SERVER['SERVER_NAME']); //define('ASSETS_VERSION', ASSETS_REVISION ? ASSETS_REVISION : time('U')); define('ASSETS_VERSION', ASSETS_REVISION ? ASSETS_REVISION : time()); function asset_path($path) { return ASSETS_HOST . ASSETS_DIRECTORY . '/' . $path . '?v=' . ASSETS_VERSION; } I am getting a Call to undefined function env() for line 5. Any idea for that fix? I am getting a Call to undefined function env() for line 5. Any idea for that fix? You have to add use function Env\env; to this file as well - the use statements must be defined per file, see the docs. @chesio Nice touch about vlucas/phpdotenv v5! Do you mind make a PR for it? @TangRufus Sure, I'll try to get a PR ready over the coming weekend. SO how do we fix this ? I am stuck here aswell.. SO how do we fix this ? I am stuck here aswell.. Run composer require vlucas/phpdotenv:^5.2 oscarotero/env:^2.1 Update your application.php to reflect https://github.com/roots/bedrock/blob/master/config/application.php#L31
2025-04-01T04:35:22.858112
2021-11-05T16:06:38
1046010598
{ "authors": [ "klnit", "openoms" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10420", "repo": "rootzoll/raspiblitz", "url": "https://github.com/rootzoll/raspiblitz/issues/2701" }
gharchive/issue
(UNKNOWN) [<IP_ADDRESS>] 50001 (?) : Connection refused after upgrade to 1.7.1 I cannot connect to electrs (at first it did a new sync and I thougt that was the reason, but now after rebooot still) STATUS ELECTRS SERVICE configured=1 serviceInstalled=1 serviceRunning=1 syncProgress=100.00% tipSynced=1 (UNKNOWN) [<IP_ADDRESS>] 50001 (?) : Connection refused electrumResponding=0 initialSynced=1 localIP='<IP_ADDRESS>' portTCP='50001' localTCPPortActive=0 publicTCPPortAnswering=0 portSSL='50002' localHTTPPortActive=1 publicHTTPPortAnswering=0 TorRunning=1 nginxTest=1 ● electrs.service - Electrs Loaded: loaded (/etc/systemd/system/electrs.service; enabled; vendor preset: enabled) Active: active (running) since Fri 2021-11-05 15:45:30 GMT; 12min ago Main PID: 6245 (electrs) Tasks: 16 (limit: 4915) CGroup: /system.slice/electrs.service └─6245 /home/electrs/electrs/target/release/electrs --electrum-rpc-addr=<IP_ADDRESS>:50001 Nov 05 15:45:30 raspberrypi electrs[6245]: Config { network: Bitcoin, db_path: "/mnt/hdd/app-storage/electrs/db/bitcoin", daemon_dir: "/home/electrs/.bitcoin", daemon_auth: UserPass("raspibolt Nov 05 15:45:30 raspberrypi electrs[6245]: [2021-11-05T15:45:30.141Z INFO electrs::metrics::metrics_impl] serving Prometheus metrics on <IP_ADDRESS>:4224 Nov 05 15:45:30 raspberrypi electrs[6245]: [2021-11-05T15:45:30.838Z INFO electrs::db] "/mnt/hdd/app-storage/electrs/db/bitcoin": 319 SST files, 39.025542469 GB, 4.027567841 Grows Nov 05 15:45:35 raspberrypi electrs[6245]: [2021-11-05T15:45:35.493Z INFO electrs::chain] loading 708330 headers, tip=0000000000000000000b93efaa186fe13b8f96eececa5560021ea2036a42ab6c Nov 05 15:45:39 raspberrypi electrs[6245]: [2021-11-05T15:45:39.586Z INFO electrs::chain] chain updated: tip=0000000000000000000b93efaa186fe13b8f96eececa5560021ea2036a42ab6c, height=708330 Nov 05 15:45:39 raspberrypi electrs[6245]: [2021-11-05T15:45:39.630Z INFO electrs::index] indexing 12 blocks: [708331..708342] Nov 05 15:45:40 raspberrypi electrs[6245]: [2021-11-05T15:45:40.626Z INFO electrs::chain] chain updated: tip=0000000000000000000531ede5e077d5595576928cb1ab3cc404ff6e24a5c2db, height=708342 Nov 05 15:45:40 raspberrypi electrs[6245]: [2021-11-05T15:45:40.964Z INFO electrs::db] starting config compaction Nov 05 15:45:40 raspberrypi electrs[6245]: [2021-11-05T15:45:40.987Z INFO electrs::db] starting headers compaction Nov 05 15:45:45 raspberrypi electrs[6245]: [2021-11-05T15:45:45.016Z INFO electrs::db] starting txid compaction admin@raspberrypi:~ $ sudo -u electrs lsof -i COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME electrs 6245 electrs 4u IPv4 1515489 0t0 TCP localhost:xtell (LISTEN) electrs 6245 electrs 15u IPv4 1521738 0t0 TCP localhost:36664->localhost:8333 (ESTABLISHED) electrs 6245 electrs 16u IPv4 1521738 0t0 TCP localhost:36664->localhost:8333 (ESTABLISHED) Any idea? starting txid compaction Electrs is finishing reindexing after the update. Just need too wait.
2025-04-01T04:35:22.860350
2023-01-17T14:04:28
1536435411
{ "authors": [ "davebhc71", "roovo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10421", "repo": "roovo/obsidian-card-board", "url": "https://github.com/roovo/obsidian-card-board/issues/131" }
gharchive/issue
UI option to change color indicators in Dated board Would like to change colors for Past and Current items. Orange and Red are too similar!😉 thanks btw :) Just to follow up on this FR - Green for Today would be nice. And, I’m not asking for the User to be able to change the color, but that you consider a ‘universal’ change for Today to Green in the next update. 😉 Just realised that you should be able to use a css snippet along these lines to customise the colours: .card-board-view .card-board-card-highlight-area.critical { background-color: red; } .card-board-view .card-board-card-highlight-area.good { background-color: green; } .card-board-view .card-board-card-highlight-area.important { background-color: orange; } I have updated the readme with info on doing this and have just released a related feature: via snippets, you can set the color indicators based on tags too.
2025-04-01T04:35:23.006732
2020-03-10T14:03:12
578602622
{ "authors": [ "gavanderhoorn", "jontje" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10422", "repo": "ros-industrial/abb_librws", "url": "https://github.com/ros-industrial/abb_librws/issues/80" }
gharchive/issue
Python wrapper? Being able to use RWS from a Python application would be perfect. I've titled this issue "Python wrapper" as that could of course be one way to do this. It may however complicate things needlessly. Searching around, I've been able to find some mentions of using requests to interface with RWS, but I haven't been able to find any proper Python package. As RWS purports to be a REST based interface: @jontje would you happen to know whether RWS comes with an API spec (OpenAPI, Swagger, RADL, RAML, RSDL, something else) which would make it possible auto-generate (Python) wrappers or runtime code generate them? I'm not entirely up-to-date on REST and web APIs, but there are definitely more advanced (and convenient) ways of consuming a REST interface than Python requests and composing raw URIs. As RWS purports to be a REST based interface: @jontje would you happen to know whether RWS comes with an API spec (OpenAPI, Swagger, RADL, RAML, RSDL, something else) which would make it possible auto-generate (Python) wrappers or runtime code generate them? RWS comes in two versions: RWS 1.0 (RobotWare 6) doesn't have any such API specs (as far as I know). RWS 2.0 (RobotWare 7) is compliant with OpenAPI and Swagger can be used to generate client code, but I have not worked with RobotWare 7 systems yet. Ah, great. Is abb_libegm RW7 compatible? Would there be a way to make sure? :) Of course there are ;) I can check with a colleague tomorrow, but I am quite certain there aren't any. Is abb_libegm RW7 compatible? Did you mean abb_libegm or abb_librws? abb_librws: Unfortunately not compatible, since RWS 1.0 has been deprecated in RW7. And I haven't really started thinking about a good approach to solved this yet, mainly since I haven't worked with RW7 systems. abb_libegm: Should be compatible with RW7, but I haven't had time to test this yet. Hm. At least for RW7 we should be able to use something like OpenAPITools/openapi-generator. @gavanderhoorn, FYI, I got a no regarding auto-generation of client code for RWS 1.0. Too bad, but thanks for asking. Closing this now.
2025-04-01T04:35:23.018476
2019-01-18T20:40:06
400881154
{ "authors": [ "christian-rauch", "jproberge" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10423", "repo": "ros-industrial/robotiq", "url": "https://github.com/ros-industrial/robotiq/pull/142" }
gharchive/pull-request
Mesh colour This PR adds colour (black) to the articulated meshes. Fixes #141 Edit: I just saw that I had other commits in my devel branch. There are two commits to fix the installation and running of the simulation launch file. @jproberge Is there any news on an "official" colour/material definition from Robotiq? If this is not happening in the short term, I am happy if you change the colour to something slightly lighter. Maybe the issue is also with the material definition and not the colour per se. In Blender, the palm mesh shows more shading than in Gazebo: hi @christian-rauch , Thanks for your follow-up! I will have the information this week and I'll post back as soon as possible. Also, I agree that it depends on the visualisation environment. At the end, it's just aesthetic, so I don't want to delay this PR too much, and I'll post back here this week as soon as I get the information. Thanks again! jproberge @christian-rauch, I'm accepting this PR, since I've reviewed / tested it and found that it works flawlessly and is an improvement. Also, Robotiq told me that there were no precise color definition. In fact they ask their supplier to provide black anodized parts, but then these parts are also subject to a tumbler surface finish process, which might also alter the colour a little. Thus no precise colour definition. Thanks!
2025-04-01T04:35:23.071624
2016-05-02T12:44:06
152545650
{ "authors": [ "dirk-thomas", "harvcode", "tfoote" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10424", "repo": "ros-visualization/python_qt_binding", "url": "https://github.com/ros-visualization/python_qt_binding/issues/38" }
gharchive/issue
kinetic version breaks rqt_graph command ran: rosrun rqt_graph rqt_graph error: File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui/ros_plugin_provider.py", line 83, in load raise e ImportError: cannot import name QGraphicsPathItem error originates from here: File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_dotgraph/edge_item.py", line 32, in <module> from python_qt_binding.QtGui import QBrush, QGraphicsPathItem, QGraphicsPolygonItem, QGraphicsSimpleTextItem, QPainterPath, QPen, QPolygonF ImportError: cannot import name QGraphicsPathItem It seems like you are using an older version of qt_dotgraph. Version 0.3.2 imports that class from QWidgets: https://github.com/ros-visualization/qt_gui_core/blob/kinetic-devel/qt_dotgraph/src/qt_dotgraph/edge_item.py#L33 I am using the version that gets installed automatically from the kinetic binaries, so perhaps the repository binaries need to be changed to reflect this? Please state the exact version number of qt_dotgraph as well as rqt_graph you are using. @harvcode The versions available from binary have changed recently. That's why we need to confirm the version on your system. ros-kinetic-qt-dotgraph/xenial,now 0.3.2-0xenial-20160428-192006-0700 amd64 ros-kinetic-rqt-graph/xenial,now 0.4.0-1xenial-20160429-165937-0700 amd64 The code in the error message in your original comment (edge_item.py, line 32) does not match the code available in these versions. Version 0.3.0 has the line you quoted above: https://github.com/ros-visualization/qt_gui_core/blob/0.3.0/qt_dotgraph/src/qt_dotgraph/edge_item.py#L32 Version 0.3.1 as well as 0.3.2 have this instead which should work: https://github.com/ros-visualization/qt_gui_core/blob/0.3.1/qt_dotgraph/src/qt_dotgraph/edge_item.py#L32 https://github.com/ros-visualization/qt_gui_core/blob/0.3.2/qt_dotgraph/src/qt_dotgraph/edge_item.py#L32 Just tried it again and I still get "ImportError: cannot import name QFormLayout". Apt says I am up to date on both of those packages and lists those versions. Since the error is different then what you reported above please post the complete error message and steps you did to reproduce. From command line: rqt From menu: Plugins > Robot Tools > Controller Manager Error appears in console. Please post the full error message. ` PluginManager._load_plugin() could not load plugin "rqt_controller_manager/ControllerManager": Traceback (most recent call last): File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/plugin_handler.py", line 99, in load self._load() File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/plugin_handler_direct.py", line 54, in _load self._plugin = self._plugin_provider.load(self._instance_id.plugin_id, self._context) File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/composite_plugin_provider.py", line 71, in load instance = plugin_provider.load(plugin_id, plugin_context) File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/composite_plugin_provider.py", line 71, in load instance = plugin_provider.load(plugin_id, plugin_context) File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui_py/ros_py_plugin_provider.py", line 60, in load return super(RosPyPluginProvider, self).load(plugin_id, plugin_context) File "/opt/ros/kinetic/lib/python2.7/dist-packages/qt_gui/composite_plugin_provider.py", line 71, in load instance = plugin_provider.load(plugin_id, plugin_context) File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui/ros_plugin_provider.py", line 83, in load raise e ImportError: cannot import name QFormLayout RosPluginProvider.load(rqt_controller_manager/ControllerManager) exception raised in builtin.import(rqt_controller_manager.controller_manager, [ControllerManager]): Traceback (most recent call last): File "/opt/ros/kinetic/lib/python2.7/dist-packages/rqt_gui/ros_plugin_provider.py", line 77, in load module = builtin.import(attributes['module_name'], fromlist=[attributes['class_from_class_type']], level=0) File "/root/catkin_ws/src/ros_control/rqt_controller_manager/src/rqt_controller_manager/controller_manager.py", line 35, in from python_qt_binding.QtGui import QCursor, QFont, QFormLayout, QHeaderView, ImportError: cannot import name QFormLayout` The stack trace indicates that the problem is in the package rqt_controller_manager which is in your local workspace (I assume it hasn't been released into Kinetic yet?). You might want to fill a ticket in the repository which contains that package: https://github.com/ros-controls/ros_control/issues
2025-04-01T04:35:23.081395
2022-07-27T07:58:37
1319169258
{ "authors": [ "rhaschke", "ysl-design" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10425", "repo": "ros-visualization/rviz", "url": "https://github.com/ros-visualization/rviz/issues/1758" }
gharchive/issue
The plug-in in the display panel sometimes does not shrink as the display panel compresses. Describe your issue here and explain how to reproduce it. Your environment OS Version: Ubuntu 20.04 ROS Distro: Noetic RViz, Qt, OGRE, OpenGl version as printed by rviz:[ INFO] [1658890616.733419800]: rviz version 1.14.14 [ INFO] [1658890616.733471700]: compiled against Qt version 5.12.8 [ INFO] [1658890616.733501700]: compiled against OGRE version 1.9.0 (Ghadamon) [ INFO] [1658890616.741053500]: Forcing OpenGl version 0. [ INFO] [1658890616.894986100]: Stereo is NOT SUPPORTED [ INFO] [1658890616.895090400]: OpenGL device: llvmpipe (LLVM 12.0.0, 256 bits) [ INFO] [1658890616.895148600]: OpenGl version: 3.1 (GLSL 1.4). System locale, zh_CN.UTF-8 When I stretch the displays panel first, then compress it, and then double-click the plug-in component in the displays panel (such as the drop-down list), the picture below appears. It seems that the components don't adapt to size as the displays panel compresses. I do not know whether the problem is caused by the QT or the displays panel. I cannot reproduce this issue. Could you please provide more detailed instructions and a video as well? https://user-images.githubusercontent.com/5376030/181205794-3ee24b7f-cc94-46ae-82f4-a135c89b42a6.mp4 Sorry, I can't upload the video right now. The procedure is as follows: step1: step2: step3: result: Thanks for the clarification. Now, I can reproduce the issue. Could you please test https://github.com/ros-visualization/rviz/pull/1760? I tried #1760 and the problem did not recur. Closed via #1760.
2025-04-01T04:35:23.083947
2020-12-15T08:23:43
767314141
{ "authors": [ "stertingen" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10426", "repo": "ros/catkin", "url": "https://github.com/ros/catkin/pull/1127" }
gharchive/pull-request
Added --force-color flag to catkin_make and catkin_make_isolated This can be used to enforce colored output even on non-interactive output. --force-color can still be overriden by --no-color. This PR is for CI environments where the output is stored in a file but a colored output makes it easier for a user to read the log. There are some alternatives to the current implementation: --force-color overrides --no-color Error/Warning if --force-color and --no-color are both set
2025-04-01T04:35:23.095033
2015-07-10T10:06:29
94271157
{ "authors": [ "asmodehn", "dirk-thomas" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10427", "repo": "ros/catkin", "url": "https://github.com/ros/catkin/pull/747" }
gharchive/pull-request
fixing package_dir location logic. package_dir: {'foo':'lib'} means 'lib/init.py' and not 'lib/foo/init.py' It seems some tests should also be changed... I d rather have some feedback about this pull request before diving into all the tests that might need to be changed to follow the same behavior as the distutils behavior ( that you can test with python setup.py install ) so that our devel/ and install/ folder endup being similar. references : https://docs.python.org/2/distutils/setupscript.html#listing-whole-packages https://docs.python.org/2/distutils/examples.html#pure-python-distribution-by-package Can you please provide some more information what the current problem is as well as step-by-step instructions how to reproduce the problem. Current problem : If your files structure to access your python modules is : python-mod-name/setup.py python-mod-name/python_module_1/__init__.py python-mod-name-2/setup.py python-mod-name-2/python_module_2/__init__.py which enables you to have submodules pointing to the source repository of a package by the way, instead of the usual : src/python_module_1/__init__.py src/python_module_2/__init__.py and if you have in your setup.py : d = generate_distutils_setup( packages=[ 'python_module_1', 'python_module_2', ], package_dir={ 'python_module_1': 'python-mod-name', 'python_module_2': 'python-mod-name-2', }) the catkin_make build works fine. The __init__.py is found in devel/lib/python2.7/dist-packages/python_module. However catkin_make install doesn't install the package properly. instead the setup.py file ends up in install/lib/python2.7/dist-packages/python_module. Also python setup.py install is completely broken. You can test that on this revision : https://github.com/asmodehn/flask-ext-catkin.git @ de04b20773460d86f7d19b3970eb5728c09e901b Using the proper distutils parameter for the catkin version also doesn't work. : Using 'python_module_1': 'python-mod-name/python_module_1', makes catkin expect the __init__.py file in python-mod-name/python_module_1/python_module_1 which is incorrect. The reason is that the logic to find the location implemented in catkin is different than the logic implemented in distutils ( which makes things very confusing ) Solution : fixing the logic used to determine locations, and d = generate_distutils_setup( packages=[ 'python_module_1', 'python_module_2', ], package_dir={ 'python_module_1': 'python-mod-name/python_module_1', 'python_module_2': 'python-mod-name-2/python_module_2', }) which matches the distutils logic and solves both problem. catkin_make build and catkin_make install now work properly. also python setup.py install works as expected. Implemented here : https://github.com/asmodehn/flask-ext-catkin.git @ 41fd3f7077ec7d8cb4a54a89ed830a212c387a5d Thank you for the patch! I updated the condition I have commented on in #751. In that branch I have also updated the tests to match the new behavior. Can you please check and confirm that the new PR also works for you? Thanks ! I will check it over the next few days and let you know. On Aug 6, 2015 8:02 AM, "Dirk Thomas"<EMAIL_ADDRESS>wrote: Thank you for the patch! I updated the condition I have commented on in #751 https://github.com/ros/catkin/pull/751. In that branch I have also updated the tests to match the new behavior. Can you please check and confirm that the new PR also works for you? — Reply to this email directly or view it on GitHub https://github.com/ros/catkin/pull/747#issuecomment-128176931. I am closing this in favor of the new PR #751.
2025-04-01T04:35:23.109071
2016-02-12T12:15:29
133226713
{ "authors": [ "dirk-thomas", "nlyubova", "tfoote" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10428", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/10486" }
gharchive/pull-request
romeo_moveit_actions: 0.0.7-2 in 'indigo/distribution.yaml' [bloom] Increasing version of package(s) in repository romeo_moveit_actions to 0.0.7-2: upstream repository: https://github.com/nlyubova/romeo_moveit_actions.git release repository: https://github.com/nlyubova/romeo_moveit_actions-release.git distro file: indigo/distribution.yaml bloom version: 0.5.20 previous version for package: 0.0.7-0 romeo_moveit_actions * Merge pull request #1 <https://github.com/nlyubova/romeo_moveit_actions/issues/1> from IanTheEngineer/remove_shape_tools Convert deprecated shape_tools dependency * Convert deprecated shape_tools dependency shape_tools functionality was merged into geometric_shapes: https://github.com/ros-planning/geometric_shapes/pull/32 and removed from moveit_core https://github.com/ros-planning/moveit_core/pull/242 which caused this issue. This commit updates the pick and place tutorial and adds geometric_shapes to the package.xml and CMakeLists.txt to prevent the ROS buildfarm from failing to build this package. * Contributors: Ian McMahon, Natalia Lyubova Please either change the patch to not change the comment line manually or create a new patch using bloom-release ... -r after updating to the latest release of python-rosdistro. That's bloom-release with --pull-request-only or -p not -r You need python-rosdisto 0.4.4 or higher.
2025-04-01T04:35:23.112324
2022-04-29T02:44:22
1219652609
{ "authors": [ "ihadzic", "methylDragon" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10429", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/32944" }
gharchive/pull-request
Two new packages to add to index: move_base_swp and lsm_localization Please Add This Package to be indexed in the rosdistro. melodic, noetic The source is here: https://github.com/ihadzic/move_base_swp https://github.com/ihadzic/lsm_localization Checks [x] All packages have a declared license in the package.xml [x] This repository has a LICENSE file [x] This package is expected to build on the submitted rosdistro Holding for Noetic sync
2025-04-01T04:35:23.113885
2023-03-31T18:10:56
1649808141
{ "authors": [ "ChrisThrasher", "audrow" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10430", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/36672" }
gharchive/pull-request
Revert "Revert "rsl: 0.2.2-1 in 'humble/distribution.yaml' [bloom]"" Reverts ros/rosdistro#36613 @ChrisThrasher, the Humble sync just went out. We can see if this breaks any downstream packages and try to fix them before the next Humble sync in 2-3 weeks. This release of pick_ik should fix at least one of the build farm errors were previously encountering.
2025-04-01T04:35:23.118498
2024-04-22T06:10:20
2255664809
{ "authors": [ "Timple" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10431", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/40717" }
gharchive/pull-request
azure_iot_sdk_c: 1.13.0-3 in 'rolling/distribution.yaml' [bloom] Increasing version of package(s) in repository azure_iot_sdk_c to 1.13.0-3: upstream repository: https://github.com/Azure/azure-iot-sdk-c.git release repository: https://github.com/ros2-gbp/azure_iot_sdk_c-release.git distro file: rolling/distribution.yaml bloom version: 0.12.0 previous version for package: null Replaces https://github.com/ros/rosdistro/pull/40536 I might need a pointer here... The build fails on: https://build.ros2.org/job/Rdev__azure_iot_sdk_c__ubuntu_noble_amd64/1/console File "/usr/lib/python3/dist-packages/colcon_cmake/task/cmake/build.py", line 173, in _reconfigure raise RuntimeError("Could not find 'cmake' executable") However cmake seems correctly listed as build dependency: https://github.com/ros2-gbp/azure_iot_sdk_c-release/blob/master/patches/package.xml#L11C1-L12C1
2025-04-01T04:35:23.182620
2021-02-23T19:56:24
814784326
{ "authors": [ "clalancette", "stephenstarkie" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10432", "repo": "ros2/ros2_documentation", "url": "https://github.com/ros2/ros2_documentation/issues/1131" }
gharchive/issue
Install ROS2 on macOS fails with pygraphviz I tried to install according to the instructions here: https://index.ros.org/doc/ros2/Installation/Foxy/macOS-Install-Binary/ Installation fails on the line: python3 -m pip install pygraphviz pydot with; Collecting pygraphviz Using cached pygraphviz-1.7.zip (118 kB) Building wheels for collected packages: pygraphviz Building wheel for pygraphviz (setup.py) ... error ERROR: Command errored out with exit status 1: command<EMAIL_ADDRESS>-u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-install-j0l_0s3c/pygraphviz_ac93efbdfc764377ae19b8ff1b8c3d95/setup.py'"'"'; file='"'"'/private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-install-j0l_0s3c/pygraphviz_ac93efbdfc764377ae19b8ff1b8c3d95/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(file);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, file, '"'"'exec'"'"'))' bdist_wheel -d /private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-wheel-gsy173sm cwd: /private/var/folders/7x/9ydhg6pn74s30xzlsltt5h1w0000gp/T/pip-install-j0l_0s3c/pygraphviz_ac93efbdfc764377ae19b8ff1b8c3d95/ Complete output (311 lines): running bdist_wheel running build running build_py creating build creating build/lib.macosx-11-x86_64-3.8 creating build/lib.macosx-11-x86_64-3.8/pygraphviz copying pygraphviz/scraper.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz copying pygraphviz/graphviz.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz copying pygraphviz/init.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz copying pygraphviz/agraph.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz copying pygraphviz/testing.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz creating build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_unicode.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_scraper.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_readwrite.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_string.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/init.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_html.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_node_attributes.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_drawing.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_subgraph.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_close.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_edge_attributes.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_clear.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_layout.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_attribute_defaults.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests copying pygraphviz/tests/test_graph.py -> build/lib.macosx-11-x86_64-3.8/pygraphviz/tests running egg_info writing pygraphviz.egg-info/PKG-INFO writing dependency_links to pygraphviz.egg-info/dependency_links.txt writing top-level names to pygraphviz.egg-info/top_level.txt reading manifest file 'pygraphviz.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '.png' under directory 'doc' warning: no files found matching '.txt' under directory 'doc' warning: no files found matching '.css' under directory 'doc' warning: no previously-included files matching '~' found anywhere in distribution warning: no previously-included files matching '*.pyc' found anywhere in distribution warning: no previously-included files matching '.svn' found anywhere in distribution no previously-included directories found matching 'doc/build' writing manifest file 'pygraphviz.egg-info/SOURCES.txt' copying pygraphviz/graphviz.i -> build/lib.macosx-11-x86_64-3.8/pygraphviz copying pygraphviz/graphviz_wrap.c -> build/lib.macosx-11-x86_64-3.8/pygraphviz running build_ext building 'pygraphviz._graphviz' extension creating build/temp.macosx-11-x86_64-3.8 creating build/temp.macosx-11-x86_64-3.8/pygraphviz clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk -I/usr/local/include<EMAIL_ADDRESS>-I/usr/local/opt/sqlite/include -I/usr/local/opt/tcl-tk/include<EMAIL_ADDRESS>-c pygraphviz/graphviz_wrap.c -o build/temp.macosx-11-x86_64-3.8/pygraphviz/graphviz_wrap.o In file included from pygraphviz/graphviz_wrap.c:154: In file included from /usr/local/Cellar/python@3.8/3.8.8/Frameworks/Python.framework/Versions/3.8/include/python3.8/Python.h:34: In file included from /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/stdlib.h:66: In file included from /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/sys/wait.h:110: In file included from /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/sys/resource.h:72: In file included from /usr/local/include/stdint.h:59: In file included from /usr/local/include/stdint.h:59: ...lots of the same line In file included from /usr/local/include/stdint.h:59: In file included from /usr/local/include/stdint.h:59: /usr/local/include/stdint.h:2:10: error: #include nested too deeply macos 11.2.1 We don't officially support Big Sur with any of our platforms, so it is not entirely surprising. That being said, this looks like a problem in pygraphviz, nothing specific to ROS 2 here. So I'm going to close this out; I'll suggest that you open an issue against https://github.com/pygraphviz/pygraphviz instead.
2025-04-01T04:35:23.220750
2023-12-03T22:46:11
2022749262
{ "authors": [ "WardBrian", "bob-carpenter", "roualdes" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10433", "repo": "roualdes/bridgestan", "url": "https://github.com/roualdes/bridgestan/issues/190" }
gharchive/issue
Python callback into BridgeStan with C types I have an implementation of Stan that iterates and does warmup calculations in Python and leapfrog/trajectories in C++. The gradient calculations implicit in each leapfrog step happen via a Python callback which then calls BridgeStan: Python -> C++ -> Python/BridgeStan -> C/C++/BridgeStan. Since the C++ code is calling a BridgeStan function, and Python/BridgeStan expects numpy arrays, I've got to convert a double* to a numpy array just so that BridgeStan can convert the numpy array back to double*. The conversion to a numpy array is unnecessarily costly, especially since everything is using the same underlying memory. Consider BridgeStan's python/bridgestan/model.py lines 219 to 229. This code chunk defines the Python/BridgeStan interface to the C/BridgeStan function bs_log_density_gradient. Lines 225 and 227 specify numpy arrays as argument types from Python/BridgeStan -> C/C++/BridgeStan. Simply adding the following code to python/bridgestan/model.py::StanModel's __init__() method allows a Python callback from C++ to re-use the original double*: self._ldg = self.stanlib["bs_log_density_gradient"] self._ldg.restype = ctypes.c_int self._ldg.argtypes = [ ctypes.c_void_p, ctypes.c_int, ctypes.c_int, ctypes.POINTER(ctypes.c_double), # changed from double_array ctypes.POINTER(ctypes.c_double), ctypes.POINTER(ctypes.c_double), # changed from double_array star_star_char, ] What do you think of building such functionality into BridgeStan? Just adding the above code won't break our API in anyway, but does offer a whole new world of possibility for BridgeStan. I'm not sold on the name _ldg, I'm just asking about the possibility of including such functionality. If y'all are into this idea, would you brainstorm with me what a better design might look like that allow callbacks, from the various higher level languages, into BridgeStan using plain C types? What do you think of building such functionality into BridgeStan? I couldn't quite follow all the callbacks, so I'm not entirely sure what you are suggesting adding to BridgeStan and what that would enable for clients that they can't achieve now. My understanding of Edward's request is that a user may be working with other code that uses ctypes, and as a result have a double array which is represented as a ctypes.POINTER(ctypes.c_double), rather than a numpy array object. There is nothing at the C level of BridgeStan to differentiate between these, and in principle either is fine, but for convenience BridgeStan only speaks the language of numpy ndarrays, not 'raw' ctypes pointers. You can convert from a ctypes pointer to an ndarray, but since all BridgeStan really does is some bounds checking and then convert it back to a pointer internally, this comes with an extra cost at no extra benefit. The concrete code Ed provided would give you a different function which accepts ctypes objects as opposed to numpy ones, but I also have a minimal patch prepared which would make it so either can be accepted anywhere we currently accept numpy arrays, meaning no extra functions, just slightly broader types being accepted. Got it. Thanks! Right, this doesn't seem to be an issue outside of Python, for the reasons you mention. Good call. Brian proposed some code, in branch python/allow-ctypes-double-pointers, that allows direct access to BridgeStan's log_density_gradient() for both numpy arrays and ctypes. The solution involves an extra if statement, relative to what lives in main branch. To help us determine whether or not we want to include such functionality in BridgeStan, I ran some experiments and report the results below. I ran some numerical simulations to better understand the computational cost trade-offs associated with this extra if statement. From a numpy array user perspective, we expect a slight slow down (one extra if statement). From a ctypes user perspective, we expect a major gain (no unnecessary numpy array creation). The code used for the simulations lives in the github repository bridgestan-speed. The simulations were across three models: gaussian with data (2 parameters), logistic regression (25 parameters), and standard normal with no data (1_000 parameters). Numpy array perspective. 500 runs of 1_000 evaluations of log_density_gradient across two branches. Times reported are means per run plus/minus standard deviations. dimensions / branch main python/allow-ctypes-double-pointers 2 (normal w/ data) 13.9 ms +/- 40.9 μs 14.6 ms +/- 493 μs 25 (logistic regression) 96.2 ms +/- 8.42 ms 82.7 ms +/- 1.13 ms 1_000 (std_normal no data) 27.5 ms +/- 1.48 ms 27.1 ms +/- 721 μs Ctypes perspective. 20 runs of Stan sampling with 1_000 warmup and 1_000 iterations across two branches. Times reported are means per Stan run plus/minus standard deviations. RNG seeds were coordinated to ensure no branch lucked into a worse part of the model parameter space than the other branch. dimensions / branch main python/allow-ctypes-double-pointers 2 (normal w/ data) 251 ms +/- 7.36 ms 89.9 ms +/- 5.3 ms 25 (logistic regression) 2.56 s +/- 132 ms 2 s +/- 74.9 ms 1_000 (std_normal no data) 1.51 s +/- 30.3 ms 908 ms +/- 92 ms What do you all think, should we add this functionality to BridgeStan? Thanks for the report @roualdes. I couldn't quite tell what was being compared. I see two tables of results, but don't know what they mean. What does "across two branches" mean for the "Numpy array perspective"? Is it that the first one is passing in an numpy array and the second passing in some raw memory? What does "across two branches" mean for the "Numpy array perspective"? The first table is meant to show costs associated with the proposed solution to this issue for the most common Python/BridgeStan user, a user who is only dealing with numpy arrays. The table presents mean run times for calling BridgeStan's log_density_gradient with numpy arrays. The first column sets the baseline for run times in the branch main. The second column displays run times for the proposed solution from branch python/allow-ctypes-double-pointers. Is it that the first one is passing in an numpy array and the second passing in some raw memory? Correct. The second table is meant to show benefits associated with the proposed solution when a user has a double* to raw memory. Let me know if I can offer any other words to help clarify what my results are about. Thanks---that's what I suspected and this looks close enough to be worth doing. If the results weren't clear enough here, I'd be tempted to do about 10 times as many evals over several batches because the inconsistent sd in the tests you list (e.g., "13.9 ms +/- 40.9 μs 14.6 ms +/- 493 μs" in the first line of the first table, where the new method has ten times the standard deviation). This kind of micro-benchmarking is notoriously tricky, so probably not worth it. They align with my expectation - for numpy arrays it is a wash and for raw pointers the overhead is much less. If it had ended up being a significant increase for numpy arrays I would be more hesitant toward adding it. However, it's really just adding one more if into a part of the code that already has a bunch of ifs for validation logic, so I'm not shocked the impact is essentially nothing.
2025-04-01T04:35:23.222438
2019-09-22T00:56:03
496718354
{ "authors": [ "roughike", "srihamat" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10434", "repo": "roughike/flutter_facebook_login", "url": "https://github.com/roughike/flutter_facebook_login/pull/182" }
gharchive/pull-request
Update README I spent 2 days looking for solution then eventually found this is the solution, look like someone also face the same issues as well. Update solution here should help save time for more people Thanks! I'll modify this a tiny bit, but LGTM.
2025-04-01T04:35:23.293456
2022-08-03T14:02:06
1327270684
{ "authors": [ "remorses", "rpearce" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10435", "repo": "rpearce/react-medium-image-zoom", "url": "https://github.com/rpearce/react-medium-image-zoom/issues/333" }
gharchive/issue
Document how to change background color In v4 you could use the overlayBgColorStart props to change background color, now you can use the following css (in this case i am also changing background when in dark mode) [data-rmiz-modal-overlay='hidden'] {˙ background-color: rgba(255, 255, 255, 0); } [data-rmiz-modal-overlay='visible'] { background-color: rgba(255, 255, 255, 0.6); backdrop-filter: saturate(180%) blur(60px); -webkit-backdrop-filter: saturate(180%) blur(60px); } .dark [data-rmiz-modal-overlay='hidden'] { background-color: rgba(0, 0, 0, 0); } .dark [data-rmiz-modal-overlay='visible'] { background-color: rgba(0, 0, 0, 0.6); } @remorses Thanks for calling that out. Yes, nearly all style-related things (including transition times) have moved to being CSS controlled now, and you're right — it would be helpful to document that somehow. @remorses Thank you for bringing this up and, in doing so, contributing to this project! I added info to the Styles section in the readme, including a link to an example of how to do custom modal styling. Thank you, amazing project btw!
2025-04-01T04:35:23.298500
2015-02-11T20:08:14
57369930
{ "authors": [ "rpetz" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10436", "repo": "rpetz/SharpNetSH", "url": "https://github.com/rpetz/SharpNetSH/issues/1" }
gharchive/issue
Maintenance nightmare Currently responses to each method call are parsed by hand on a method by method basis. This creates a maintenance nightmare that I do not want to personally support. Considering we are following convention over configuration for the method calls themselves (no concrete classes, all interface proxying) we should do the same thing for the response objects. Rather than using a hand coded response object class for each method call, response objects should be standardized with the following code: public int ExitCode { get; } public Boolean IsNonStandardExitCode { get { return ExitCode != 0; } } public IEnumerable<String> Response { get; } public dynamic ResponseObject { get; } The dynamic response object should follow a set of standardized parsing conventions which should be configured using an attribute on the method (since different methods have different conventions). This has been resolved in the latest code for the master branch
2025-04-01T04:35:23.552588
2022-03-30T06:42:56
1185965249
{ "authors": [ "ened", "rrousselGit" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10438", "repo": "rrousselGit/riverpod", "url": "https://github.com/rrousselGit/riverpod/issues/1326" }
gharchive/issue
examples/stackoverflow: Fast Scroll/Small Page Size will cause App to hang Describe the bug I try to understand the stackoverflow example, specifically how the pagination is done. The forked version of the repo is here: https://github.com/ened/river_pod/tree/examples/stackoverflow/pageSize5 The page size was changed to 5 and I started scrolling quickly. This caused a few errors and eventually, the App is not loading any new data. Expected behavior Open requests to missing pages should load appropriately and backfill what is missing. App should never hang and stop loading This caused a few errors and eventually, the App is not loading any new data. What error? Errors like: Exception has occurred. DioError (DioError [DioErrorType.cancel]: Source stack: #0 DioMixin.fetch (package:dio/src/dio_mixin.dart:473:35) #1 DioMixin.request (package:dio/src/dio_mixin.dart:468:12) #2 DioMixin.requestUri (package:dio/src/dio_mixin.dart:422:12) #3 DioMixin.getUri (package:dio/src/dio_mixin.dart:72:12) #4 paginatedQuestionsProvider.<anonymous closure> (package:stackoverflow/home.dart:79:8) #5 paginatedQuestionsProvider.<anonymous closure> (package:stackoverflow/home.dart:54:37) #6 AutoDisposeFutureProviderFamily.create.<anonymous closure> (package:riverpod/src/future_provider/auto_dispose.dart:132:23) #7 AutoDisposeFutureProvider.create.<anonymous closure> (package:riverpod/src/future_provider/auto_dispose.dart:68:43) #8 _FutureProviderElementMixin._listenFuture (package:riverpod/src/future_provider.dart:89:27) #9 AutoDisposeFutureProvider.create (package:riverpod/src/future_provider/auto_dispose.dart:68:16) #10 ProviderElementBase._buildState (package:riverpod/src/framework/provider_base.dart:476:26) #11 AutoDisposeProviderElementBase._buildState (package:riverpod/src/framework/auto_dispose.dart:152:11) #12 ProviderElementBase.mount (package:riverpod/src/framework/provider_base.dart:348:5) #13 _StateReader._create (package:riverpod/src/framework/container.dart:104:11) #14 _StateReader.getElement (package:riverpod/src/framework/container.dart:92:52) #15 ProviderContainer.readProviderElement.<anonymous closure> (package:riverpod/src/framework/container.dart:490:38) #16 ProviderContainer.readProviderElement (package:riverpod/src/framework/container.dart:523:6) ════════ Exception caught by image resource service ════════════════════════════ The following HandshakeException was thrown resolving an image codec: Connection terminated during handshake When the exception was thrown, this was the stack Image provider: NetworkImage("[https://lh3.googleusercontent.com/a/AATXAJwfy-25HkvYx5KfDm8sxqtm0yMCLIoLgxTofHrJ=k-s256]()", scale: 1.0) Image key: NetworkImage("[https://lh3.googleusercontent.com/a/AATXAJwfy-25HkvYx5KfDm8sxqtm0yMCLIoLgxTofHrJ=k-s256]()", scale: 1.0) ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by image resource service ════════════════════════════ Connection terminated during handshake ════════════════════════════════════════════════════════════════════════════════ ════════ Exception caught by image resource service ════════════════════════════ If waited for long enough (or release build) the errors seem to have no impact yet interrupt the development workflow. If waited for long enough (or release build) the errors seem to have no impact yet interrupt the development workflow. Maybe too many requests were made at the same time due to the small page side. Closing that the Riverpod error shouldn't happen anymore. (well the exception should be thrown, but that's on purpose. On the other hand, it shouldn't be considered "uncaught")
2025-04-01T04:35:23.554780
2023-12-05T01:26:56
2025101624
{ "authors": [ "rruckley" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10439", "repo": "rruckley/tmflib", "url": "https://github.com/rruckley/tmflib/pull/15" }
gharchive/pull-request
Fix quote item Added missing fields from QuoteItem Added missing ProductRefOrValue struct Added missing fields from Quote Changed visibility to public for most fields in QuoteItem Again merging forward.
2025-04-01T04:35:23.561273
2020-06-11T07:10:06
636776248
{ "authors": [ "dadepo", "koivunej" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10440", "repo": "rs-ipfs/rust-ipfs", "url": "https://github.com/rs-ipfs/rust-ipfs/issues/186" }
gharchive/issue
Using rust-ipfs as a library This is not a feature request but a request for clarification on the usage of rust-ipfs as a library Currently I am interacting with IPFS by calling out to an installed binary. So far, here are the commands I need for my interactions: //1 // Ensure ipfs is started with --enable-pubsub-experiment and --enable-namesys-pubsub //2 ipfs add -r /path_to_dir //3 ipfs name publish --key=hash-public-key cid_to_publish //4 ipfs get "ipns/path"--output=destination_path My questions now are: I know rust-ipfs is pre-alpha, but does it currently support --enable-pubsub-experiment and --enable-namesys-pubsub? Does it also currently support adding a directory, publishing to ipns and requesting from ipns? Would there be a theoretical/real performance benefit of using rust-ipfs as a library compared with interacting with an installed binary? Thanks! Thanks for reaching out! I know rust-ipfs is pre-alpha, but does it currently support --enable-pubsub-experiment and --enable-namesys-pubsub? The current build has --enable-pubsub-experiment always on, and in fact, it cannot be turned off via configuration. The inability to configure it off with the #132 makes it so that you cannot connect to go-ipfs 0.5 which is running with --enable-pubsub-experiment. We currently do not support IPNS at all, so no support for IPNS over pubsub either. Does it also currently support adding a directory, publishing to ipns and requesting from ipns? We do not currently have an unixfs importer or the add but we have planned implementing it rather soon. Currently I am working over at the unixfs reading support #172 and /cat is looking ok with #184 which will leave the main library or crate missing out the IpfsPath resolving, which has had some iterations over it. The IpfsPath enhancements should be following rather soon. I don't think we have any concrete plans for IPNS. On my mind the path has been to first fix the dns story #83 to enable dnslink, then implement keystore management followed by initial work on IPNS over DHT. At that point I expect the libp2p-gossipsub to be compatible with go-ipfs and go-ipfs defaulting to it as well, so the IPNS over pubsub would be next logical step. We are currently gathering Phase 3 ideas over at #181 but since this issue was already created: Could you open a bit your use case up? Any details would be much appreciated. Would there be a theoretical/real performance benefit of using rust-ipfs as a library compared with interacting with an installed binary? At least the theoretical benefit is there but at least for now, it might be non-existent or only caused by the fact that we implement so few features compared to {go,js}-ipfs. In the long run we aim to provide a predictable and trustworthy experience as crate and of course rust and LLVMs optimization capabilities will shine compared to communicating over HTTP when we have the features working. Thanks for the response. I have enough information to make my decision regarding usage of the current version rust-ipfs. Tnanks!
2025-04-01T04:35:23.574035
2018-07-17T08:47:26
341822948
{ "authors": [ "bpoplauschi", "codecov-io", "dreampiggy" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10441", "repo": "rs/SDWebImage", "url": "https://github.com/rs/SDWebImage/pull/2388" }
gharchive/pull-request
Update the migration guide with the latest version. Update one changelog New Pull Request Checklist [x] I have read and understood the CONTRIBUTING guide [x] I have read the Documentation [x] I have searched for a similar pull request in the project and found none [x] I have updated this branch with the latest master to avoid conflicts (via merge from master or rebase) [x] I have added the required tests to prove the fix/feature I am adding [x] I have updated the documentation (if necessary) [x] I have run the tests and they pass [x] I have run the lint and it passes (pod lib lint) This merge request fixes / reffers to the following issues: ... Pull Request Description This PR contains any update for the wiki/documentation/changelog. You can just put anything to change for these into single PR. Everyone who need to change documentations can directlly add commit here. So we don't need to merge again and again. Until final update is done, we can merge and release the 5.0 version (Or another beta version if we find something important to change and test) Good idea @dreampiggy Codecov Report Merging #2388 into 5.x will decrease coverage by 0.06%. The diff coverage is n/a. @@ Coverage Diff @@ ## 5.x #2388 +/- ## ========================================== - Coverage 68.77% 68.71% -0.07% ========================================== Files 47 47 Lines 6607 6607 ========================================== - Hits 4544 4540 -4 - Misses 2063 2067 +4 Flag Coverage Δ #iOS 69.72% <ø> (-0.02%) :arrow_down: #macOS 68.11% <ø> (-0.02%) :arrow_down: Impacted Files Coverage Δ SDWebImage/SDWebImageManager.m 81.3% <0%> (-0.82%) :arrow_down: SDWebImage/SDWebImageDownloaderOperation.m 87.62% <0%> (-0.69%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 957de6d...ffad516. Read the comment docs. @bpoplauschi This one now been updated to the latest 5.x. Maybe it's time to merge.
2025-04-01T04:35:23.598133
2021-11-05T12:47:11
1045813029
{ "authors": [ "TheTams", "rsh249" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10442", "repo": "rsh249/cRacle", "url": "https://github.com/rsh249/cRacle/issues/4" }
gharchive/issue
Define the 7 outputs from get_optim() in documentation Currently the titles of the lists in the output from get_optim() do not make it clear to the user, which output corresponds which method as discussed in Harbert and Baryiames 2019. I have been unable to find any documentation that defines them elsewhere. The categories are: conintkde conintgauss origk origg dircontint means sds I will work on this documentation as well. These names are a holdover from how I initially set up the models and output collection, but we worked on the terms in the last publication a bit more to better align with the modeling that was actually taking place. For a quick reference: The "*kde" and "*k" methods are non-parametric models described in the Harbert & Baryiames, 2019 paper. The "*gauss" and "*g" methods are the parametric models. "conint" returns 95% confidence intervals of both methods "orig" returns an estimate of the maximum likelihood value (the exact values aren't calculated but a narrow interval around the maximum is easy to get. dirconint is functionally equivalent to conintgauss but uses the weighted means method instead of the ML method. means is equivalent to origg but uses the weighted means method instead of the ML method. sds is a calculation to go with means of the weighted standard deviation (something you don't get with the ML method but I haven't looked into much). I hope this helps for now. Thank you for that, I thought I had figured some of it out... So is the weighted means the equivalent of the MCRM from Thompson et al. (2012)? And none of these are equivalent to the Coexistence Approach? The weighted means is equivalent to the parametric CRACLE. There is a separate function for MCR a la Thompson et al. 2012 that does the unweighted MCR which (I think) is the same as the coexistence approach. Sorry, my bad, I see the MCR one there. Great!
2025-04-01T04:35:23.603608
2020-03-03T21:40:21
574965784
{ "authors": [ "StriveForBest", "WoLpH", "bharling", "pablolmedorado" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10443", "repo": "rsinger86/drf-flex-fields", "url": "https://github.com/rsinger86/drf-flex-fields/issues/46" }
gharchive/issue
Adding FlexFieldsFilterBackend immediately breaks the app I am utilizing drf-flex-fields heavily and everything works fine until the moment i add: REST_FRAMEWORK = { ... 'DEFAULT_FILTER_BACKENDS': ( 'rest_flex_fields.filter_backends.FlexFieldsFilterBackend', ), ... } It crashes the app with the following traceback: INFO 2020-03-03 21:38:35,864 autoreload 90410<PHONE_NUMBER> Watching for file changes with StatReloader Performing system checks... Exception in thread django-main-thread: Traceback (most recent call last): File<EMAIL_ADDRESS>line 932, in _bootstrap_inner self.run() File<EMAIL_ADDRESS>line 870, in run self._target(*self._args, **self._kwargs) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/autoreload.py", line 53, in wrapper fn(*args, **kwargs) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/management/commands/runserver.py", line 117, in inner_run self.check(display_num_errors=True) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/management/base.py", line 392, in check all_issues = self._run_checks( File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/management/base.py", line 382, in _run_checks return checks.run_checks(**kwargs) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/checks/registry.py", line 72, in run_checks new_errors = check(app_configs=app_configs) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/checks/urls.py", line 13, in check_url_config return check_resolver(resolver) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/core/checks/urls.py", line 23, in check_resolver return check_method() File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/resolvers.py", line 407, in check for pattern in self.url_patterns: File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/functional.py", line 48, in __get__ res = instance.__dict__[self.name] = self.func(instance) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/resolvers.py", line 588, in url_patterns patterns = getattr(self.urlconf_module, "urlpatterns", self.urlconf_module) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/functional.py", line 48, in __get__ res = instance.__dict__[self.name] = self.func(instance) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/resolvers.py", line 581, in urlconf_module return import_module(self.urlconf_name) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1014, in _gcd_import File "<frozen importlib._bootstrap>", line 991, in _find_and_load File "<frozen importlib._bootstrap>", line 975, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 671, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 783, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/urls.py", line 6, in <module> path('api/v1/', include('gagosian.api.v1.urls', namespace='api')) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/urls/conf.py", line 34, in include urlconf_module = import_module(urlconf_module) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1014, in _gcd_import File "<frozen importlib._bootstrap>", line 991, in _find_and_load File "<frozen importlib._bootstrap>", line 975, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 671, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 783, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/api/v1/urls.py", line 4, in <module> from .contacts import urls as contacts_urls File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/api/v1/contacts/urls.py", line 3, in <module> from .viewsets import ( File "/Users/alex.zagoro/projects/gagosian/noya/gagosian/api/v1/contacts/viewsets.py", line 1, in <module> from rest_framework.viewsets import ModelViewSet File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/viewsets.py", line 27, in <module> from rest_framework import generics, mixins, views File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/generics.py", line 24, in <module> class GenericAPIView(views.APIView): File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/generics.py", line 43, in GenericAPIView filter_backends = api_settings.DEFAULT_FILTER_BACKENDS File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 220, in __getattr__ val = perform_import(val, attr) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 168, in perform_import return [import_from_string(item, setting_name) for item in val] File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 168, in <listcomp> return [import_from_string(item, setting_name) for item in val] File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_framework/settings.py", line 177, in import_from_string return import_string(val) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/django/utils/module_loading.py", line 17, in import_string module = import_module(module_path) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_flex_fields/__init__.py", line 3, in <module> from .views import FlexFieldsModelViewSet File "/Users/alex.zagoro/venv/noya-TS4FQGPC/lib/python3.8/site-packages/rest_flex_fields/views.py", line 21, in <module> class FlexFieldsModelViewSet(FlexFieldsMixin, viewsets.ModelViewSet): AttributeError: partially initialized module 'rest_framework.viewsets' has no attribute 'ModelViewSet' (most likely due to a circular import) Adding the backend directly to the view works fine. Yeah seeing this too, but I get: ImportError: Could not import 'rest_flex_fields.filter_backends.FlexFieldsFilterBackend' for API setting 'DEFAULT_FILTER_BACKENDS'. ImportError: cannot import name 'GenericViewSet'. looks like a circular import also As a workaround (I also posted this at #59 which appears to be a duplicate) you can set the filter backend at the view: import rest_flex_fields.filter_backends as flex_filters from rest_framework import viewsets class ModelViewSet(viewsets.ModelViewSet): filter_backends = viewsets.ModelViewSet.filter_backends + [ flex_filters.FlexFieldsFilterBackend, ] Starting from @WoLpH idea, I have created a mixin to replace the "FlexFieldsMixin" that includes the FilterBackend and also respects the user settings. I'll leave the code here just in case there was anyone interested. from rest_flex_fields.filter_backends import FlexFieldsFilterBackend from rest_flex_fields.views import FlexFieldsMixin class FlexFieldsOptimizedMixin(FlexFieldsMixin): filter_backends = [FlexFieldsFilterBackend] + api_settings.DEFAULT_FILTER_BACKENDS
2025-04-01T04:35:23.607048
2024-09-25T01:00:14
2546665427
{ "authors": [ "jeremy-then" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10444", "repo": "rsksmart/rootstock-integration-tests", "url": "https://github.com/rsksmart/rootstock-integration-tests/pull/98" }
gharchive/pull-request
Adds 'should reject a basic pegin v1 with value exactly below minimum… Adds 'should reject a basic pegin v1 with value exactly below minimum' test. This test sends a pegin v1 with a value exactly below minimum (the minimum pegin value minus 1 satoshis), to assert that it should be rejected with rejection reason 5 (invalid amount). Asserts that the sender funds are lost and that the sender's derived rsk address nor the pegin v1 rsk recipient address get any funds, but the federation balance is increased by the pegin amount while the Bridge doesn't mark that pegin tx as processed. https://github.com/rsksmart/rootstock-integration-tests/pull/108
2025-04-01T04:35:23.610263
2023-10-03T12:56:54
1924095814
{ "authors": [ "marcgarba", "otterdahl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10445", "repo": "rsmp-nordic/rsmp_core", "url": "https://github.com/rsmp-nordic/rsmp_core/issues/129" }
gharchive/issue
Clarify 4.5.3. Unimplemented statuses or commands This sentence can be written more clearly: If a status (sCI) or command (cCI) is unimplemented, the site answers with CommandResponse/Response where q or age is set according to the table below. v should be set to null. To understand better the subtlety compared with the other case of "unknown" status/command (SXL mismatch), we could have this written instead : If a status (sCI) or command (cCI) is recognized in relation to its SXL but not implemented the site answers with StatusResponseOrUpdate/CommandResponse where q or age is set according to the table below. v should be set to null. For the 4.5.1 (unknown component object), we could have the same list "StatusResponseOrUpdate/CommandResponse" to be consistent. Opened a PR here: https://github.com/rsmp-nordic/rsmp_core/pull/136 The PR has been merged. Closing this issue
2025-04-01T04:35:23.792024
2017-02-15T08:50:55
207741311
{ "authors": [ "FinScience", "javierluraschi" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10446", "repo": "rstudio/sparklyr", "url": "https://github.com/rstudio/sparklyr/pull/492" }
gharchive/pull-request
Normalize jar paths for windows under spark 1.6 Normalize jar paths for windows under spark 1.6 to fix #491 #490 #414 #410 #398 #305 Planning to backport this one to CRAN, since this is causing a bunch of issues. The problem is scoped to only Windows and Spark 1.6.x; however, a lot of users use this configuration since our documentation defaults to 1.6.2. The fix is straightforward, we need to normalize the paths to jars we specify for the embedded CSV parser. Unfortunately, this affects almost any operation since other non-csv operations also cause Spark to enumerate jar dependencies which trigger an java.io.IOException: No FileSystem for scheme exception. @jjallaire @kevinushey I'm planning to push this as a patch to CRAN and release as sparkly 0.5.2, see branch: https://github.com/rstudio/sparklyr/tree/bugfix/hotfix-0.5.2 for complete changes. This problem makes 1.6.x unusable in Windows, a workaround in the config is possible but very inconvenient since all our docs are based on 1.6. I've validated that normalizePath has no effect in OS X and that this correctly resolves the windows issue. I tested the fix in Windows with 2.0, 1.6 and OS X with 2.0 and 1.6 as well. I'm also backporting a Livy fix which makes spark_connect fail intermittently when using Livy, also really safe fix. @kevinushey if there is something very safe worth backporting, let me know. @javierluraschi . With respect to issue #491 , the errors still persist. I used devtools::install_github("rstudio/sparklyr") to get the latest version. Please let me know in case anything else needs to be done. @FinScience since this PR was not merged at the time, you could have used: devtools::install_github("rstudio/sparklyr", ref = "bugfix/copy-to-windows-spark-1-6") However, now that is merged you can use: devtools::install_github("rstudio/sparklyr") @javierluraschi I tried it now. For some reason, I am still getting the same errors. sc <- spark_connect(master = "local") Warning messages: 1: In value[3L] : java.lang.RuntimeException: java.lang.NullPointerException at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522) at org.apache.spark.sql.hive.client.ClientWrapper.(ClientWrapper.scala:204) at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:238) at org.apache.spark.sql.hive.HiveContext.executionHive$lzycompute(HiveContext.scala:218) at org.apache.spark.sql.hive.HiveContext.executionHive(HiveContext.scala:208) at org.apache.spark.sql.hive.HiveContext.setConf(HiveContext.scala:440) at org.apache.spark.sql.SQLContext$$anonfun$4.apply(SQLContext.scala:272) at org.apache.spark.sql.SQLContext$$anonfun$4.apply(SQLContext.scala:271) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) at scala.collection.AbstractIterable.foreach(Iterable.scala:54) at [... truncated] 2: In create_hive_context_v1(sc) : Failed to create Hive context, falling back to SQL. Some operations, like window-functions, will not work @FinScience I have a couple suggestions for you, please see: https://github.com/rstudio/sparklyr/issues/491 and reply on that thread to continue this investigation. Thank you!
2025-04-01T04:35:23.835826
2018-03-28T10:30:31
309309361
{ "authors": [ "GeekRishabh", "morenoh149" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10447", "repo": "rt2zz/react-native-contacts", "url": "https://github.com/rt2zz/react-native-contacts/issues/278" }
gharchive/issue
Docs: Using it with Pod How to use this package when we are using Cocoapods. I am also using react-native-firebase . Please provide docs how to use it. Sharing My Podfile platform :ios, '8.0' target 'test' do # Uncomment the next line if you're using Swift or would like to use dynamic frameworks # use_frameworks! # Required by RNFirebase pod 'Firebase/Core' pod 'Firebase/Auth' pod 'Firebase/Firestore' pod 'Firebase/Database' pod 'Firebase/Messaging' pod 'RNVectorIcons', :path => '../node_modules/react-native-vector-icons' pod 'react-native-contacts', :path => '../node_modules/react-native-contacts' end Have upgrade to platform :ios, '9.0' but this breaks the firebase linking modules I don't use podfiles, perhaps borrow inspiration from https://github.com/msand/SVGPodTest/blob/master/ios/Podfile#L12
2025-04-01T04:35:23.845654
2021-01-07T05:20:40
781037038
{ "authors": [ "klorel", "teamxSxoptimiz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10448", "repo": "rte-france/challenge-roadef-2020", "url": "https://github.com/rte-france/challenge-roadef-2020/issues/25" }
gharchive/issue
Semi-final phase The semi final phase will end on March 8 2021 and we will be competing on set B of instances. Are we allowed/requested to send a new version of the program before that deadline? Thanks Dear competitors, Please refer to the dedicated web page : https://www.roadef.org/challenge/2020/en/calendrier.php New contribution are to be sent before March 16th 2021. Great, thanks. What is the protocol for the submission? The same protocol used for the qualification phase but changing the subject to "ROADEF/EURO challenge 2020 semi final material from XX"?
2025-04-01T04:35:23.854807
2018-12-11T02:28:40
389570077
{ "authors": [ "Jacob-Stevens-Haas", "SilleBille", "humitos" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10449", "repo": "rtfd/readthedocs.org", "url": "https://github.com/rtfd/readthedocs.org/issues/4987" }
gharchive/issue
Pandoc support Details Read the Docs project URL: Build URL (if applicable): Read the Docs username (if applicable): Expected Result This request may be turned down. But, giving my shot I'm pretty new to readthedocs and I was impressed by the features. Thanks for this awesome project! While researching, I came across pandoc which isn't supported by readthedocs (or at least, I don't find it in the tutorial). I am trying to achieve the following: A set of markdown files that: a. is generated in the form of linux man pages b. is generated into HTML With the help of pandoc, I'm able to achieve the above 2. However, I am not able to host it on readthedocs. Is there a possiblity to add support for pandoc? Actual Result No way to build with pandoc and deploy to readthedocs Hi @SilleBille! Read the Docs currently achieve 2 -- generate HTML from Markdown. Also, Sphinx -- the builder that RTD uses -- has the ability to generate man pages but we are not doing that currently. Although, there is an issue already open for that #4458 Would it be enough for you to enable man page generation for Sphinx builder? @humitos thanks for the reply! To summarize my needs with a simplified diagram: javadoc is generated using the official javadoc package man pages are currently processed using pandoc Documentation can also be converted to HTML by pandoc. Using sphinx/mkdocs for just this workflow makes the pipeline complicated. I want to keep the pipeline as simplified as possible. Though sphinx supports all the above, as per your previous comment, it seems rtfd doesn't provide support for man pages? Thanks for the explanation. I think this is a very fair feature request considering that we support sphinx and mkdocs. Why not pandoc also? I don't have too much experience with pandoc (I've only used it to convert files between formats and it's great) and I don't know how much complicated it would be, but I'm sure that it's not an easy feature to add to our platform. On the other hand, considering our current roadmap, I don't think that this will happen soon, unfortunately. I marked this issue as Needed: design decision so other core developers can write their thoughts here. To answer your question, though sphinx supports all the above, as per your previous comment, it seems rtfd doesn't provide support for man pages? Yes, I can switch over to sphinx if man pages can be generated in both roff and HTML formats Currently, we are not generating man pages. So, I'd say that it's not supported. The discussion for this will happen in #4458 @humitos Thanks for an optimistic reply. The main concern for me is that I don't want my team to follow different formats to write down the documentation. I want them to follow consistent formatting across all the docs they write. I'm going to close this issue since it's not going to happen soon, unfortunately. We can revisit later if we consider valuable to add another builder. Thanks! Adding a link here because I found this issue before the relevant, older one, but it looks like readthedocs has pandoc support already: https://github.com/readthedocs/readthedocs.org/issues/579#issuecomment-31547081
2025-04-01T04:35:23.857670
2018-09-05T22:55:31
357433159
{ "authors": [ "agjohnson", "ericholscher", "humitos" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10450", "repo": "rtfd/readthedocs.org", "url": "https://github.com/rtfd/readthedocs.org/pull/4608" }
gharchive/pull-request
Add a contrib Dockerfile for local build image on Linux This is necessary as permissions are all incorrect on the paths that are shared between the host system and the Docker build container. Closes #2692 Anything we can do to move this along? Seems like a useful improvement. I rebased this branch and push a new commit to always use latest to build the image and also guide the user to re-tag the Docker image after building it so Read the Docs uses it without changing any setting that could break in other places. I've updated this pr to use a configurable label in the docker script. The last piece is how to handle DOCKER_IMAGE_SETTINGS, and I think the answer is maybe alter this setting conditionally to point to the dev images. I added a quick hack to settings.dev to allow for manual triggering of the docker image name replacement. One last review and I think this is ready. Let's see what new and interesting problems this will likely create for us eventually! :crossed_fingers:
2025-04-01T04:35:23.860919
2015-12-05T01:32:35
120519150
{ "authors": [ "agjohnson", "danroth27" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10451", "repo": "rtfd/sphinx-autoapi", "url": "https://github.com/rtfd/sphinx-autoapi/issues/45" }
gharchive/issue
Default output folder for docfx changed to _site Looks like the .NET mapper assumes that the default output folder will be _api_, but it recently go changed to _site. The .NET mapper should probably explicitly specify the output folder when running docfx to make sure it can find the output metadata. This can be moved to a mapper-specific configuration option, though we should also specify the path with the docfx option -o or --output. It appears the -o option only prepends to the output path, it never overrides. Forcing -o _api creates an output path -o _api/_api. If there is a dest setting of 'docs/apiin thedocfx.jsonfile, forcing-o _apicreates an output path of_api/docs/api`. So, seems we can't use this to set a path quite yet. Not sure if this is a docfx-dev vs docfx mainline issue, but I get the output path of _api. Addressing this in a PR, if the path does change to _site we can update again.
2025-04-01T04:35:23.871774
2021-08-25T01:39:47
978619249
{ "authors": [ "korken89", "pdgilbert" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10452", "repo": "rtic-rs/cortex-m-rtic", "url": "https://github.com/rtic-rs/cortex-m-rtic/issues/517" }
gharchive/issue
trait implementation fails on task shared object (Using 0.6-alpha from current github #22ec841.) I am trying to implement a trait but the implementation is not recognized for the shared object. The following code compiles with the cx.shared.led.on() line commented out Click to expand #![deny(unsafe_code)] #![no_std] #![no_main] #[cfg(debug_assertions)] use panic_semihosting as _; #[cfg(not(debug_assertions))] use panic_halt as _; use rtic::app; #[cfg_attr(feature = "stm32l4xx", app(device = stm32l4xx_hal::pac, peripherals = true, dispatchers = [USART1, USART2]))] mod app { use dwt_systick_monotonic::DwtSystick; use rtic::time::duration::{Seconds, Milliseconds}; const PULSE: u32 = 1; // 1 second const PERIOD: u32 = 10 * PULSE; // 10 seconds use stm32l4xx_hal::{ gpio::{gpioc::PC13, Output, PushPull}, pac::Peripherals, prelude::*, }; use embedded_hal::digital::v2::OutputPin; type LedType = PC13<Output<PushPull>>; fn setup(dp: Peripherals) -> LedType { let mut rcc = dp.RCC.constrain(); let mut gpioc = dp.GPIOC.split(&mut rcc.ahb2); let led = gpioc.pc13.into_push_pull_output(&mut gpioc.moder, &mut gpioc.otyper); impl LED for PC13<Output<PushPull>> { fn on(&mut self) -> () { self.set_low().unwrap() } fn off(&mut self) -> () { self.set_high().unwrap()} } led } pub trait LED { fn on(&mut self) -> (); fn off(&mut self) -> (); } #[monotonic(binds = SysTick, default = true)] type DwtMono = DwtSystick<80_000_000>; #[local] struct Local {} #[init] fn init(cx: init::Context) -> (Shared, Local, init::Monotonics) { let mut dcb = cx.core.DCB; let dwt = cx.core.DWT; let systick = cx.core.SYST; let mono = DwtSystick::new(&mut dcb, dwt, systick, 8_000_000); let device = cx.device; let mut led = setup(device); led.off(); //WORKS HERE led.on(); led.off(); crate::app::one::spawn_after(Milliseconds(PULSE * 1000 - 700)).ok(); (Shared {led}, Local {}, init::Monotonics(mono)) } #[shared] struct Shared { led: LedType, } #[task(shared = [led])] fn one(cx: one::Context) { //cx.shared.led.on(); //FAILS HERE compiles with this line commented out crate::app::one::spawn_after(Seconds(PULSE)).ok(); } } but with the line uncommented I get 3 | #[cfg_attr(feature = "stm32l4xx", app(device = stm32l4xx_hal::pac, peripherals = true, dispatchers = [USART1, USART2]))] | -------------------------------------------------------------------------------------- method `on` not found for this ... 85 | cx.shared.led.on(); //FAILS HERE compiles with this line commented out | ^^ method not found in `led<'_>` | = help: items from traits can only be used if the trait is implemented and in scope note: `LED` defines an item `on`, perhaps you need to implement it --> examples/zzled.rs:46:5 | 46 | pub trait LED { | ^^^^^^^^^^^^^ The implementation works in the #[init] section, it is only in the #[task(shared = [led])] section that I get the error. I've been testing mostly with stm32l4xx but it also occurs with stm32f1xx. Hi, You need to lock the led fits. A more complete working and run tested example, from which the above was extracted, is now at #490.
2025-04-01T04:35:23.892863
2016-12-07T21:50:09
194180657
{ "authors": [ "IljaDaderko", "Stringsaeed", "ads1018", "digitalmaster", "faddee", "jachiike-madubuko", "louisgv", "rtsao", "tajo" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10453", "repo": "rtsao/styletron", "url": "https://github.com/rtsao/styletron/issues/26" }
gharchive/issue
React Native Support? The state of managing css in React-Native is still kinda clunky. This would be a huge improvement.. Do you know if it's supported? As someone who hasn't actually had a chance to use React Native myself, I'm interested to learn more about which parts of Styletron could help! I don't know much about it, to be honest! In my mind, the coolest thing about Styletron is the under-the-hood mapping of declarations to atomic classes, with the result of producing tiny stylesheets for server-rendering for browsers -- I wouldn't be able to say how that transfers into the context of React Native. React Native support would be awesome! I know styled-components support its -> https://github.com/styled-components/styled-components#react-native Haven't got a chance to work with React Native so I'm not that familiar of what's happening under the hood but shouldn't this work? function styled(Base, style) { return (props) => ( <Base {...props} style={{ ...(typeof style === 'function' ? style(props) : style), ...props.style }} /> ); } With of course some minor tweaks so you're able to compose styled components. Any updates on react-native support / implementation examples? With the new v4 architecture, all that needs to be done here is a rendering engine for react native (which uses RN-shaped style objects). There's no assumptions about atomic rendering or even the shape of style objects. Is this issue resolved? If not I can take a whack at it this week Is this issue resolved? It's not. We don't use RN. If not I can take a whack at it this week That would be great! Very excited to see this extension of styletron. Is there any news for 2020 on progress. Thanks for being badass! @jachiike-madubuko We still don't have any plans for RN engine since we don't use RN. anyone work on this or should work on it? @Stringsaeed No. The status is unchanged.
2025-04-01T04:35:23.906979
2024-04-03T18:15:04
2223603297
{ "authors": [ "Blad3forc3", "ruanformigoni" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10454", "repo": "ruanformigoni/gameimage", "url": "https://github.com/ruanformigoni/gameimage/issues/52" }
gharchive/issue
Wineprefix not created error When I get to the configure wine options during setup (install dxvk etc) i get this error: Writing image to "/media/ade/Temp/Temp/build/Age-Of-Empires/icon/icon.png" I :: Writing grayscale image to "/media/ade/Temp/Temp/build/Age-Of-Empires/icon/icon.grayscale.png" I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/Age-Of-Empires/gameimage.json"' as UPDATE I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: Open file '"/media/ade/Temp/Temp/build/gameimage.json"' as READ I :: application: Age-Of-Empires I :: image: "/media/ade/Temp/Temp/build/wine.flatimage" I :: prefix: "/media/ade/Temp/Temp/build/Age-Of-Empires/wine" I :: [e] :: bwrap: Can't create file at /tmp/fim/instance/502568f_20240217183447/mount/KIdoON.mount.dwarfs.usr//lib/libnvidia-gtk2.so.510.47.03: Function not implemented I :: [e] :: Killed I :: Finished Command: '/media/ade/Temp/Temp/build/wine.flatimage fim-exec wine.sh regedit' Wine prefix does not exist Thanks for the report @Blad3forc3 ! I found the issue and fixed it, it was on the wine runner when used with nvidia, erase the build folder and re-download the wine packages. That should do it :smile: Fixed in latest wine runner, thanks! The issue is that the HOME was set I :: [o] :: [wine.sh] HOME : /home/ade, and it was building from an external drive. Just resolved other issues with winetricks, please wait for this pipeline to finish. I'll assume this is fixed, feel free to re-open if it is not. :smile: This was fixed yes :)
2025-04-01T04:35:23.914896
2020-02-07T11:09:21
561572254
{ "authors": [ "michae1" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10455", "repo": "rubicon-project/prebid-server-java", "url": "https://github.com/rubicon-project/prebid-server-java/pull/612" }
gharchive/pull-request
Add influxDb tags Sometimes influxDb tags can come in handy to split data by region, hostname etc. I find there is no tagging in current prebid-server-java. As dropwizard-metrics-influxdb already has tags support, I suggest this simple implementation. Fixed empty tags exception
2025-04-01T04:35:23.920061
2018-07-04T03:45:37
338120737
{ "authors": [ "composerinteralia", "koic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10456", "repo": "rubocop-hq/rubocop-performance", "url": "https://github.com/rubocop-hq/rubocop-performance/issues/11" }
gharchive/issue
Add RuboCop rake task And run RuboCop in CI (see #10) I'm preparing for minimal rake tasks now. I will prepare #27 and some rake tasks. Basic rake tasks have been prepared with #27, #28, and #30. I will close the issue.
2025-04-01T04:35:23.922547
2018-07-02T18:39:28
337622162
{ "authors": [ "egze", "jfelchner" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10457", "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/issues/6069" }
gharchive/issue
ForceEqualSignAlignment should only align block content The current alignment makes it not obvious what variables are set inside and outside the block. I think rubocop should align only what is inside the block. Here is an example: my_secret_values = [1,2,3,4].map do |i| x = i * 2 secret = x*x end Expected behavior my_secret_values = [1, 2, 3, 4].map do |i| x = i * 2 secret = x * x end Actual behavior my_secret_values = [1, 2, 3, 4].map do |i| x = i * 2 secret = x * x end Steps to reproduce the problem Run rubocop with autocorrect. RuboCop version Include the output of rubocop -V. Here's an example: $ rubocop -V 0.54.0 (using Parser <IP_ADDRESS>, running on ruby 2.4.2 x86_64-darwin17) This should be fixed in #7211
2025-04-01T04:35:23.924434
2019-10-03T13:16:33
502060316
{ "authors": [ "bbatsov", "mvz", "scottmatthewman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10458", "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/pull/7394" }
gharchive/pull-request
[Fix #6759] Reference configuration in Basic Usage Within the Basic Usage page, make a brief mention of the use of .rubocop.yml to change RuboCop's behaviour. The additional content links to the Configuration page to go into more detail. Fixes #6759. Looks good. Thanks! Thanks, @scottmatthewman!
2025-04-01T04:35:23.925681
2020-08-10T18:39:16
676336307
{ "authors": [ "fatkodima", "marcandre" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10459", "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/pull/8511" }
gharchive/pull-request
Fix a false positive for Style/CaseLikeIf when conditional contains comparison with a class Closes #8508 Marked it as unsafe. I'll merge after the merge conflict is resolved... 😢 Thanks!
2025-04-01T04:35:23.932296
2024-02-22T05:53:21
2148264711
{ "authors": [ "TJNII", "jonas054" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10460", "repo": "rubocop/rubocop", "url": "https://github.com/rubocop/rubocop/issues/12706" }
gharchive/issue
Add ability to globally disable caching Is your feature request related to a problem? Please describe. While backing up my system today I noticed Rubocop has begun writing into ~/.cache. I find this behavior undesirable as I try to keep my code and it's dependencies tightly contained in working directories, it's never pleasant to find when a tool escapes where I put it and begins writing to other paths on my filesystem. I found the documentation at https://docs.rubocop.org/rubocop/usage/caching.html, but I was disappointed to see that this option, which writes outside the per-project tree, is currently only settable in per-project config. I'm not looking forward to having to remember to turn this off on every project I work on. Describe the solution you'd like As Rubocop has begun writing outside the bounds of the project tree, please add a config file that reads from outside the project tree (Say, ~/.config/rubocop.yml) where environmental options like this can be globally specified. For this case I'm not seeking any deep configs so a simple read and merge should be sufficient. Describe alternatives you've considered Per https://docs.rubocop.org/rubocop/usage/caching.html there aren't any. chmod 000 ~/.cache/rubocop_cache I guess? Per https://docs.rubocop.org/rubocop/configuration.html reading from ~/.rubocop.yml is supported, but if $(pwd)/.rubocop.yml exists it will not be loaded. The behavior added in https://github.com/rubocop/rubocop/pull/8314 appears to be at odds with with the behavior I want. I maintain my assertion that if Rubocop is going to write outside the project tree, then it must accept config from outside the project tree as well. Perhaps an ENV var? Per https://github.com/rubocop/rubocop/blob/master/lib/rubocop/cache_config.rb#L8 setting RUBOCOP_CACHE_ROOT to an empty string will not disable it. What about the --cache false command line option? Have you considered that? I have, but the spirit of the request is for when I forget about this behavior and don't remember I need to add a flag or tune the project config file. I'm looking for something I can set in my home directory or bashrc to configure Rubocop to stay within the project tree. The problem with a command line argument is I often run Rubocop from Make or bundler, and bundler (apparently) doesn't use bash aliases so I can't alias rubocop="rubocop --cache-false" in my .bashrc. I actually just realized that there is a simple, supported solution to this: My grouse isn't with the caching itself, it's with the default cache directory being outside the project tree. RUBOCOP_CACHE_ROOT supports relative paths. RUBOCOP_CACHE_ROOT='./tmp/rubocop/cache' works. As this caching feature has been around for a while and nobody else is asking for a global disable option I think we can close this. I found a solution to my concern. And for the record chmod 000 ~/.cache/rubocop_cache does work, but that's only good for making sure ~/.cache/rubocop_cache is never written and reminding that the project config needs updated via the error. It's heavy handed, but workable.
2025-04-01T04:35:23.937781
2023-01-03T05:03:30
1516933340
{ "authors": [ "koic", "sambostock" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10461", "repo": "rubocop/rubocop", "url": "https://github.com/rubocop/rubocop/pull/11378" }
gharchive/pull-request
Fix OperatorMethodCall when forwarding arguments ... cannot be used as an argument to an operator, so it should not be an offense. For example, def m(...) = obj.==(...) would be autocorrected as def m(...) = obj ==(...) which is a syntax error, as would be def m(...) = obj == (...) and def m(...) = obj == ... Before submitting the PR make sure the following are checked: [x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences. [x] Wrote good commit messages. [x] Commit message starts with [Fix #issue-number] (if the related issue exists). [x] Feature branch is up-to-date with master (if not - rebase it). [x] Squashed related commits together. [x] Added tests. [x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code. [x] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details. Closes #11377 Thanks!
2025-04-01T04:35:23.958215
2023-08-25T18:14:09
1867498391
{ "authors": [ "jcouball", "srem1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10462", "repo": "ruby-git/ruby-git", "url": "https://github.com/ruby-git/ruby-git/issues/666" }
gharchive/issue
--allow-empty how add oprions --allow-empty to commit or --allow-empty-message # assuming your current directory is in a git working directory g = Git.open('.') message = '' g.commit(message, allow_empty: true, allow_empty_message: true) why not g.config() ? this is not commit, I need setup for whole repo This library does not provide that capability.
2025-04-01T04:35:24.010198
2023-09-13T11:43:14
1894349442
{ "authors": [ "eregon", "seven1m" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10463", "repo": "ruby/yarp", "url": "https://github.com/ruby/yarp/issues/1461" }
gharchive/issue
How to differentiate WhileNode in modifier form What is the way to determine if a WhileNode was in modifier form? irb(main):001:0> n = YARP.parse('foo while bar').value.statements.body.first => @ WhileNode (location: (0...13)) ... irb(main):002:0> ls n YARP::WhileNode#methods: accept begin_modifier? child_nodes closing closing_loc comment_targets copy deconstruct deconstruct_keys inspect keyword keyword_loc predicate set_newline_flag statements YARP::Node#methods: location newline? pretty_print slice instance variables: @closing_loc @flags @keyword_loc @location @predicate @statements => nil irb(main):003:0> n.closing => nil The only difference I can see is that WhileNode#closing is nil (vs "end"), but that doesn't seem appropriate to use. Should there be (or is there already) a method to determine if the modifier form was used? I suppose the same question goes for UntilNode. Via LoopFlags Ah sorry I thought you meant checking whether it's a do/while vs a while loop (there is LoopFlags.BEGIN_MODIFIER for that). $ bin/parse -e 'foo while bar' @ ProgramNode (location: (0...13)) ├── locals: [] └── statements: @ StatementsNode (location: (0...13)) └── body: (length: 1) └── @ WhileNode (location: (0...13)) ├── keyword_loc: (4...9) = "while" ├── closing_loc: ∅ ├── predicate: │ @ CallNode (location: (10...13)) │ ├── receiver: ∅ │ ├── call_operator_loc: ∅ │ ├── message_loc: (10...13) = "bar" │ ├── opening_loc: ∅ │ ├── arguments: ∅ │ ├── closing_loc: ∅ │ ├── block: ∅ │ ├── flags: variable_call │ └── name: "bar" ├── statements: │ @ StatementsNode (location: (0...3)) │ └── body: (length: 1) │ └── @ CallNode (location: (0...3)) │ ├── receiver: ∅ │ ├── call_operator_loc: ∅ │ ├── message_loc: (0...3) = "foo" │ ├── opening_loc: ∅ │ ├── arguments: ∅ │ ├── closing_loc: ∅ │ ├── block: ∅ │ ├── flags: variable_call │ └── name: "foo" └── flags: I think you could check via while_node.statements.location.start_offset < while_node.predicate.location.start_offset. And the same for UntilNode and also IfNode and UnlessNode. Note that semantically there is 0 difference between foo while bar and while bar; foo; end. Yes, I was confused myself. My question actually should have been: how do you know if it's a begin while vs a regular while. Yep, begin_modifier? is actually what I need. Thank you @eregon! ❤️
2025-04-01T04:35:24.019100
2022-09-09T20:32:07
1368293949
{ "authors": [ "cielf", "dorner", "edwinthinks" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10464", "repo": "rubyforgood/human-essentials", "url": "https://github.com/rubyforgood/human-essentials/pull/3143" }
gharchive/pull-request
3088 distribution pdf Resolves #3088 Description This updates the output of the distribution PDF when there is an associated request. I've done a minor refactor of the LineItem class to accommodate the requested changes. Please see screenshots. The "request" version of it is missing a border on the bottom right of the data table - I couldn't figure out how to put it back and I ran out of time. We can decide if that's a blocker or if we can merge as is. Type of change New feature (non-breaking change which adds functionality) How Has This Been Tested? Local/unit tests Screenshots Thanks @dorner -- I probably won't get to this until at least Wednesday, because life. Still pending some fixes and after that I think we mergey merge Ack, forgot to push my commit! 😮 @edwinthinks I think this should be good now. @edwinthinks thanks for the QA - it should be fixed now! Awesome! Thanks @dorner just QA'ed this and it works. A bit of a side thought for a future issue, perhaps we can use something like https://adminlte.io/themes/v3/pages/examples/invoice.html from the theme which allows us to consolidate the printing and the show page of distributions. I figure this data could be helpful outside of the PDF. Lemme know what you think. That looks interesting! Definitely worth looking into next time we do something like this.
2025-04-01T04:35:24.046625
2017-07-02T15:06:06
240015532
{ "authors": [ "tyranja", "zaziemo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10465", "repo": "rubymonsters/speakerinnen_liste", "url": "https://github.com/rubymonsters/speakerinnen_liste/pull/588" }
gharchive/pull-request
Update all the gems https://github.com/rubymonsters/speakerinnen_liste/issues/470 Too many gems :( Travis checks are failing. You have to update the ruby version in the travis.yml as well I think. I am not yet ready, just half way through.
2025-04-01T04:35:24.071138
2023-03-02T12:50:07
1606752734
{ "authors": [ "maany" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10466", "repo": "rucio/webui", "url": "https://github.com/rucio/webui/issues/134" }
gharchive/issue
Fix storybook deployment Description The deploy-to-gh-pages script is broken since #117. It expects the storybook build to be stored in a docs directory, which we do not do anymore. The solution would be to modify the package.json to build storybook into the docs directory. The would result in successful deployments of storybook to rucio.cern.ch/webui Steps to reproduce No response Rucio WebUI Version No response Additional Information No response @ThePhisch could you work on this while I am away? Already fixed by @ThePhisch
2025-04-01T04:35:24.124191
2014-12-16T11:17:29
52100770
{ "authors": [ "andrewchambers", "nickzuck" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10467", "repo": "rui314/8cc", "url": "https://github.com/rui314/8cc/issues/36" }
gharchive/issue
segfault Put this in test.c long long a = 0LLU; run 8cc -S ./test.c. -> segfault. Hey there can you please give the link to the file of which you are talking about What file? the contents of the file are what I showed. Simply that single line. "What file? the contents of /*the file */ are what I showed. Simply that single line." According to your answer ...the file in the /* */ ......The file which you are talking about
2025-04-01T04:35:24.128680
2021-01-20T14:22:37
790021317
{ "authors": [ "ruimarinho", "torkelrogstad" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10468", "repo": "ruimarinho/docker-bitcoin-core", "url": "https://github.com/ruimarinho/docker-bitcoin-core/pull/108" }
gharchive/pull-request
CI: remove deprecated set-env, use build outputs set-env was removed a while ago from GitHub actions, because it was a securit risk. We replace this with the recommended workaround to get CI working again. @ruimarinho a bit unsure of what's going on here. Is the docker password empty because I'm submitting a PR from a fork, or because you've removed it? @ruimarinho a bit unsure of what's going on here. Is the docker password empty because I'm submitting a PR from a fork, or because you've removed it? Hmm, I see DOCKER_HUB_PASSWORD is correctly set so I need to look at what's going on. It might if you open a PR from this repo, with the same changes, I know that there are some restrictions on reading secrets from forks It might if you open a PR from this repo, with the same changes, I know that there are some restrictions on reading secrets from forks Thank you @torkelrogstad 🙏 fixed in https://github.com/ruimarinho/docker-bitcoin-core/pull/112.
2025-04-01T04:35:24.136677
2024-10-01T13:08:59
2559181366
{ "authors": [ "LawrenceGB", "marcusschiesser" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10469", "repo": "run-llama/LlamaIndexTS", "url": "https://github.com/run-llama/LlamaIndexTS/pull/1282" }
gharchive/pull-request
fix: ensure id_ is correctly passed during creation Fixes: #1278 This PR addresses issue #1278 where the id_ field was not being correctly passed as the id when creating a TextNode. This caused the upsert operation to the vector database to rely on the idGenerator instead of using the provided document ID, if available. Changes: Updated the logic for creating TextNode to ensure the id_ is set as the document ID when available. This ensures that when upserting to the vector database, the correct document ID is used, avoiding the generation of unnecessary random IDs. This fix improves consistency in ID handling, ensuring that document IDs are properly retained in the vector database. @LawrenceGB and @himself65, this PR was causing us a little headache: A document is parsed into several nodes. This PR causes all these nodes to have the same ID as the parent. So we get duplicate IDs, and these IDs are no longer UUIDs (which is required by some vector DBs). I, therefore, reverted this PR in https://github.com/run-llama/LlamaIndexTS/pull/1311 To get the ID of the source document from a split node, please check node.relationships[NodeRelationship.SOURCE] @marcusschiesser Could you please help clarify how I can get the actual ID that is being set as I can't seem to get it working. I am creating a unique document like so, it's not using that ID but generates the random one when upserting to Pinecone and also the doc returned relationships are empty. new Document({ id_: `${fileId}-${index}`, text: page.text, metadata: { page: page.index } }) Thanks!
2025-04-01T04:35:24.139092
2024-05-16T07:11:20
2299542010
{ "authors": [ "monoboard1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10470", "repo": "run-llama/llama_index", "url": "https://github.com/run-llama/llama_index/issues/13530" }
gharchive/issue
[Question]: How can I create a tool that enables querying a specific pandas DataFrame and then using the query result to perform calculations through a custom function? Question Validation [X] I have searched both the documentation and discord for an answer. Question I need to develop a tool that allows me to query a specific pandas DataFrame and then apply a custom function to the query result. How can I achieve this functionality in an efficient manner? @dosu thanks but could you please give me an example how to build the agent that first query pandas dataframe with specific LLM using the "pandas_query_engine" and then use the query result into LLM custom calling function? @dosu Ok now please give me an example of FunctionTool that can query data from pandas dataframe and then use the query result in a custom function
2025-04-01T04:35:24.147436
2024-01-02T20:16:31
2062896571
{ "authors": [ "ji21" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10471", "repo": "run-llama/llama_index", "url": "https://github.com/run-llama/llama_index/pull/9801" }
gharchive/pull-request
Integrations/deepeval observability Description Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change. Fixes #9800 Requires deepeval (v0.20.43) Type of Change Please delete options that are not relevant. [x] New feature (non-breaking change which adds functionality) [x] This change requires a documentation update How Has This Been Tested? Please describe the tests that you ran to verify your changes. Provide instructions so we can reproduce. Please also list any relevant details for your test configuration [x] Added new notebook (that tests end-to-end) (on DeepEval repo: https://github.com/confident-ai/deepeval) Suggested Checklist: [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have made corresponding changes to the documentation [ ] I have added Google Colab support for the newly added notebooks. [x] My changes generate no new warnings [x] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes [x] I ran make format; make lint to appease the lint gods @logan-markewich thanks for the prompt review! happy new year
2025-04-01T04:35:24.150270
2017-12-28T22:44:35
285003059
{ "authors": [ "xiaods" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10472", "repo": "runconduit/conduit", "url": "https://github.com/runconduit/conduit/issues/98" }
gharchive/issue
cargo test came across failed. $ cargo test --all Compiling abstract-ns v0.4.2 Compiling env_logger v0.4.3 error[E0277]: the trait bound `std::sync::Arc<str>: std::convert::From<&str>` is not satisfied --> /Users/xiaods/.cargo/registry/src/github.com-1ecc6299db9ec823/abstract-ns-0.4.2/src/name.rs:67:23 | 67 | Ok(Name(value.into())) | ^^^^ the trait `std::convert::From<&str>` is not implemented for `std::sync::Arc<str>` | = help: the following implementations were found: <std::sync::Arc<T> as std::convert::From<T>> = note: required because of the requirements on the impl of `std::convert::Into<std::sync::Arc<str>>` for `&str` error: aborting due to previous error error: Could not compile `abstract-ns`. warning: build failed, waiting for other jobs to finish... error: build failed $ rustc -V rustc 1.20.0 (f3d6973f4 2017-08-27) the rust version should be 1.21+
2025-04-01T04:35:24.222508
2024-04-26T06:26:16
2265051362
{ "authors": [ "Ian-cmd-ops", "runeharlyk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10473", "repo": "runeharlyk/SpotMicroESP32-Leika", "url": "https://github.com/runeharlyk/SpotMicroESP32-Leika/issues/37" }
gharchive/issue
ESP32 Cam not working I can't get the my ESP32 cam to stream or view after I got my ESP32 Uploaded with via PIO. do I need the other modules attached first before it runs #define USE_PSRAM true #define USE_WIFI true #define WAIT_FOR_WIFI false #define USE_WEBSERVER true #define USE_WEBSERVER_SSL false #define USE_WEBSOCKET true #define USE_OAT false #define USE_NTP false #define USE_MDNS true #define USE_DNS_SERVER false #define USE_REMOTE_SERIAL false #define USE_LOW_POWER false #define USE_CAMERA true #define USE_MPU true #define USE_POWER_BUTTON true #define USE_USS true The camera stream should work out of the box. Is there anything in the browser console log? Bad ESP32 Cam
2025-04-01T04:35:24.224624
2024-06-04T00:08:46
2332272166
{ "authors": [ "Ben10164", "Unmoon", "aHooder", "capslock13" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10474", "repo": "runelite/plugin-hub", "url": "https://github.com/runelite/plugin-hub/pull/6125" }
gharchive/pull-request
Update advanced-raid-tracker Add TOA, Inferno, Colosseum Improved UI using icons and handling the new raids Various chart improvements and expansion adds Chart Creator Apologies for the size of this update; any future changes will not be nearly this large or as comprehensive in the future. No expectations of this being reviewed anytime soon. Are you sure you need this many changes? I don't have time to review a 20k+ line diff. Closing & re-opening to re-run checks. Any updates on the review process?
2025-04-01T04:35:24.225819
2020-03-28T14:25:59
589587143
{ "authors": [ "Pingvinszar2", "raiyni" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10475", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/11117" }
gharchive/issue
Miscellaneous: Inventory/Bank Slot binder. Self-explanatory. Locks your items/objects in your inventory/bank, or other interfaces. The option will prevent you from accidentally dragging/switching items while doing click-intensive tasks like stringing bows, crafting jewelleries, making arrows, etc. Anti drag works in bank now.
2025-04-01T04:35:24.228360
2022-08-03T14:18:44
1327294686
{ "authors": [ "Adam-", "ghostoo666" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10476", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/15436" }
gharchive/issue
Max hit hitsplat not interacting with Special Attack counter plugin Type Incorrect behavior Operating System Windows OS Version Win10 Bug description The special attack counter records either number of successful hits per weapon (dwh) or total damage in special attacks per weapon (bgs). With the new max hit hitsplats, if you hit your max hit during the special attack, it won't be recorded to the counters. Screenshots or videos No response RuneLite version No response Logs No response fixed in d81463e4ad60eceab47470519ac49c91eea53139
2025-04-01T04:35:24.230320
2018-07-30T09:08:26
345672580
{ "authors": [ "generiskk", "memementor" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10477", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/4563" }
gharchive/issue
Achievement diary typo In the hard lumbridge/draynor diary... says i need 68 farming but really i know i need 63 Since its under the same, I will add this too.
2025-04-01T04:35:24.233415
2019-03-21T13:14:45
423723001
{ "authors": [ "Abextm", "deathbeam", "yyruneliter" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10478", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/8278" }
gharchive/issue
Camera Zoom Settings Not Working Outer zoom limit, inner zoom limit, require control down in "Camera Zoom" settings do not work after Runescape update for March 21, 2019. Seems to be at default zoom levels. Vertical camera is functioning properly. Steps to reproduce the behavior: Go to 'Runelite' Login Click on 'Configuration' Scroll down to 'Camera Zoom' Click on 'Edit Plugin Configuration' See that settings do not work when changed (except 'Vertical camera') OS: Windows 10 RuneLite version: 1.5.16 Launcher version: 1.6 This was fixed in 776d27cc4772d9ad3c779a2799ecef9a12aa0904 Restart your client Restart your client
2025-04-01T04:35:24.246204
2020-12-29T01:39:48
775654028
{ "authors": [ "Adam-", "Nightfirecat", "chaticon" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10479", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/pull/12990" }
gharchive/pull-request
opponentinfo: make health bar more active Made the health bar continue to display while the opponent attacks the player. Useful for keeping track of enemy hp while recoiling. This looks like it would break whenever you are fighting something in a multi combat area and then had something else attack you. This looks like it would break whenever you are fighting something in a multi combat area and then had something else attack you. My bad, I had tested it in multi but clearly not well enough. Should correctly prioritise the NPC you're attacking now. The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent? This change very much would benefit from some tests being added, given its complexity. The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent? This change very much would benefit from some tests being added, given its complexity. The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent? This change very much would benefit from some tests being added, given its complexity. I added some tests to clarify behaviour in some ambiguous scenarios. To address the specific scenarios you brought up, you were partially right that I goofed the second scenario. It wouldn't start tracking the hellhound if it wasn't already active opponent, but if it was, it would stay the active opponent until that NPC died or some other NPC became active opponent. I've fixed that now. As for the first scenario you mentioned, I don't think it's as weird as you say. If the player is attacking an NPC, then that NPC will always be the active opponent regardless of what happens in multi. The only time the most recent attacker will become the active opponent is if the player is not attacking any NPC at all. That might sound odd, but if you're, say, cannoning while idle, that NPC is probably the most recent enemy your cannon hit, in which case it's behaving similarly to the current version of the plugin. Regardless, I hope the tests are to your liking and can help illustrate my explanation. Cheers. The behavior this has now seems very odd. Also, if a monster attacking you changed targets to another player--for instance, a hellhound attacking you became aggressive to a player running through your area--it would start tracking that no-longer-attacking-you hellhound as the active opponent? This change very much would benefit from some tests being added, given its complexity. I added some tests to clarify behaviour in some ambiguous scenarios. To address the specific scenarios you brought up, you were partially right that I goofed the second scenario. It wouldn't start tracking the hellhound if it wasn't already active opponent, but if it was, it would stay the active opponent until that NPC died or some other NPC became active opponent. I've fixed that now. As for the first scenario you mentioned, I don't think it's as weird as you say. If the player is attacking an NPC, then that NPC will always be the active opponent regardless of what happens in multi. The only time the most recent attacker will become the active opponent is if the player is not attacking any NPC at all. That might sound odd, but if you're, say, cannoning while idle, that NPC is probably the most recent enemy your cannon hit, in which case it's behaving similarly to the current version of the plugin. Regardless, I hope the tests are to your liking and can help illustrate my explanation. Cheers. While reviewing this, I've thought of another case which causes this code to fail, and I've written a brief test case which demonstrates it: @Test public void singleCombatIssue() { Player localPlayer = mock(Player.class), two = mock(Player.class), three = mock(Player.class); when(client.getLocalPlayer()).thenReturn(localPlayer); // TODO: find some unique var/condition representing a player being in multi-combat and mock it to be false here opponentInfoPlugin.onInteractingChanged(new InteractingChanged(localPlayer, two)); assertSame(two, opponentInfoPlugin.getLastOpponent()); // Another player tries to attack or cast magic on the player which causes interaction target changed from null -> player -> null in succession opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null)); assertSame(two, opponentInfoPlugin.getLastOpponent()); opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, localPlayer)); assertSame(two, opponentInfoPlugin.getLastOpponent()); // this line fails opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null)); assertSame(two, opponentInfoPlugin.getLastOpponent()); } Keep in mind this is, as far as I'm aware, the only case that would be affected by the player being in a multi-combat zone (or not, in this case) If I'm understanding correctly, the reason this appears to change the value of lastOpponent is because you didn't also mock the localPlayer.getInteracting() method to return 'two'. It uses that to check if the player is still attacking a different actor when an actor attacks them, to prevent lastOpponent from taking on the value of the actor they aren't attacking. I did a test for this from lines 73-85. While reviewing this, I've thought of another case which causes this code to fail, and I've written a brief test case which demonstrates it: @Test public void singleCombatIssue() { Player localPlayer = mock(Player.class), two = mock(Player.class), three = mock(Player.class); when(client.getLocalPlayer()).thenReturn(localPlayer); // TODO: find some unique var/condition representing a player being in multi-combat and mock it to be false here opponentInfoPlugin.onInteractingChanged(new InteractingChanged(localPlayer, two)); assertSame(two, opponentInfoPlugin.getLastOpponent()); // Another player tries to attack or cast magic on the player which causes interaction target changed from null -> player -> null in succession opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null)); assertSame(two, opponentInfoPlugin.getLastOpponent()); opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, localPlayer)); assertSame(two, opponentInfoPlugin.getLastOpponent()); // this line fails opponentInfoPlugin.onInteractingChanged(new InteractingChanged(three, null)); assertSame(two, opponentInfoPlugin.getLastOpponent()); } Keep in mind this is, as far as I'm aware, the only case that would be affected by the player being in a multi-combat zone (or not, in this case) If I'm understanding correctly, the reason this appears to change the value of lastOpponent is because you didn't also mock the localPlayer.getInteracting() method to return 'two'. It uses that to check if the player is still attacking a different actor when an actor attacks them, to prevent lastOpponent from taking on the value of the actor they aren't attacking. I did a test for this from lines 73-85. The tests do add full coverage of the changed code, though I'm still not 100% sure I understand this code inside and out. I'll review this more in depth at a later time. I've added some further tests and cleaned up the test class a bit. @chaticon, does this still look good to you? From some quick ingame testing (in addition to the full test coverage provided by this PR's tests) it seems to still work just fine. I've added some further tests and cleaned up the test class a bit. @chaticon, does this still look good to you? From some quick ingame testing (in addition to the full test coverage provided by this PR's tests) it seems to still work just fine. Sorry for the late response. Looks good to me. And I think I may have commented on something outdated that you'd already fixed, sorry about that. This throws an exception on login java.lang.NullPointerException: null at net.runelite.client.plugins.opponentinfo.OpponentInfoPlugin.onInteractingChanged(OpponentInfoPlugin.java:161) at net.runelite.client.eventbus.EventBus$Subscriber.invoke(EventBus.java:67) at net.runelite.client.eventbus.EventBus.post(EventBus.java:238) at net.runelite.client.callback.Hooks.post(Hooks.java:169) Additionally we should probably add comments to all of this logic since it looks complicated and I don't follow all of it. We may want to keep an eye on this since there is no despawn event hooks for the opponent actor, and it is still invoking getInteracting() on it. But I think it is probably okay.
2025-04-01T04:35:24.248557
2019-04-08T01:05:29
430215513
{ "authors": [ "xDemoN" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10480", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/pull/8442" }
gharchive/pull-request
World Map: Identify Both Shield of Arrav Quest Start Points Closes #8437 Let me know if you'd like the syntax to be different. Enum names updated @Nightfirecat
2025-04-01T04:35:24.253347
2023-08-24T20:33:41
1865844411
{ "authors": [ "austinvazquez", "vsiravar" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10481", "repo": "runfinch/finch-core", "url": "https://github.com/runfinch/finch-core/pull/142" }
gharchive/pull-request
fix: Change rootfs compression to gzip Issue #, if available: Description of changes: Change compression of rootfs bundle to gzip since windows does not come with zstd installed. Testing done: Yes, locally. [X] I've reviewed the guidance in CONTRIBUTING.md License Acceptance By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Do we need to update Makefile and bin/update-rootfs.sh now as well? See https://github.com/search?q=org%3Arunfinch .tar.zst&type=code Do we need to update Makefile and bin/update-rootfs.sh now as well? See https://github.com/search?q=org%3Arunfinch .tar.zst&type=code This is a good point. However the Makefile is updated when https://github.com/runfinch/finch-core/blob/main/.github/workflows/update-dependencies.yaml#L32 action runs. For the next run we still need https://github.com/runfinch/finch-core/blob/main/bin/update-rootfs.sh#L5 as the Makefile has a zst compressed tar. Once the change is merged, I will update the update-rootfs.sh to AARCH64_FILENAME_PATTERN="common/aarch64/finch-rootfs-production-arm64-[0-9].*\.tar.gz$" AMD64_FILENAME_PATTERN="common/x86-64/finch-rootfs-production-amd64-[0-9].*\.tar.gz$"
2025-04-01T04:35:24.257866
2023-10-03T23:35:23
1925108567
{ "authors": [ "pendo324", "vsiravar" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10482", "repo": "runfinch/finch", "url": "https://github.com/runfinch/finch/pull/594" }
gharchive/pull-request
feat: windows persistent disk and other misc fixes/refactors Issue #, if available: Description of changes: Adds persistent disk support to Windows This took more work than anticipated, in order to deal with non-Admin users pkg/disk/dpgo/ is brand new code, and should be a focus of the review pkg/winutil/run_windows.go is also new and requires careful review. This is what allows Finch to run as the regular user, except for when it needs Admin access to call diskpart (to create the persistent disk) Fix paths in nerdctl_config_applier to make the post-boot/init shelling work Added winres to allow the finch.exe to have metadata attached to it. This is WIP, need final icons and descriptions etc. Large (in terms of lines changed) refactor of pkg/path/finch.go, but it should have no impact on functionality (needs careful review) Fixed the Makefile's clean target for Windows Most of the other changes are just noise from refactoring (like, literally renaming things). Nothing major, but take a look if possible. Sorry the diff is so large Testing done: [x] I've reviewed the guidance in CONTRIBUTING.md License Acceptance By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Really cool PR, I left a few minor comments from the first pass, especially some files missing licenses.
2025-04-01T04:35:24.259242
2023-12-04T22:16:20
2024860212
{ "authors": [ "elboy3" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10483", "repo": "rungalileo/dataquality", "url": "https://github.com/rungalileo/dataquality/pull/814" }
gharchive/pull-request
fix: disallow emb cols as metadata https://app.shortcut.com/galileo/story/9579/dq-adding-x-y-data-x-or-data-y-as-metadata-succeeds-but-should-not Customer ran into issues when logging x as metadata @setu4993 yes we were getting quite an odd error on the runners side!
2025-04-01T04:35:24.272651
2023-12-19T21:54:56
2049495631
{ "authors": [ "Sirri69", "alpayariyak", "christopherwolfsaal", "martinkallstrom" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10484", "repo": "runpod-workers/worker-vllm", "url": "https://github.com/runpod-workers/worker-vllm/issues/24" }
gharchive/issue
Runpod I tried to load the docker image and use in runpod. But i got the following message. 023-12-19T21:52:21.224549804Z Traceback (most recent call last): 2023-12-19T21:52:21.224594641Z File "/handler.py", line 3, in 2023-12-19T21:52:21.224600084Z import runpod 2023-12-19T21:52:21.224605461Z File "/usr/local/lib/python3.11/dist-packages/runpod/init.py", line 6, in 2023-12-19T21:52:21.224610387Z from . import serverless 2023-12-19T21:52:21.224616241Z File "/usr/local/lib/python3.11/dist-packages/runpod/serverless/init.py", line 16, in 2023-12-19T21:52:21.224627070Z from .modules import rp_fastapi 2023-12-19T21:52:21.224631744Z File "/usr/local/lib/python3.11/dist-packages/runpod/serverless/modules/rp_fastapi.py", line 10, in 2023-12-19T21:52:21.224636270Z from fastapi import FastAPI, APIRouter 2023-12-19T21:52:21.224640611Z File "/usr/local/lib/python3.11/dist-packages/fastapi/init.py", line 7, in 2023-12-19T21:52:21.224644971Z from .applications import FastAPI as FastAPI 2023-12-19T21:52:21.224649164Z File "/usr/local/lib/python3.11/dist-packages/fastapi/applications.py", line 16, in 2023-12-19T21:52:21.224653711Z from fastapi import routing 2023-12-19T21:52:21.224691454Z File "/usr/local/lib/python3.11/dist-packages/fastapi/routing.py", line 22, in 2023-12-19T21:52:21.224716007Z from fastapi import params 2023-12-19T21:52:21.224724461Z File "/usr/local/lib/python3.11/dist-packages/fastapi/params.py", line 5, in 2023-12-19T21:52:21.224816202Z from fastapi.openapi.models import Example 2023-12-19T21:52:21.224821635Z File "/usr/local/lib/python3.11/dist-packages/fastapi/openapi/models.py", line 4, in 2023-12-19T21:52:21.224825700Z from fastapi._compat import ( 2023-12-19T21:52:21.224829829Z File "/usr/local/lib/python3.11/dist-packages/fastapi/_compat.py", line 20, in 2023-12-19T21:52:21.224913810Z from fastapi.exceptions import RequestErrorModel 2023-12-19T21:52:21.224923693Z File "/usr/local/lib/python3.11/dist-packages/fastapi/exceptions.py", line 3, in 2023-12-19T21:52:21.224950980Z from pydantic import BaseModel, create_model 2023-12-19T21:52:21.224958616Z File "/usr/local/lib/python3.11/dist-packages/pydantic/init.py", line 372, in getattr 2023-12-19T21:52:21.225128094Z module = import_module(module_name, package=package) 2023-12-19T21:52:21.225144639Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2023-12-19T21:52:21.225149112Z File "/usr/lib/python3.11/importlib/init.py", line 126, in import_module 2023-12-19T21:52:21.225160188Z return _bootstrap._gcd_import(name[level:], package, level) 2023-12-19T21:52:21.225196585Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2023-12-19T21:52:21.225211329Z File "/usr/local/lib/python3.11/dist-packages/pydantic/main.py", line 11, in 2023-12-19T21:52:21.225294129Z import pydantic_core 2023-12-19T21:52:21.225303857Z File "/usr/local/lib/python3.11/dist-packages/pydantic_core/init.py", line 30, in 2023-12-19T21:52:21.225314032Z from .core_schema import CoreConfig, CoreSchema, CoreSchemaType, ErrorType 2023-12-19T21:52:21.225320699Z File "/usr/local/lib/python3.11/dist-packages/pydantic_core/core_schema.py", line 15, in 2023-12-19T21:52:21.225353524Z from typing_extensions import deprecated 2023-12-19T21:52:21.225360153Z ImportError: cannot import name 'deprecated' from 'typing_extensions' (/usr/local/lib/python3.11/dist-packages/typing_extensions.py) Yep, known issue because of locking dependencies runpod==1.4.2 Absolutely silly behaviour. Pliz fix. ✨ Finding same Should be fixed now Will there be a new docker image or where it is fixed?
2025-04-01T04:35:24.275770
2017-06-30T22:11:35
239900189
{ "authors": [ "d3zd3z", "utzig" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10485", "repo": "runtimeco/mcuboot", "url": "https://github.com/runtimeco/mcuboot/pull/77" }
gharchive/pull-request
Always validate slot 0 when requested The MCUBOOT_VALIDATE_SLOT0 feature only verifies the signature when there is no swapping happening. The assumption was that if there is a swap being done, the code will verify the signature of slot 1 before doing the slot. However, either due to bugs, or intentional trickery, it may be possible to confuse the code into continuing a swap operation. If the data is modified before this, the bootloader can be tricked into booting the resulting image in slot 0 without having verified the signature. Fix this by always verifying slot 0's signature before booting it. JIRA: MCUB-64 Signed-off-by: David Brown<EMAIL_ADDRESS> Don't merge yet, until we can create some enhancements to the sim that will allow us to test this. @d3zd3z I'm doing some "final" changes for #81 which affect some of the same lines as this patch. Not sure there's something remaining to be done, otherwise would be glad if it is merged soon!
2025-04-01T04:35:24.286458
2023-10-29T17:47:15
1967090341
{ "authors": [ "Anmol-Baranwal", "aslezar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10486", "repo": "rupali-codes/LinksHub", "url": "https://github.com/rupali-codes/LinksHub/pull/2145" }
gharchive/pull-request
Added link to material tailwind Fixes Issue Fixes #2144 Changes proposed New UI Library Material Tailwind added to the database of frontend ui- libraries @rupali-codes The issue of not showing in subcategories. In this case, the resource is limited.
2025-04-01T04:35:24.288144
2024-06-19T17:13:48
2362831320
{ "authors": [ "Riyazul555" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10487", "repo": "rupali-codes/LinksHub", "url": "https://github.com/rupali-codes/LinksHub/pull/2397" }
gharchive/pull-request
Yaak added in Backend Testing Fixes Issue #2380 Changes proposed Screenshots Note to reviewers @rupali-codes please review this pR Thanks
2025-04-01T04:35:24.289527
2019-06-30T18:23:23
462432380
{ "authors": [ "bernardesrodrigo", "ramezanifar" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10488", "repo": "ruscito/pycomm", "url": "https://github.com/ruscito/pycomm/issues/60" }
gharchive/issue
Should it work with AB Emulator 5000? Hi, Greats for the library, I´m trying to make it work with RSLogixEmulator 5000 running in remote machine over ethernet. Should it work like that or only with real hardware? When reading a tag I´m getting the error: "pycomm.cip.cip_base.CommError: can only concatenate str (not "bytes") to str" RSLogixEmulator 5000 does not let read and write. Instead use SoftLogix5800. It makes your PC like a PLC.
2025-04-01T04:35:24.303470
2016-11-12T03:09:49
188888785
{ "authors": [ "edtanous", "ruslo" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10489", "repo": "ruslo/hunter", "url": "https://github.com/ruslo/hunter/pull/580" }
gharchive/pull-request
Fix an issue in OpenSSL build when cross compiling The commit 44c681e1a551e0599a8ca170cb0cfcc19c4a5d54 seems to have broken cross compiling builds, as openssl ./config seems to require both the CC and MACHINE variables to be set, otherwise it seems to default to build host compiler and machine variables. Someone appears to have already seen this issue when cross compiling for android, and fixed it in commit bb657ff41c874ca3bba1ca394ff1ea09ee8049b3. This patch moves the MACHINE argument and adds it to ALL build types, not just android. I'm unclear on the architecture of the PACKAGE_INTERNAL_DEPS_ID, but I saw it being incremented for other similar changes, so I assume this will require an update as well. If I'm wrong, let me know and I will update the PR. I'm unclear on the architecture of the PACKAGE_INTERNAL_DEPS_ID I've added some details here: https://docs.hunter.sh/en/latest/reference/user-modules/hunter_download.html The commit 44c681e seems to have broken cross compiling builds So what platforms do we fix by this change? Testing: https://ci.appveyor.com/project/ingenue/hunter/build/1.0.1165 https://travis-ci.org/ingenue/hunter/builds/175285201 This fixes cross compiling from Linux x86_64 to Armv6 32 bit. My CMAKE_TOOLCHAIN_FILE looks like the following SET(CMAKE_SYSTEM_NAME Linux) SET(CMAKE_SYSTEM_VERSION 1) set(CMAKE_SYSTEM_PROCESSOR armv6) # specify the cross compiler SET(CMAKE_C_COMPILER arm-linux-gnueabi-gcc) SET(CMAKE_CXX_COMPILER arm-linux-gnueabi-g++) # where is the target environment SET(CMAKE_FIND_ROOT_PATH /usr/arm-linux-gnueabi) # search for programs in the build host directories SET(CMAKE_FIND_ROOT_PATH_MODE_PROGRAM NEVER) # for libraries and headers in the target directories SET(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY) SET(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY) This is on an Ubuntu 16.04 machine with a 5.4 gcc toolchain standard, although it's not clear if that matters. it's not clear if that matters I will merge it because it doesn't break old behavior and seems to fix the issue for you. However I can't guarantee that there will be no problem with your toolchain in future. If you want to be sure you have to extend testing. Uploading: https://ci.appveyor.com/project/ingenue/hunter/build/1.0.1171 https://travis-ci.org/ingenue/hunter/builds/175429374 Released: https://github.com/ruslo/hunter/releases/tag/v0.16.27
2025-04-01T04:35:24.313403
2021-07-12T13:52:53
942088120
{ "authors": [ "0x2b00b1e5", "gwenn", "thomcc" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10490", "repo": "rusqlite/rusqlite", "url": "https://github.com/rusqlite/rusqlite/issues/992" }
gharchive/issue
Parsing keys for SQLCipher databases Hi, I am probably just missing something but I tried to connect to a sqlcipher database but it didn't work. I first did conn.execute_batch("PRAGMA key = 'test123';") and then conn.execute_batch("SELECT count(*) FROM sqlite_master;") as a test, but this did not decrypt the database correctly for some reason (returned message: file is not a database). I have tested the decryption using the sqlcipher CLI tool, which just worked fine. I am using the sqlcipher-bundled feature. execute_batch should be fine but you should use: conn.pragma_update(None, "key", &passphrase)?; And do you know which kind of error is expected when the passphrase is incorrect ? Page https://www.zetetic.net/sqlcipher/sqlcipher-api/#PRAGMA_key doesn't help. https://discuss.zetetic.net/t/file-is-not-a-database/3320 https://discuss.zetetic.net/t/error-file-is-not-a-database/4434 let cipher_version: String = db.pragma_query_value(None, "cipher_version", |row| row.get(0))?; And do you know which kind of error is expected when the passphrase is incorrect ? It gives the error "not a database" when you try to use it (e.g. on the first query, not on the PRAGMA key). It is somewhat weird... This sounds like your key is wrong to me. This wouldn't explain why it works on the CLI though. Perhaps the page size doesn't match though (it must match for sqlcipher, and perhaps it has a different default on the CLI)... execute_batch should be fine but you should use: conn.pragma_update(None, "key", &passphrase)?; Tried that too, same result. Just to clarify that the conn.execute_batch("PRAGMA key = 'test123;'); ... does succeed. The problems start when trying to touch/read/alter the database. Perhaps the page size doesn't match though (it must match for sqlcipher, and perhaps it has a different default on the CLI)... Maybe? How would I change the page size? Update: used the rusqlite api to create a db instead. This does work. So I guess there is indeed a pagesize/encoding issue there.
2025-04-01T04:35:24.315481
2020-11-11T03:50:50
740435956
{ "authors": [ "Veykril", "jrmuizel" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10491", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/issues/6527" }
gharchive/issue
Renaming type renames variables Given the following code: struct tag { x: i32 } fn main() { let tag = Some(4); if let Some(tag) = tag { dbg!(tag); } } renaming struct tag to struct Tag incorrectly produces: struct Tag { x: i32 } fn main() { let tag = Some(4); if let Some(tag) = Tag { dbg!(Tag); } } This seems to have been fixed.
2025-04-01T04:35:24.317671
2022-02-03T12:20:06
1123019532
{ "authors": [ "flodiebold" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10492", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/pull/11399" }
gharchive/pull-request
Fix assoc type shorthand from method bounds In code like this: impl<T> Option<T> { fn as_deref(&self) -> T::Target where T: Deref {} } when trying to resolve the associated type T::Target, we were only looking at the bounds on the impl (where the type parameter is defined), but the method can add additional bounds that can also be used to refer to associated types. Hence, when resolving such an associated type, it's not enough to just know the type parameter T, we also need to know exactly where we are currently. This fixes #11364 (beta apparently switched some bounds around). bors r+
2025-04-01T04:35:24.318861
2020-02-18T11:26:40
566828109
{ "authors": [ "flodiebold", "lnicola" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10493", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/pull/3215" }
gharchive/pull-request
Exclude methods from non-parameter types introduced by generic constraints Fixes #3184. r? @flodiebold bors r+