added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:40:15.169427
2014-08-13T09:24:59
40136988
{ "authors": [ "RQuadling", "evought", "maxgalbu", "robmorgan", "rquadling", "shadowhand" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10265", "repo": "robmorgan/phinx", "url": "https://github.com/robmorgan/phinx/issues/287" }
gharchive/issue
Make the change method the default and comment out up/down This is potentially a big change moving forwards. In newly created migrations up and down would be commented out by default and change would become the default method. This is the opposite of the present behaviour. Background The change method is far simpler to use and suffices for most users. The up / down methods are more flexible but should be used less. Does change() work correctly with data migration, not just schema migration? In other words, if I modify a column in a table and have to adjust existing rows based on a SELECT, won't I have to create a custom down() method? Or if I drop a table/column in a migration and need to recreate it on rollback? Of course, you are just talking about changing the default here and I will still be able to override it by creating an up()/down() method when I need it? I am looking at moving to phinx on a project and am trying to understand what the approximate limits are to understand the impact of planned changes and how I should organize my initial setup. @evought see the note about limitations at the end of http://docs.phinx.org/en/latest/migrations.html#the-change-method "see the note about limitations..." Understood, thank you. Been experimenting with the behavior this evening and it seems fairly logical. With regard to the current state, and the proposal to make change() the default operation, would it be advantageous (and more defensive) to validate that the migration can be rolled-back BEFORE the migration takes place? Maybe by using a variant of the adapters that only allow the reversible methods? So, up() and down() would use the current adapter and I am at will to do whatever I want. But change() would use a limited adapter that simply does not have any method that cannot be reversed. For me, the advantage here is that the developer is protected from making non-reversible mistakes. Make their/our lives easier. Some people run Phinx on massive production systems with 100's of tables and terrabytes of data. I'd sleep much better knowing they are in full control of their migrations (even if they write something nasty). If Phinx started to become opinionated about migrations I'm worried it might lead to problems. Better to keep it simple. p.s moving to 0.5.0 now. I agree with @RQuadling, those people would not have full control on their migration because they would unknowingly make a change in the change() method that can't be reverted. The full control is having up() and down() and being able to make mistakes because of my code and not phinx's. What I was saying is that the methods that can be called in the change() method should be limited to reversible ones. If a user requires something that cannot be automatically reversed in a migration then they need to use up() and down(). done in 5ffb2324f5789c8919271dd2c6bb0b448a0b02a1 I suppose we need to see a use case of a non reversible change() call. I'm still not sure that every use of change IS reversible. up()/down() IS the "in the hands of the developer" control that is required. On the other hand, I never use change(), so meh!
2025-04-01T06:40:15.189828
2016-06-21T19:33:28
161520658
{ "authors": [ "LettError", "benkiel", "jackjennings", "typesupply" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10266", "repo": "robofab-developers/fontParts", "url": "https://github.com/robofab-developers/fontParts/issues/40" }
gharchive/issue
Deprecate showInterface from Font initialization? I'm not sure how desirable or immediately necessary this change is, but given that there's potential for a clean break with the roboFab implementation, I'm opening this issue as a point of discussion. Usually when you see a boolean option for any method, it's either a signal that there is a method that is doing double duty, or domain model that hasn't been defined. I'm wondering if this isn't a place where another object might be introduced in order to better encapsulate the design space that fontParts works within: Font('foo.ufo', showInterface=True) # vs. proposed: Editor().open(Font('foo.ufo')) I'd propose introducing some kind of Editor or Application concept that would be a generic container that explicitly interacts with whichever application is the subject of a fontParts implementation. I think that this could be a cleaner separation of responsibilities, and would allow the implementation of Font to be ignorant of the details of the manipulating the UI, which I like. There are ramifications for testing (isolating UI concerns to fewer areas), and for conceptually decoupling interface from editor (the latter of which may or may not include an interface that can be shown, but still have the concept of opening a font). It could also mean that a specialized Editor object could open different kinds of resources (e.g. just glyphs, features, etc.). It would still be possible to maintain the old showInterface API (with a deprecation warning?). Thoughts? My two cents: Most folks writing scripts will have a much easier time with: Font('foo.ufo', showInterface=True) I see the appeal of Editor().open(Font('foo.ufo')) on a higher level, but would not want the first to go away. This is based on my experience teaching typeface designers to write code, they are not out to be the world's best programmers, they just want to easily understand something and get on with it. I guess it comes down somewhat to the philosophy of the library… Keeping both feels like it violates the Pythonic tenant of one right way, so I suppose this could be something that just can't change. Hm, I kinda like the general idea. I think there are some gnarly backwards compatibility and environment things that we need to think through. For example, not all environments have an interface and not all environments have a no-interface way of opening a font. OpenFont("foo.ufo") in FontLab and RoboFont will show an interface. The same thing in NoneLab will not. Let me think on it... I was looking through the OpenFont and friends after I opened this issue. For me these have always felt like convenience functions wrapping the mid-level interfaces, to mask some of the underlying implementation details. Perhaps it's worth considering what the primary interface to the library is, and which APIs are changeable, even if this proposal goes too far in introducing new functionality. The the case of NoneLab Editor#open could simply be a no-op. I also don't want to imply that Editor is the most appropriate name for this—nor Application. I can think of several situations where one or the other may not be the appropriate term for the context. Interface could be general enough, or at least on the opposite end from Editor I would agree with @benkiel. One thing to consider is that these calls are part of almost all robofab centric scripts. It would break whatever portability they have. "Clean break" is expensive. You could kind of use the same reason against the creation of this library entirely. I guess my question is still, if portability is a goal, then which interfaces are changeable and which ones are frozen? Is there a priority among interfaces? One of the stated goals is to remove cruft, and I think that—arguably—this keyword arg is a wart, in not at least crufty. The second question that I have would be, if these changes are not possible—or not a priority—now, is it possible they could be in the future? Is there a plan to make a long-term roadmap possible? I think that having this library versioned and available on pypi would go a long way to being about to say "now that this is v.2 here are the breaking changes" and offer deprecation warnings up to that point. I guess my question is still, if portability is a goal, then which interfaces are changeable and which ones are frozen? There isn't a hard and fast rule. It's more of an intuitive, "How many scripts do we think this will break?" vs, "Ugh, I can't live with this mistake from 2003 anymore so everyone can just deal with the break." kind of thing. The second question that I have would be, if these changes are not possible—or not a priority—now, is it possible they could be in the future? Yes. Right now the focus is on trying to get reasonable compatibility with RoboFab so we can stop answering RoboFab tech support questions. After that, we're going to need to bring the API up to date with new stuff. All that said, we need to make sure that we keep in mind that this is supposed to be a simple scripting API for non-programmers. I've been struggling with that from time to time as evidenced by the looks I got when I proposed a couple of weird things at Typographics. 😕 I've been thinking about an Editor object a bit more. It could be useful for abstracting some of the other common interface interactions that we have needed. For example, robofab.interface.all.dialogs. <bigboldtype>This was a huge headache in RoboFab so I'm sweating a bit even bringing it up.</bigboldtype> But, I know how valuable being able to have a common API for triggering extremely simple user interactions from my experiences writing scripts that had to work equally well in RoboFont and Windows FontLab. Something like this: app = Application() path = app.showGetFile(...) go = app.askYesNo(...) # etc. <bigboldtype>FontParts would not implement any of this. Ever. At all. It would only define the API.</bigboldtype> Still, I'm shaking a bit just thinking about it. It's a short road from here to <insert horrible "universal" windowing API name here>. We've put UI things into fontParts.ui, this is explicitly for an implementation of fontParts to handle, fontParts is not going to make ui components. These are simple, for common things that a scripter would likely want, anything fancier can be done with something else (dialogKit, vanilla, etc). As such, I'm going to close this for now, but it may be worth coming back to the idea of an Editor object at some point.
2025-04-01T06:40:15.239326
2016-04-29T04:15:51
151790041
{ "authors": [ "gharris1727" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10267", "repo": "robotics-at-maryland/qubo", "url": "https://github.com/robotics-at-maryland/qubo/pull/27" }
gharchive/pull-request
Finished main development of the DVL drivers I also moved the drivers directory up a level like was in the refactor notes. Also need to work on the actual build integration, at the moment the makefile just spits things into the bin directories next to the source, which probably isn't the best idea.
2025-04-01T06:40:15.261497
2020-07-05T02:30:48
650987842
{ "authors": [ "benkiel", "gferreira", "jpt", "typemytype" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10268", "repo": "robotools/fontParts", "url": "https://github.com/robotools/fontParts/issues/534" }
gharchive/issue
Path intersections between (x,y, x,y) Hi, The Glyphs API has intersectionBetweenPoints and while looking for a UFO equivalent, I found robofab.path.intersect; I was wondering if this feature is missing from fontParts or if it was intentionally left out as out-of-scope? Or maybe it is implemented with a different name and I just missed it? A lot of half-working apis where removed in the transition between robofab and fontParts: easier to maintain, smaller codebase, ... . The path module was not ported over to fontParts. BooleanOperations has getIntersections of list of contours. see https://github.com/typemytype/booleanOperations from booleanOperations import getIntersections f = CurrentFont() g1 = f["A"] g2 = f["B"] result = getIntersections(g1.contours + g2.contours) print(result) good luck! Hi @typemytype - thank you, but I think I did not do a great job wording my question - or possibly I misunderstood robofab.path.intersect - but this is not the kind of intersection I am looking for. Rather than the intersection of glyphs themselves, I'm trying to run a straight line run across a glyph from arbitrary point (x,y) to another arbitrary point (x,y) - pretty much the measurement or slice tool but without creating new points, and without a GUI Hopefully this demonstrates the values I am looking for in such an API - imagine I were to run some imaginary code like getLineIntersection((-1000,200), (1000,200)) and would get something like {51,200}, {263, 200}, {446, 200}, {658, 200} as an output. I would love to know what you recommend! (And if it is something that might make sense in fontParts, I would be happy to try and contribute) if you are in a RoboFont world then you can use IntersectGlyphWithLine from mojo.tools import IntersectGlyphWithLine result = IntersectGlyphWithLine(CurrentGlyph(), ((startX, startY), (endX, endY)) print(result) see https://robofont.com/documentation/building-tools/api/mojo/mojo-tools/#mojo.tools.IntersectGlyphWithLine good luck! Something along these lines could work in fontShell on the command line: `from booleanOperations import getIntersections from fontParts.fontshell.glyph import RGlyph def getLine(sx,sy,ex,ey): line = RGlyph() line.width = ey pen = line.getPen() pen.moveTo((sx, sy)) pen.lineTo((ex,ey)) pen.closePath() return line.contours def getLineIntersection(glyph, sx,sy,ex,ey): return getIntersections(getLine(sx,sy,ex,ey) + glyph.contours)` I've tested a bit, didn't quite get it working, but that's the idea Thanks again for the quick response! That's exactly the functionality I am looking for, but outside of RoboFont - hopefully I would just be able to point a Python script to a UFO file @jpt It's not in fontParts, as @typemytype said; it didn't make the move over from roboFab. Hopefully the bit above helps set you down the path to do it with a script. Thank you @benkiel - my previous reply happened at the same time as yours; I see now how the two libraries can work together with RGlyph as the basis -- I'll get experimenting! Closing the issue, thanks again @jpt note that that code doesn't pull in a font, but you can do so with an import of RFont. there’s also MarginPen in fontPens, but it only works for vertical or horizontal lines — see this example Ah, that's likely easier, @jpt if you draw it first into a transformPen with the correct angle, then it also works of arbitrary lines!! and done: from fontPens.marginPen import MarginPen from fontTools.pens.transformPen import TransformPen from fontTools.misc.transform import Transform import math def intersectGlyphWithLine(glyph, line): # expand the line into seperate variables (startx, start), (endx, endy) = line # calculate the diff yDiff = endy - starty xDiff = endx - startx # use the atan2 to get the angle of the line angle = math.atan2(yDiff, xDiff) # create a marginPen, starting from the starty pen = MarginPen(g.font, starty, isHorizontal=True) # create an empty transform matrix matrix = Transform() # set the starting point tot the origin matrix = matrix.translate(startx, starty) # rotate the matrix around the origin matrix = matrix.rotate(-angle) # translate back matrix = matrix.translate(-startx, -starty) # draw in a transform pen with the margin pen glyph.draw(TransformPen(pen, matrix)) # get the inverse matrix inverseMatrix = matrix.inverse() # create an empty list of possible intersections intersections = [] # calculate the min max of the x's minx = min((startx, endx)) maxx = max((startx, endx)) # get all margin x values for x in pen.getAll(): # convert the straight line back to with an inverse matrix x, y = inverseMatrix.transformPoint((x, starty)) # if the point is inbetween the min, max its a point on the line if minx < x < maxx: # add it to the intersections intersections.append((x, y)) return intersections g = CurrentGlyph() startx, starty = 94, 356 endx, endy = 60, 142 results = intersectGlyphWithLine(g, ((startx, starty), (endx, endy))) print(results) Perhaps we should add this to fontTools? wow, I can't overstate my thanks for this @gferreira @typemytype @benkiel - it solves the problem, but it's also so generous and educational. I can have trouble locating the right library for this or that, I appreciate the demonstration of interoperability between them!
2025-04-01T06:40:15.289791
2024-12-07T05:59:46
2724399014
{ "authors": [ "trevorhauter", "tversteeg" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10269", "repo": "rockerBOO/awesome-neovim", "url": "https://github.com/rockerBOO/awesome-neovim/pull/1382" }
gharchive/pull-request
Add trevorhauter/gitportal.nvim Repo URL: https://github.com/trevorhauter/gitportal.nvim Checklist: [x] The plugin is specifically built for Neovim, or if it's a colorscheme, it supports treesitter syntax. [x] The lines end with a .. This is to conform to awesome-list linting and requirements. [x] The title of the pull request is Add/Update/Remove `username/repo` (notice the backticks around `username/repo`) when adding a new plugin. [x] The description doesn't mention that it's a Neovim plugin, it's obvious from the rest of the document. No mentions of the word plugin unless it's related to something else. No .. for Neovim. [x] The description doesn't contain emojis. [x] Neovim is spelled as Neovim (not nvim, NeoVim or neovim), Vim is spelled as Vim (capitalized), Lua is spelled as Lua (capitalized), Tree-sitter is spelled as Tree-sitter. [x] Acronyms should be fully capitalized, for example LSP, TS, YAML, etc. Thanks for the PR!
2025-04-01T06:40:15.293170
2023-08-22T19:28:34
1862071271
{ "authors": [ "Nichebiche", "Valdorff" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10270", "repo": "rocket-pool/RPIPs", "url": "https://github.com/rocket-pool/RPIPs/issues/66" }
gharchive/issue
Dispute text for GMC operations is not totally clear There are 3 bullets like this Anyone MAY file an RPIP disputing a grant, bounty, or retrospective award within two weeks of the announcement of recipients. Such an RPIP SHALL be subject to a snapshot vote. If a majority of the GMC agrees that a grant recipient is failing to provide the specified services to the protocol in a timely manner (as documented in the original application and in subsequent monthly updates), the GMC SHALL publicly announce such a decision and cease any future payments. This decision MAY be disputed by anyone through the creation of an RPIP within two weeks of the GMC’s announcement. The RPIP SHALL be subject to a snapshot vote. Any group or individual MAY submit a publicly-available document to the GMC claiming successful completion of the bounty. The GMC SHALL discuss all such applications. If a majority of the GMC agrees then the GMC SHALL announce the award of the bounty. Anyone MAY dispute the awarding of the bounty through the creation of an RPIP within two weeks of the GMC’s announcement. The RPIP SHALL be subject to a snapshot vote. The term "creation an RPIP", is not as clear as it should be. In discussion, https://discord.com/channels/405159462932971535/1143603646592987237/1143616540600717332, we thought it should be a merged Draft that was actively being pursued. This could be measured by having a Draft RPIP and an open sentiment poll. Niche
2025-04-01T06:40:15.295248
2020-07-11T17:44:49
655250032
{ "authors": [ "andresfcamacho", "reidmorrison" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10271", "repo": "rocketjob/iostreams", "url": "https://github.com/rocketjob/iostreams/pull/10" }
gharchive/pull-request
Enable usage of IOStreams::Pgp with keys that don't have an email. Description of changes Allow usage of keys without email. Also so allow key_id to be used for delete_keys and trust methods. By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Excellent work, impressive tests. :100: Thank you for adding support for using the key_id. Had not encountered keys before with an email address. This enhancement will help anyone that runs into this scenario. Changes are now available in gem: iostreams v1.3.0
2025-04-01T06:40:15.300396
2021-04-06T11:40:36
851339033
{ "authors": [ "andrew-newell", "yourivdlans" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10272", "repo": "rocketjob/rails_semantic_logger", "url": "https://github.com/rocketjob/rails_semantic_logger/issues/125" }
gharchive/issue
Can't disable started, processing and rendered logs when using log_level debug Environment Ruby Version: 2.6.6 Rails Version: <IP_ADDRESS> Semantic Logger Version: 4.7.4 Rails Semantic Logger Version: 4.5.0 Puma: 4.3.6 Rails configuration: config.log_level = :debug STDOUT.sync = true config.rails_semantic_logger.rendered = false config.rails_semantic_logger.add_file_appender = false config.semantic_logger.add_appender(io: STDOUT, level: config.log_level, formatter: config.rails_semantic_logger.format) Expected Behavior Using above configuration I should not see the ActionView rendered output, but I am. Actual Behavior I am still seeing the rendered logs being output, same goes for the started and processing logs (These are disabled by default as well). It seems when I set the log_level to :info the started, processing and rendered config settings do take effect. I believe this is the offending line for the "Processing" log, https://github.com/rocketjob/rails_semantic_logger/blob/47112b2c9effe7ab72f4f99d46875ed8d67d0965/lib/rails_semantic_logger/action_controller/log_subscriber.rb#L8 which has a comment: # Log as debug to hide Processing messages in production As far as the other "started" and "rendered" logs, you can set their log level using: RailsSemanticLogger::Rack::Logger.started_request_log_level = :trace RailsSemanticLogger::ActionView::LogSubscriber.rendered_log_level = :trace @andrew-newell Thanks! In the end I chose to go with another gem because this didn't feel like the best solution for our problem.
2025-04-01T06:40:15.302136
2021-06-25T16:29:41
930327857
{ "authors": [ "naveen8801", "rockingrohit9639" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10273", "repo": "rockingrohit9639/project-zone", "url": "https://github.com/rockingrohit9639/project-zone/issues/129" }
gharchive/issue
Adding dashboard data route on backend Describe feature Add a protected route on the backend for fetching and updating data for user dashboard like Profile image, social links,, badges, bio/description, and others. Add a new collection in db for uploading these details. Additional context @rockingrohit9639 @kanak22, please assign it to me, I'd like to work on it :) You can start working on it
2025-04-01T06:40:15.333729
2023-10-19T11:30:44
1951978955
{ "authors": [ "roebi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10274", "repo": "roebi/slack-to-ticket", "url": "https://github.com/roebi/slack-to-ticket/issues/9" }
gharchive/issue
use of env see https://github.com/slackapi/bolt-js/blob/main/examples/getting-started-typescript/src/utils/env.ts ups - solved in branch 7 ups - solved in branch 7
2025-04-01T06:40:15.380248
2024-05-28T18:47:51
2321709414
{ "authors": [ "HagenFritz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10275", "repo": "rogers-obrien-rad/smartpm-python-sdk", "url": "https://github.com/rogers-obrien-rad/smartpm-python-sdk/issues/15" }
gharchive/issue
[BUG] Activities: baseline starts/finishes don't match values from SmartPM UI Current Behavior I am attempting to recreate this figure: But for baseline starts and finishes, the values are way off. Expected behavior For this specific project, I am plotting 967 starts/finishes, but the SmartPM UI is only showing 464. So my thought is that we are not filtering activities that we probably should. Steps To Reproduce Steps to reproduce the behavior: Run python snippets/explore_activities.py and view the activity_distribution.csv file in the references directory. Find the same project on SmartPM and find this figure Export the data Compare Baseline Starts/Finishes Additional context Might be worthwhile talking to SmartPM users to see why or why not certain activities would not be included. The Planned Start/Finishes are also wrong, but only for the month that the data date falls within
2025-04-01T06:40:15.397111
2022-09-30T16:04:52
1392656709
{ "authors": [ "Nandinisaagar", "sakshi135" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10276", "repo": "rohansaini886/Hacktoberfest2022", "url": "https://github.com/rohansaini886/Hacktoberfest2022/pull/31" }
gharchive/pull-request
Added a C++ code in the C++ folder [Linear Search in C++] ( Issue #4) Written the code for Linear Search in C++ programming language. Resolved the issue. Please merge The PR got excluded from hacktoberfest pls merge and add the label after 1st oct if possible I want to work on this. Please assign me this.
2025-04-01T06:40:15.399471
2022-09-01T17:12:43
1359162130
{ "authors": [ "benfoster", "rohit-gohri" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10277", "repo": "rohit-gohri/redocusaurus", "url": "https://github.com/rohit-gohri/redocusaurus/issues/228" }
gharchive/issue
How to pass redoc options Redoc has a number of configurable options documented here. How do you pass these to the Redocusaurus plugin? I saw one issue that suggested to add this to the theme object but this does not work. For example: // Redocusaurus config [ 'redocusaurus', { // Plugin Options for loading OpenAPI files specs: [ { spec: 'openapi/openapi.yaml', route: '/api/', config: { disableSearch: true } }, ], // Theme Options for modifying how redoc renders them theme: { // Change with your site colors primaryColor: '#1890ff', redocOptions: { disableSearch: true, hideDownloadButton: true } } }, ], You can pass it as options - https://redocusaurus.vercel.app/docs/getting-started/theme-options
2025-04-01T06:40:15.405466
2022-04-01T16:28:48
1190029901
{ "authors": [ "sandeep-ps" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10278", "repo": "rokwire/rokwire-building-blocks-api", "url": "https://github.com/rokwire/rokwire-building-blocks-api/issues/892" }
gharchive/issue
[TASK] Release 1.13.0 Description Release version 1.13.0 and submit the following modules for deployment: Events BB API Doc Acceptance Criteria Version 1.13.0 tagged in GitHub and the above modules submitted for deployment. Completed and submitted for deployment.
2025-04-01T06:40:15.409654
2023-03-07T20:52:12
1614191205
{ "authors": [ "dreimer1986", "roleoroleo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10279", "repo": "roleoroleo/yi-hack-Allwinner-v2", "url": "https://github.com/roleoroleo/yi-hack-Allwinner-v2/issues/595" }
gharchive/issue
y211ga: Installed FW 12 and Hack 0.2.6, motion detection not working After FW <IP_ADDRESS>_202211241030 installing by official app and 0.2.6 hack reinstalled the camera does not really allow any motion detection anymore (MQTT in HA). Even more interesting, in camera settings the switch for motion detection always switches back to off after I refresh the page. Trying the same over REST link http://x.x.x.x:8080/cgi-bin/camera_settings.sh?motion_detection=yes brings back this: { "error":"false" } This happens on two cams, not just one, so I think that a problem on flashing is not our problem here... Check this commit: https://github.com/roleoroleo/yi-hack-Allwinner-v2/commit/b40b880afa217446582eeaa6664541f661ccb59c Ooops, I thought this one was in 0.2.6 already. OK, then this is fine as it's already fixed upstream. Regarding the motion detection failing. After I switched on the cam on Addin side and it was active on both sides (cam and ha) it did not work... Now a few hours later it indeed sends images again. I think it takes a while to start working maybe? Regardless, I think it's working as expected now.
2025-04-01T06:40:15.418374
2018-12-08T14:44:14
388928063
{ "authors": [ "ArturMoczulski", "Ramapitecus", "jessewgibbs" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10280", "repo": "rollbar/rollbar-php", "url": "https://github.com/rollbar/rollbar-php/issues/426" }
gharchive/issue
Typo in EncodedPayload->decreaseSize https://github.com/rollbar/rollbar-php/blob/c2f8be87983ecdd3c46ceb3c459de02376af4707/src/Payload/EncodedPayload.php#L26 $this->size =- $amount @ramapitecus thanks for bringing this to our attention (as well as #427 and #428). @ArturMoczulski can you take a look when you get a chance? I'm not sure I understand what the problem is here I'm sorry I wasn't clear. The method is called decreaseSize so I assume that if $this->size is 10, and you call $obj->decreaseSize(2) you'ld want $this->size === 8 but there is a typo there, the =- is flipped. And $this->size === -2 after that call. Fixed in master. Will be included in the next release. Released in v1.7.3
2025-04-01T06:40:15.422600
2023-03-07T12:15:31
1613325387
{ "authors": [ "jfbaraky", "rnovacek" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10281", "repo": "rollbar/rollbar-react", "url": "https://github.com/rollbar/rollbar-react/issues/87" }
gharchive/issue
Incorrect typescript type on fallbackUI prop on ErrorBoundary The fallbackUI prop of ErrorBoundary has incorrect type in index.d.ts: fallbackUI?: ReactNode; But the docs says it should be a react component. If I pass a react component: const FatalError = () => { return 'Error'; } <ErrorBoundary fallbackUI={FatalError}> ... </ErrorBoundary> typescript will complain with No overload matches this call. Overload 1 of 2, '(props: ErrorBoundaryProps | Readonly<ErrorBoundaryProps>): ErrorBoundary', gave the following error. Type 'FC<{}>' is not assignable to type 'ReactNode'. Overload 2 of 2, '(props: ErrorBoundaryProps, context: any): ErrorBoundary', gave the following error. Type 'FC<{}>' is not assignable to type 'ReactNode'. The doc is passing the component correctly. If you try to satisfy the typescript, you'll get this error: Unhandled Runtime Error Error: Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: object. Check the render method of `ErrorBoundary`. What I did was hard cast the type on my side: <ErrorBoundary errorMessage="Error in React render" fallbackUI={Error as unknown as React.ReactNode} // Prop is wrongly typed extra={(_error, _info) => ({ // Extra information that need to be sent to rollbar })} > {children} </ErrorBoundary>
2025-04-01T06:40:15.432846
2020-12-08T09:22:28
759274182
{ "authors": [ "jmcvetta" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10282", "repo": "rollbar/terraform-provider-rollbar", "url": "https://github.com/rollbar/terraform-provider-rollbar/issues/154" }
gharchive/issue
TestAccTeamDeleteOnAPIBeforeApply no longer passes The code now contained in TestAccTeamDeleteOnAPIBeforeApply used to pass reliably. Now, however, it does not pass. Why? Disabling this test caused a decrease in coverage: https://coveralls.io/builds/35531461 The original purpose of this test was indeed to cover handleErrNotFound(). And it used to work. Again bringing us to the puzzling question: why does it no longer work? During testing I have uncovered similar bugs when reading a deleted project or a deleted project access token. Terraform should recognize that the resource has been deleted from the server and re-create it. Instead, Terraform fails. Maybe, before recent changes, Rollbar API allowed "deleted" teams to be read if requested by ID, even tho the "deleted" teams were not included in the list of teams?
2025-04-01T06:40:15.439109
2024-03-21T19:44:07
2201026059
{ "authors": [ "Reecepbcups" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10283", "repo": "rollchains/spawn", "url": "https://github.com/rollchains/spawn/issues/112" }
gharchive/issue
ICS consumer chain generation TODO [x] Add / remove ICS as expected with tags [x] Add new 'staking' option. This feature is on by default but hidden If ICS is used, then staking is disabled, removes all references for staking. \ Also remove Distr, Mint, etc. [x] Integrate Interchaintest w/ simple ICS test ( https://github.com/strangelove-ventures/interchaintest/pull/1088 ) [x] Add test_ics_node.sh instead of test_node.sh w/ Stide's provider hack [x] CI: Run test_ics_node.sh for ics chain (matrix?) [x] Speed up CI [ ] Unit Test (ref: stride, neutron) [ ] rename to ics_basic_test.go - see if we can abstract in setup.go so we can reuse standard test instead of new ones [ ] Update to ICTest v8.3.0 + IBC validation (from upstream ICT) [ ] Local Interchain proper generation (remove default) [ ] In UI, if ICS selected, deselect POA Test Basic make install && rm -rf testics1/ spawn new testics1 --disable=poa,tf,globalfee,packetforward,ibc-ratelimit,cw,wasm-lc,ignite-cli spawn new chain1 --disable=ics spawn new chain2 [ ] merge different ChainSepc array setups into 1. This way it can be used in all the test. Then just have the provider setup automatically. We just switch between Ethos, CosmosHub, and Saga (only allow 1 for an ICS chain, Hub default) [ ] Local-Interchain have ICS be renamed to the normal file start testnet (For JSON generation, maybe have a struct instead of the raw JSON files. (we only bech32 convert wasm1 & wasmd info, so other networks like cosmoshub, neutron, ethos, are safe to support directly in the file) [ ] Local-IC delete encoding-options which are not used (probably part of the above JSON builder) TODO: push up ICS 4.1.0 with heighliner to confirm make template-ics and make template-staking's make ictest-basic function as expected
2025-04-01T06:40:15.445379
2023-09-13T14:51:15
1894705454
{ "authors": [ "gupadhyaya", "nashqueue" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10284", "repo": "rollkit/rollkit", "url": "https://github.com/rollkit/rollkit/pull/1190" }
gharchive/pull-request
updated README.md with releases Overview Checklist [ ] New and updated code has appropriate documentation [ ] New and updated code has new and/or updated testing [ ] Required CI checks are passing [ ] Visual proof for any user facing features like CLI or documentation updates [ ] Linked issues closed with keywords We are not compatible with mocha-4 / arabica 10 until we do a new release with #1166 Markdown lint still breaks as we don't have a rc13 release yet
2025-04-01T06:40:15.488129
2016-04-21T23:44:10
150221042
{ "authors": [ "cgilmour", "jbrendel" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10285", "repo": "romana/romana", "url": "https://github.com/romana/romana/pull/99" }
gharchive/pull-request
More installer changes A handful of changes to support installing Romana with different CIDRs. Adds a docker_cidr variable to specify the IP and subnet that docker0 interface will bind to. Makes sure the docker configs are in place before it's installed, because it gets started automatically. (Applying config after means it needs a restart) Make sure we calculate the host bits based on the romana_cidr (16 - subnet prefix length) and put that in the config. Sort the hosts (for AWS) so the inventory is consistent on multiple runs of the installer. They're discovered dynamically, so the values would be in an arbitrary order. Sorting by private IP seemed reasonable. A command-line that uses allthethings is like this: ./romana-setup -n newcidrs -s kubernetes install -e romana_cidr=<IP_ADDRESS>/12 -e host_cidr=<IP_ADDRESS>/24 -e docker_cidr=<IP_ADDRESS>/24 Special note for AWS: The CIDR prefix needs to be at least 16. It really doesn't like having a 10/8. Is the docker_cidr even used? If I recall correctly, our docker0 interface doesn't even carry an address? Could be wrong, I don't currently have a cluster to look at. It's not used in a useful way, but it is there by default after installation. Most of the time, nobody cares because 172.16/12 is an uncommon CIDR to use. But when we want to use it, this variable helps us put docker in a different spot. As long as this is documented in that way somewhere... Added documentation (in PR #100) so that we can get this closed.
2025-04-01T06:40:15.490338
2015-02-24T11:31:04
58726047
{ "authors": [ "janareit", "romanbican" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10286", "repo": "romanbican/roles", "url": "https://github.com/romanbican/roles/issues/12" }
gharchive/issue
Initiating roles and permissions Hey, on L5 where would you put this initiating roles and permissions file? Do you create some database/rbac/rbac_init.php file for it? I understand that later for new users I would integrate this to users management but for initial start do you create them in some init file? Hello. Well, sorry, but I don't understand you. When you are starting, create you first roles and permissions through database seeds and attach permissions for example to a administrator role. Then create administration panel, where you can attach (or detach of course) roles and permissions to a new users. Ok, forgot about db seeds. Thanks for hint!
2025-04-01T06:40:15.513588
2020-05-21T15:21:56
622575491
{ "authors": [ "bitpshr", "sebmck" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10287", "repo": "romejs/rome", "url": "https://github.com/romejs/rome/pull/502" }
gharchive/pull-request
feat: add react/react-jsx-no-duplicate-props This pull request implements the react/jsx-no-duplicate-props lint rule, disallowing duplicate props on JSX elements. References #341 Supersedes #351 Coauthored by @zackargyle Looks great, thank you!
2025-04-01T06:40:15.538501
2021-10-13T20:26:23
1025666100
{ "authors": [ "ronnypolley" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10288", "repo": "ronnypolley/posh-sdkman", "url": "https://github.com/ronnypolley/posh-sdkman/issues/38" }
gharchive/issue
How do I change the default path of posh-gvm? (https://github.com/flofreud/posh-gvm/issues/16) Copy of https://github.com/flofreud/posh-gvm/issues/16 I found that posh-gvm default set its path as C:\Users\user.posh_gvm. However, in my case, the space of my C: is not very large. So, can I set the default path of posh-gvm to other path ? such as D:\posh_gvm (in my case). Thanks a lot. Copy of https://github.com/flofreud/posh-gvm/issues/16#issuecomment-178069660 You need to declare $Global:PGVM_DIR before loading posh-gvm.
2025-04-01T06:40:15.541209
2024-09-24T07:07:49
2544538362
{ "authors": [ "goldgeer", "ronvanderheijden", "stof" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10289", "repo": "ronvanderheijden/openid-connect", "url": "https://github.com/ronvanderheijden/openid-connect/issues/30" }
gharchive/issue
ERROR syntax error, unexpected ')', expecting variable (T_VARIABLE) ERROR syntax error, unexpected ')', expecting variable (T_VARIABLE) in vendor/ronvanderheijden/openid-connect/src/IdTokenResponse.php on line 39. There is a comma symbol beyond the last parameter in constructor . public function __construct( IdentityRepositoryInterface $identityRepository, ClaimExtractor $claimExtractor, Configuration $config, CurrentRequestServiceInterface $currentRequestService = null, $encryptionKey = null , <============================== comma ) You are using an outdated PHP version. You should upgrade to a newer version (preferable 8.3). this repo should not tell composer it supports PHP 7.4 if it uses a syntax introduced in PHP 8.0 though. This is true, but still, PHP 7.4 is not actively supported for almost 3 years.
2025-04-01T06:40:15.649400
2022-10-20T22:03:36
1417362893
{ "authors": [ "2byrds", "HunnySajid" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10311", "repo": "roots-id/roots-react-native", "url": "https://github.com/roots-id/roots-react-native/pull/15" }
gharchive/pull-request
Hunain/issue 12/settings Related Task # https://github.com/roots-id/roots-react-native/issues/12 ChangeLog: Added settings screen Added developer screen link with dummy on settings Added save screen with mock data Implemented Picker Component to look better on IOS , previously it was not working with current styling IOS https://user-images.githubusercontent.com/19255438/197067056-0619cee2-609f-42c2-bc42-c0e9881f72a7.mov Android https://user-images.githubusercontent.com/19255438/197067392-d1cdbffe-18a1-47d1-b3e4-c4111144818c.mov I had some trouble with this one @HunnySajid. Web loaded fine. But android didn't load at all and iOS i wasn't able to click around. I'll try to look into it a bit more tomorrow. I had some trouble with this one @HunnySajid. Web loaded fine. But android didn't load at all and iOS i wasn't able to click around. I'll try to look into it a bit more tomorrow. @2byrds for ios its behavior is different. It provides a swapping up/down instead of showing a dropdown. Can you provide some details about android phone you are using so that I can look into it if it is os/mobile specific issue? ah yes @HunnySajid I was able to swipe the picker on ios. thx :) I think my android emulator was in a bad state. i started a fresh one and it worked great.
2025-04-01T06:40:15.663625
2020-07-17T18:50:33
659527591
{ "authors": [ "adevoss", "frennkie", "rootzoll" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10312", "repo": "rootzoll/raspiblitz", "url": "https://github.com/rootzoll/raspiblitz/issues/1354" }
gharchive/issue
Feature request: Confirm Reboot/PowerOff When I wanted to reboot my Pi, in the menu I choose 'Reboot' and it just rebooted. Exactly what it should do of course. What about a confirmation: Are you sure? A mistake is easily made. Same with 'Off' probably. Did not try that. Feel free to give this a try - sounds like a good "first issue" :-D My first issues were #1281 and #1297 but I will give it a try this weekend. Good third issue. I don't have a screen but the Update/Patch menu has a lot of confirmation questions. So I'll take a look at that. [ ] merge and adapt PR tested and is working. closing issue.
2025-04-01T06:40:15.665053
2021-09-06T20:15:52
989418103
{ "authors": [ "rootzoll" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10313", "repo": "rootzoll/raspiblitz", "url": "https://github.com/rootzoll/raspiblitz/issues/2517" }
gharchive/issue
Test & Fix CLN-Rescue Export & Import Tested the CLN-Import from an older/veryearly c-lightning node ... the test failed. To be retested. Test worked with RC3 - closing issue.
2025-04-01T06:40:15.667295
2022-03-29T14:13:30
1184939356
{ "authors": [ "kilrau", "openoms", "rootzoll" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10314", "repo": "rootzoll/raspiblitz", "url": "https://github.com/rootzoll/raspiblitz/issues/3043" }
gharchive/issue
[CL] CLBOSS needs a warning I was close to a heart attack today when I installed CLBOSS via the Blitz menu and suddenly funds were seemingly moving out of my CL node. Just CLBOSS doing its thing, but this needs a proper warning that CLBOSS is taking control over funds and will take action right after installation. @openoms is this addressed or should I push to v1.8.1? not yet, but will add a notice to the menu also @openoms Can I close for v1.8.0 RC2 release? @openoms please signal when we can close this one - or we push to v1.8.1
2025-04-01T06:40:15.669030
2023-03-02T11:01:28
1606587321
{ "authors": [ "openoms", "rootzoll" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10315", "repo": "rootzoll/raspiblitz", "url": "https://github.com/rootzoll/raspiblitz/pull/3697" }
gharchive/pull-request
btcpay update v1.8.2, postgres database fix btcpay update and database creation modified to follow: https://docs.btcpayserver.org/BTCPayServer/db-migration/ migration is not done yet automatically for old sqlite instances - will test migrating an old store first cc @NicolasDorier covers #3701 @openoms plz merge if ready
2025-04-01T06:40:15.797465
2024-02-21T21:17:47
2147728676
{ "authors": [ "christophfroehlich", "stefanscherzinger", "traversaro" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10317", "repo": "ros-controls/control.ros.org", "url": "https://github.com/ros-controls/control.ros.org/pull/250" }
gharchive/pull-request
Add MuJoCo by FZI ros2_control system interface for MuJoCo. Not tested but looks great! @mergifyio backport humble iron @christophfroehlich Thanks for spotting and referencing that implementation! However, I'm afraid that's a little robot dependent at the moment. As far as I'm concerned, MuJoCo needs a specific configuration .xml that specifies the robot to simulate. We could adapt/change it to provide a more generic interface between MuJoCo and ros2_control so that users can simulate arbitrary robots (manipulators). I didn't continue this idea lately, but I'm interested in supporting/implementing if that's of use for the community. @christophfroehlich Thanks for spotting and referencing that implementation! However, I'm afraid that's a little robot dependent at the moment. As far as I'm concerned, MuJoCo needs a specific configuration .xml that specifies the robot to simulate. We could adapt/change it to provide a more generic interface between MuJoCo and ros2_control so that users can simulate arbitrary robots (manipulators). I didn't continue this idea lately, but I'm interested in supporting/implementing if that's of use for the community. For the scope of this PR it is fine like it is. It should just give a reference implementation for users on control.ros.org I have a personal interest in bringing MuJoCo and ros2_control closer due to a research project I'm working on. But I haven't tried it out yet and can't comment on the limitations. @traversaro pointed me to https://github.com/google-deepmind/mujoco/issues/1432, so I think there is more work to be done to provide a seamless integration like with Gazebo (classic). Once I have the time to work on that, I'd be happy to discuss future steps to find a more general solution. @traversaro pointed me to google-deepmind/mujoco#1432, so I think there is more work to be done to provide a seamless integration like with Gazebo (classic). Once I have the time to work on that, I'd be happy to discuss future steps to find a more general solution. Just to complement the link, I think most of people using URDF with MuJoCo either manually or automatically change the URDF before the loading in MuJoCo, or the resulting .mjcf after the conversion, and that is why there are not a lot of discussion like https://github.com/google-deepmind/mujoco/issues/1432 in MuJoCo issue tracker.
2025-04-01T06:40:15.800573
2023-02-20T12:17:47
1591752438
{ "authors": [ "ahcorde", "tonynajjar" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10318", "repo": "ros-controls/gazebo_ros2_control", "url": "https://github.com/ros-controls/gazebo_ros2_control/pull/179" }
gharchive/pull-request
Remove publish_rate parameter It was removed https://github.com/ros-controls/ros2_controllers/pull/468 CI is failing, the service_msgs package was added 3 weeks ago https://github.com/ros2/rcl_interfaces/commits/rolling/service_msgs and it was released last week, maybe a package that uses this dependency is not released yet and that's why is failing? just to clarify, are you asking me the question? :smile: @bmagyar or @destogl can you relaunch the Github actions ? just to clarify, because that's unrelated to my PR, are you asking me the question? smile This was not a question, it was a clarification, but for some reason I added ?. Let's relaunch CI
2025-04-01T06:40:15.830096
2024-11-21T08:29:51
2678509647
{ "authors": [ "marip8", "r7vme" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10319", "repo": "ros-industrial/reach_ros2", "url": "https://github.com/ros-industrial/reach_ros2/pull/38" }
gharchive/pull-request
Fixes for 24.04 build Related to https://github.com/ros-industrial/reach/pull/83 Fix MPI_C not found error similar to https://github.com/ros-industrial/reach/pull/82 Use tf2_eigen.hpp instead of tf2_eigen.h https://github.com/ros2/geometry2/pull/645 Closing; addressed in #35 and #39; thanks for bringing this up
2025-04-01T06:40:15.864796
2022-10-04T08:23:25
1395859136
{ "authors": [ "enriLoniterp", "felipebapa", "jwson97", "okvik", "sjahr" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10320", "repo": "ros-planning/moveit2_tutorials", "url": "https://github.com/ros-planning/moveit2_tutorials/issues/525" }
gharchive/issue
Not understand MoveItGroupInterface Description I've downloaded Universalrobots driver and followed the procedure to use moveIt!2, it actually works what showed in the README and i use "ur_manipulator" in Rviz to move my arm. But my interest is to use moveit2_tutorials to send pose and design a pick and place application with MoveItGriuoInterface. What i have designed is: #include <memory> #include <rclcpp/rclcpp.hpp> #include <moveit/move_group_interface/move_group_interface.h> int main(int argc, char * argv[]) { rclcpp::init(argc, argv); auto const node = std::make_shared<rclcpp::Node>( "hello_moveit", rclcpp::NodeOptions().automatically_declare_parameters_from_overrides(true) ); auto const logger = rclcpp::get_logger("ur_manipulator"); // Create the MoveIt MoveGroup Interface using moveit::planning_interface::MoveGroupInterface; auto move_group_interface = MoveGroupInterface(node, "ur_manipulator"); printf("picvkandplace"); // Set a target Pose auto const target_pose = []{ geometry_msgs::msg::Pose msg; msg.orientation.w = 1.0; msg.position.x = 0.28; msg.position.y = -0.2; msg.position.z = 0.5; return msg; }(); move_group_interface.setPoseTarget(target_pose); // Create a plan to that target pose auto const [success, plan] = [&move_group_interface]{ moveit::planning_interface::MoveGroupInterface::Plan msg; auto const ok = static_cast<bool>(move_group_interface.plan(msg)); return std::make_pair(ok, msg); }(); // Execute the plan if(success) { move_group_interface.execute(plan); } else { RCLCPP_ERROR(logger, "Planing failed!"); } // Shutdown ROS rclcpp::shutdown(); return 0; } i launch ros2 run hello_moveit hello_moveit as in the tutorial but i get: [ERROR] [1664871447.339399807] [hello_moveit]: Could not find parameter robot_description_semantic and did not receive robot_description_semantic via std_msgs::msg::String subscription within 10.000000 seconds. Error: Could not parse the SRDF XML File. Error=XML_ERROR_EMPTY_DOCUMENT ErrorID=13 (0xd) Line number=0 at line 715 in ./src/model.cpp [ERROR] [1664871447.342434566] [moveit_rdf_loader.rdf_loader]: Unable to parse SRDF [FATAL] [1664871447.342740966] [move_group_interface]: Unable to construct robot model. Please make sure all needed information is on the parameter server. terminate called after throwing an instance of 'std::runtime_error' what(): Unable to construct robot model. Please make sure all needed information is on the parameter server. [ros2run]: Aborted I can't understand what should i look for also because in that everything is set to use moveit, robot_description_semantic included. Hope someone could help to this endless problem! Enrico I have same problem. Who can solve this problem? @enriLoniterp Did you find an answer? I am developing a similar program. Thanks for your help! You are likely missing the step of launching the move_group node for your robot package before running the program. For example, you can try ros2 launch moveit_resources_panda_moveit_config demo.launch.py. If you are using a different robot there likely alredy exists a package and a launch file that does the same, otherwise you're gonna need to either 1) use MoveIt Setup Assistant to generate an appropriate moveit_config package, or 2) write the launch file by hand. All of this is more or less covered in the MoveIt docs. The problem is that your program (actually the MoveGroupInterface constructor) cannot find the SRDF robot description MoveGroupInterface in either of the two places it looks in: 1) in the robot_description_semantic parameter in your node's parameters list, or 2) in a published /robot_description_semantic ROS topic. It follows that you can solve this in two ways: either 1) feed the robot_description_semantic parameter to your program node when launching it, or 2) configure the move_group node to publish its own robot_description_semantic parameter to a matching topic from which other nodes in the system can read it. Personally I prefer the second method. To do this, find the launch file that is creating the move_group node and set its node parameter publish_robot_description_semantic = True . I suggest you set the publish_robot_description as well. To verify the SRDF is published correctly run ros2 topic echo /robot_description_semantic Which should spit out the .srdf file. Thanks for your fast answer!. I am going to check. Looks like the discussion is resolved. Feel free to re-open if you'd like to continue it
2025-04-01T06:40:15.867877
2022-07-14T22:08:25
1305337441
{ "authors": [ "henrygerardmoore", "vatanaksoytezer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10321", "repo": "ros-planning/moveit2_tutorials", "url": "https://github.com/ros-planning/moveit2_tutorials/pull/472" }
gharchive/pull-request
renamed cartesian limits file Description Renamed cartesian_limits.yaml to avoid warning when running dual arm demo Checklist [ ] Required by CI: Code is auto formatted using clang-format [ ] While waiting for someone to review your request, please consider reviewing another open pull request to support the maintainers Thank you @henrygerardmoore !
2025-04-01T06:40:15.871325
2021-06-29T22:26:27
933150880
{ "authors": [ "MarqRazz", "vatanaksoytezer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10322", "repo": "ros-planning/moveit2_tutorials", "url": "https://github.com/ros-planning/moveit2_tutorials/pull/93" }
gharchive/pull-request
Clean up move_group and moveit_cpp tutorials Description Added panda_hand_controller to the loaded controllers in launch files switched joint_state_controller to joint_state_broadcaster updated visual_tools text to snake case for better readability added collision object example to moveit_cpp tutorial (I am planning on making a PR to remove the demo code from moveit2) Checklist [ ] Required by CI: Code is auto formatted using clang-format [ ] While waiting for someone to review your request, please consider reviewing another open pull request to support the maintainers Turns out I cannot run because of joint_state_broadcaster is not registered in moveit_resources.
2025-04-01T06:40:15.888067
2016-03-04T20:06:45
138573467
{ "authors": [ "tfoote" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10323", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/10619" }
gharchive/pull-request
geometry_experimental: 0.5.13-0 in 'jade/distribution.yaml' [bloom] Increasing version of package(s) in repository geometry_experimental to 0.5.13-0: upstream repository: https://github.com/ros/geometry_experimental.git release repository: https://github.com/ros-gbp/geometry_experimental-release.git distro file: jade/distribution.yaml bloom version: 0.5.20 previous version for package: 0.5.12-0 geometry_experimental * Remove LGPL from license tags LGPL was erroneously included in 2a38724. As there are no files with it in the package. * add missing dependencies in the meta-package geometry_experimental This partly fixes the doc jobs in #120 <https://github.com/ros/geometry_experimental/issues/120> * Contributors: Jochen Sprickerhof, Vincent Rabaud tf2 No changes tf2_bullet * Don't export catkin includes They only point to the temporary include in the build directory. * Contributors: Jochen Sprickerhof tf2_eigen * Added missing inline * Added unit test - Testing conversion to msg forward/backward * Added eigenTotransform function * Contributors: Davide Tateo, boris-il-forte tf2_geometry_msgs * Add missing python_orocos_kdl dependency * make example into unit test * vector3 not affected by translation * Contributors: Daniel Claes, chapulina tf2_kdl * converting python test script into unit test * Don't export catkin includes * Contributors: Jochen Sprickerhof, Tully Foote tf2_msgs No changes tf2_py No changes tf2_ros * fix documentation warnings * Adding tests to package * Contributors: Laurent GEORGE, Vincent Rabaud tf2_sensor_msgs * add missing Python runtime dependency * fix wrong comment * Adding tests to package * Fixing do_transform_cloud for python The previous code was not used at all (it was a mistake in the __init__.py so the do_transform_cloud was not available to the python users). The python code need some little correction (e.g there is no method named read_cloud but it's read_points for instance, and as we are in python we can't use the same trick as in c++ when we got an immutable) * Contributors: Laurent GEORGE, Vincent Rabaud tf2_tools * casted el to string in view_frames * Contributors: g_gemignani hold for coordinated release
2025-04-01T06:40:15.891753
2018-11-19T16:02:46
382282305
{ "authors": [ "JWhitleyAStuff", "tfoote" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10324", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/19495" }
gharchive/pull-request
pacmod3: 1.2.0-0 in 'lunar/distribution.yaml' [bloom] Increasing version of package(s) in repository pacmod3 to 1.2.0-0: upstream repository: https://github.com/astuff/pacmod3.git release repository: https://github.com/astuff/pacmod3-release.git distro file: lunar/distribution.yaml bloom version: 0.6.9 previous version for package: 1.1.1-0 pacmod3 * Merge pull request #41 <https://github.com/astuff/pacmod3/issues/41> from astuff/fix/vector_comp_dbc_file Removing unnecessary line preventing Vector CANdb++ editor from opening * Merge pull request #39 <https://github.com/astuff/pacmod3/issues/39> from astuff/feat/add_comp_rpt Feature add component report for each PACMod component. * DBC: Fixing errors. CI: Adding DBC validation to workflow. * DBC: Typo on two lines. * Adding encoding for CLEAR_FAULTS flag. * Adding parsing and publishing for COMPONENT_RPT. * DBC: Add COMPONENT_RPT and CLEAR_FAULTS flag. * Merge pull request #38 <https://github.com/astuff/pacmod3/issues/38> from astuff/feature/add_veh_6 Adding VEHICLE_6. * Contributors: Daniel-Stanek, Joshua Whitley, Mike Lemm, Nate Imig Holding for sync
2025-04-01T06:40:15.895749
2020-11-26T06:30:44
751323948
{ "authors": [ "hakuturu583" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10325", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/27499" }
gharchive/pull-request
add python3-xmltodict-pip package Signed-off-by: Masaya Kataoka<EMAIL_ADDRESS>xmltodict Packages in rosdistro is Python2 version. https://packages.ubuntu.com/search?keywords=python-xmltodict&searchon=names&suite=groovy&section=all We want to use it with Python3, so we want to add python3-xmltodict-pip package to rosdistro. Sorry, I cannot understand why CI was failed. Does anyone has idea about it? changed to native packages. Ubuntu : https://packages.ubuntu.com/search?keywords=python3-xmltodict Arch : https://www.archlinux.org/packages/community/any/python-xmltodict/ Alpine:https://pkgs.alpinelinux.org/package/edge/community/x86/py3-xmltodict Debian:https://packages.debian.org/search?keywords=python3-xmltodict
2025-04-01T06:40:15.897916
2023-06-05T15:14:57
1742001258
{ "authors": [ "clalancette", "vooon" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10326", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/37469" }
gharchive/pull-request
Revert "mavlink: 2023.5.5-1 in 'melodic/distribution.yaml' [bloom] (#… …37150)" This reverts commit 18c62e19dfc6dd53258cc14c76903fc00cd6ce22. This has failed to build since it was merged: https://build.ros.org/job/Mbin_uB64__mavlink__ubuntu_bionic_amd64__binary/ . FYI @vooon @clalancette i've updated patches and can confirm build for Noetic and Iron. But i'm unsure if that's good idea to release for Melodic, as i EOL and i haven't tested for Python 2. So tomorrow i'm going to release for Noetic, Humble, Iron and Rolling. Is i'm right?
2025-04-01T06:40:15.901074
2024-11-24T20:32:27
2688342614
{ "authors": [ "clalancette", "malban" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10327", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/43645" }
gharchive/pull-request
log_view: 0.2.4-1 in 'iron/distribution.yaml' [bloom] Increasing version of package(s) in repository log_view to 0.2.4-1: upstream repository: https://github.com/hatchbed/log_view.git release repository: https://github.com/ros2-gbp/log_view-release.git distro file: iron/distribution.yaml bloom version: 0.12.0 previous version for package: null log_view * Fix build error caused by mvwprintw. (#19 <https://github.com/hatchbed/log_view/issues/19>) * Contributors: Marc Alban Given that this has never been released into Iron before, I'm going to consider this a new package and close this out.
2025-04-01T06:40:15.901898
2015-04-03T01:12:19
66045365
{ "authors": [ "dmiklic" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10328", "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/7758" }
gharchive/pull-request
Updated rosdep rules for the libaria package Updated rosdep rules to use the binary libaria-dev package, where available. Fixed.
2025-04-01T06:40:15.910727
2018-06-19T13:56:10
333683451
{ "authors": [ "nuclearsandwich", "paulbovbel" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10329", "repo": "ros2/poco_vendor", "url": "https://github.com/ros2/poco_vendor/pull/18" }
gharchive/pull-request
Add dependency on system poco. If there is a suitable system version of poco, prefer to use it rather than building poco locally. For bouncy the necessary version of poco is available from the bionic repositories. Necessary to bloom poco_vendor without patching. Thanks for your review @mjcarroll I realized just after when checking this package.xml against the one from ardent that some additional dependencies are needed in the case that poco upstream is not available. @mjcarroll @dirk-thomas ready for a second round of review with the poco build dependencies removed. Apologies I'm not aware of the background for this PR - on xenial, libpoco-dev does not pull in PCRE or zlib. As someone who builds ROS2 from source in CI on xenial and bionic, and depends on catkin_pkg/rosdep to resolve dependencies, I wonder: is there any harm in keeping those build_depends in poco_vendor, given that xenial is nominally a supported platform?
2025-04-01T06:40:15.944209
2015-12-25T10:44:24
123872783
{ "authors": [ "adamatan", "iamsebastian", "kweng2", "rosario", "stephentu" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10330", "repo": "rosario/kasper", "url": "https://github.com/rosario/kasper/issues/24" }
gharchive/issue
Jekyll build error, Wrong Argument Type, Expected String Configuration file: /Users/ziluweng/Documents/gitHub/kasper/_config.yml Deprecation: You appear to have pagination turned on, but you haven't included the jekyll-paginate gem. Ensure you have gems: [jekyll-paginate] in your configuration file. Source: /Users/ziluweng/Documents/gitHub/kasper Destination: /Users/ziluweng/Documents/gitHub/kasper/_site Incremental build: disabled. Enable with --incremental Generating... Deprecation: Collection#map should be called on the #docs array directly. Called by /Users/ziluweng/Documents/gitHub/kasper/_plugins/rssgenerator.rb:46:in block in generate'. Deprecation: Collection#count should be called on the #docs array directly. Called by /Users/ziluweng/Documents/gitHub/kasper/_plugins/rssgenerator.rb:49:in rescue in block in generate'. Deprecation: Collection#reverse should be called on the #docs array directly. Called by /Users/ziluweng/Documents/gitHub/kasper/_plugins/rssgenerator.rb:51:in block in generate'. Deprecation: Document#title is now a key in the #data hash. Called by /Users/ziluweng/Documents/gitHub/kasper/_plugins/rssgenerator.rb:53:in block (3 levels) in generate'. Deprecation: Document#excerpt is now a key in the #data hash. Called by /Users/ziluweng/Documents/gitHub/kasper/_plugins/rssgenerator.rb:55:in `block (3 levels) in generate'. jekyll 3.0.1 | Error: wrong argument type Jekyll::Excerpt (expected String) Zilus-MacBook-Pro:kasper ziluweng$ gem I cloned the repo and ran jekyll serve, and ran into this issue that I'm stuck on. I solved the paginate problem by adding gems: jekyll-paginate But still gives the last error Hey, It could be related to a newer version of Ruby/Jekyll you might be using. Patches welcome ;) Facing the same problem, with a fresh clone of Kasper. Facing the same problem, with a fresh clone of Kasper. With a simple jekyll-server: $ jekyll serve Configuration file: /private/tmp/kasper/_config.yml Source: /private/tmp/kasper Destination: /private/tmp/kasper/_site Incremental build: disabled. Enable with --incremental Generating... Deprecation: Collection#map should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:46:in `block in generate'. Deprecation: Collection#count should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:49:in `rescue in block in generate'. Deprecation: Collection#reverse should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:51:in `block in generate'. Deprecation: Document#title is now a key in the #data hash. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:53:in `block (3 levels) in generate'. Deprecation: Document#excerpt is now a key in the #data hash. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:55:in `block (3 levels) in generate'. jekyll 3.0.1 | Error: wrong argument type Jekyll::Excerpt (expected String) I've tried changing line 55 from: item.description = parser.convert(post.excerpt) To: item.description = 'Whatever' #parser.convert(post.excerpt) And then I get: $ jekyll serve Configuration file: /private/tmp/kasper/_config.yml Source: /private/tmp/kasper Destination: /private/tmp/kasper/_site Incremental build: disabled. Enable with --incremental Generating... Deprecation: Collection#map should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:46:in `block in generate'. Deprecation: Collection#count should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:49:in `rescue in block in generate'. Deprecation: Collection#reverse should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:51:in `block in generate'. Deprecation: Document#title is now a key in the #data hash. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:53:in `block (3 levels) in generate'. Dependency Error: Yikes! It looks like you don't have pygments or one of its dependencies installed. In order to use Jekyll as currently configured, you'll need to install this gem. The full error message from Ruby is: 'cannot load such file -- pygments' If you run into trouble, you can find helpful resources at http://jekyllrb.com/help/! Liquid Exception: pygments in /private/tmp/kasper/_posts/2013-11-10-welcome-to-jekyll.markdown ERROR: YOUR SITE COULD NOT BE BUILT: ------------------------------------ pygments I've tried both gem install jekyll-pygments and gem install pygments, with no luck. With a simple jekyll-server: $ jekyll serve Configuration file: /private/tmp/kasper/_config.yml Source: /private/tmp/kasper Destination: /private/tmp/kasper/_site Incremental build: disabled. Enable with --incremental Generating... Deprecation: Collection#map should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:46:in `block in generate'. Deprecation: Collection#count should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:49:in `rescue in block in generate'. Deprecation: Collection#reverse should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:51:in `block in generate'. Deprecation: Document#title is now a key in the #data hash. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:53:in `block (3 levels) in generate'. Deprecation: Document#excerpt is now a key in the #data hash. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:55:in `block (3 levels) in generate'. jekyll 3.0.1 | Error: wrong argument type Jekyll::Excerpt (expected String) I've tried changing line 55 from: item.description = parser.convert(post.excerpt) To: item.description = 'Whatever' #parser.convert(post.excerpt) And then I get: $ jekyll serve Configuration file: /private/tmp/kasper/_config.yml Source: /private/tmp/kasper Destination: /private/tmp/kasper/_site Incremental build: disabled. Enable with --incremental Generating... Deprecation: Collection#map should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:46:in `block in generate'. Deprecation: Collection#count should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:49:in `rescue in block in generate'. Deprecation: Collection#reverse should be called on the #docs array directly. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:51:in `block in generate'. Deprecation: Document#title is now a key in the #data hash. Called by /private/tmp/kasper/_plugins/rssgenerator.rb:53:in `block (3 levels) in generate'. Dependency Error: Yikes! It looks like you don't have pygments or one of its dependencies installed. In order to use Jekyll as currently configured, you'll need to install this gem. The full error message from Ruby is: 'cannot load such file -- pygments' If you run into trouble, you can find helpful resources at http://jekyllrb.com/help/! Liquid Exception: pygments in /private/tmp/kasper/_posts/2013-11-10-welcome-to-jekyll.markdown ERROR: YOUR SITE COULD NOT BE BUILT: ------------------------------------ pygments I've tried both gem install jekyll-pygments and gem install pygments, with no luck. I managed to solve this error by removing _plugins/rssgenerator.rb and using the default feed.xml which comes by default when you do a jekyll new ... Simply run: "jekyll s --safe". This should ignore broken dependencies. The error was indeed coming from _plugins/rssgenerator.rb. Looking at the author's comment: https://github.com/agelber/jekyll-rss the new Jekyll comes with a feed.xml and they suggest to use the default one. I've included the default feed.xml which comes with the standard jekyll new template and removed the old rssgenerator.rb.
2025-04-01T06:40:15.971995
2017-07-24T18:48:38
245173970
{ "authors": [ "superjax" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10331", "repo": "rosflight/firmware", "url": "https://github.com/rosflight/firmware/pull/221" }
gharchive/pull-request
Fix 202 Here are a few fixes from flight testing on Friday, and fixing unit test issues. I don't know why it is failing on Travis. It works on my machine. I'll work on that later. This should be ready to go!
2025-04-01T06:40:15.982863
2021-08-20T13:50:00
975629445
{ "authors": [ "schmerl" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10332", "repo": "rosqual/rosdiscover", "url": "https://github.com/rosqual/rosdiscover/pull/159" }
gharchive/pull-request
Added config node source nodelet indication Config file now specifies if the node is a "normal" node or a "nodelet". Note that we could make kind optional and default to NORMAL. Note that we could make kind optional and default to NORMAL. Made this change - default is now NODE.
2025-04-01T06:40:16.002006
2024-08-15T12:26:40
2467953780
{ "authors": [ "jackoliver", "rosskouk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10333", "repo": "rosskouk/asknavidrome", "url": "https://github.com/rosskouk/asknavidrome/issues/47" }
gharchive/issue
[ISSUE] 405 Method Not Allowed - AskNavidrome -> Navidrome Describe the issue I have managed to create the Alexa skill, set up the asknavidrome service and expose it via Tailscale Funnel + nginx, but I get this log when I look at my systemctl status for the asknavidrome podman service. Operating System NixOS To Reproduce Using the Test Skill interface in Alexa Skills. It connects to the skill service okay, but when I request a specific song I get these logs: Logs aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,784 - asknavidrome.subsonic_api - DEBUG - In function search_artist() aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - DEBUG - In GeneralExceptionHandler aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - ERROR - General Exception: HTTP Error 405: Method Not Allowed aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - ERROR - Request Type Was: IntentRequest aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - ERROR - Intent Name Was: NaviSonicPlaySongByArtist aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,805 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:16] "POST / HTTP/1.0" 200 - aug 15 14:24:19 nixos asknavidrome[202111]: 2024-08-15 12:24:19,717 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:19] "GET /queue HTTP/1.0" 200 - aug 15 14:24:23 nixos asknavidrome[202111]: 2024-08-15 12:24:23,911 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:23] "GET /buffer HTTP/1.0" 200 - aug 15 14:24:27 nixos asknavidrome[202111]: 2024-08-15 12:24:27,482 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:27] "GET /history HTTP/1.0" 200 - Hi Jack, Can you confirm that the skill is connecting to navidrome, it would have been included earlier and if it is can you confirm which version of Navidrome you are using? On Thu, 15 Aug 2024, 13:27 Jack Oliver, @.***> wrote: Describe the issue I have managed to create the Alexa skill, set up the asknavidrome service and expose it via Tailscale Funnel + nginx, but I get this log when I look at my systemctl status for the asknavidrome podman service. Operating System NixOS To Reproduce Using the Test Skill interface in Alexa Skills. It connects to the skill service okay, but when I request a specific song I get these logs: Logs aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,784 - asknavidrome.subsonic_api - DEBUG - In function search_artist() aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - DEBUG - In GeneralExceptionHandler aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - ERROR - General Exception: HTTP Error 405: Method Not Allowed aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - ERROR - Request Type Was: IntentRequest aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,803 - root - ERROR - Intent Name Was: NaviSonicPlaySongByArtist aug 15 14:24:16 nixos asknavidrome[202111]: 2024-08-15 12:24:16,805 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:16] "POST / HTTP/1.0" 200 - aug 15 14:24:19 nixos asknavidrome[202111]: 2024-08-15 12:24:19,717 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:19] "GET /queue HTTP/1.0" 200 - aug 15 14:24:23 nixos asknavidrome[202111]: 2024-08-15 12:24:23,911 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:23] "GET /buffer HTTP/1.0" 200 - aug 15 14:24:27 nixos asknavidrome[202111]: 2024-08-15 12:24:27,482 - werkzeug - INFO - <IP_ADDRESS> - - [15/Aug/2024 12:24:27] "GET /history HTTP/1.0" 200 - — Reply to this email directly, view it on GitHub https://github.com/rosskouk/asknavidrome/issues/47, or unsubscribe https://github.com/notifications/unsubscribe-auth/ALZJF3TMQODS6YZAWYMMMMTZRSNBNAVCNFSM6AAAAABMSDR6MWVHI2DSMVQWIX3LMV43ASLTON2WKOZSGQ3DOOJVGM3TQMA . You are receiving this because you are subscribed to this thread.Message ID: @.***> Hey! I'm having a hard time accessing those logs. I'm using systemctl status podman-asknavidrome at the moment to get a log of what's happening. Navidrome version is [0.49.3](https://github.com/navidrome/navidrome/releases/tag/v0.49.3) Here is my NixOS configuration also: { config, pkgs, ... }: { config.virtualisation.oci-containers.containers."asknavidrome" = { autoStart = true; image = "ghcr.io/rosskouk/asknavidrome"; ports = [ "5000:5000" ]; environment = { NAVI_SKILL_ID = "REDACTED"; NAVI_SONG_COUNT = "50"; NAVI_URL = "https://REDACTED.ts.net/navidrome"; NAVI_USER = "REDACTED"; NAVI_PASS = "REDACTED"; NAVI_PORT = "443"; NAVI_API_PATH = "/rest"; NAVI_API_VER = "1.16.1"; NAVI_DEBUG = "3"; }; }; } Ah, I found it through journalctl. I am getting a 'Failed to connect to Navidrome'. All the details are correct, and I can /rest/ping it no problem from Postman. Ok, that's definitely the problem. I'd recommend you triple check the config. And if possible see if you can perform the ping API call from within the container On Thu, 15 Aug 2024, 13:57 Jack Oliver, @.***> wrote: Ah, I found it through journalctl. I am getting a 'Failed to connect to Navidrome'. All the details are correct, and I can /rest/ping it no problem from Postman. — Reply to this email directly, view it on GitHub https://github.com/rosskouk/asknavidrome/issues/47#issuecomment-2291219330, or unsubscribe https://github.com/notifications/unsubscribe-auth/ALZJF3VUPA3FXYPA4WCSXBDZRSQTZAVCNFSM6AAAAABMSDR6MWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDEOJRGIYTSMZTGA . You are receiving this because you commented.Message ID: @.***> I've taken it out of systemctl for now and just doing docker run instead. It's very confusing because it says connected, and then not. 2024-08-15 13:41:29,088 - asknavidrome.subsonic_api - DEBUG - Connected to Navidrome 2024-08-15 13:41:29,088 - asknavidrome.subsonic_api - DEBUG - In function ping() 2024-08-15 13:41:29,112 - asknavidrome.subsonic_api - ERROR - Failed to connect to Navidrome If I ping the URL from Postman, with the credentials I have in the Docker run env variables, i get a <subsonic-response> back, but it still says failed to connect when i boot the container. Very odd. % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 155 0 155 0 0 8403 0 --:--:-- --:--:-- --:--:-- 8611 <subsonic-response xmlns="http://subsonic.org/restapi" status="ok" version="1.16.1" type="navidrome" serverVersion="0.49.3 (8b93962f)"></subsonic-response> Looks like I can ping it from inside the container, so I'm not sure what's up at this point lol. Got it working. Couple things I noticed / tried; Works well with the demo navidrome instance, so I don't think it's anything wrong with the code. Didn't like it being on a /navidrome proxy pass on nginx - for some reason. Set it to the root /, and had to add /navidrome/rest in the API path, and then it worked. Thanks for being a great rubber duck! 😆 🦆
2025-04-01T06:40:16.257014
2019-12-02T02:56:23
530857874
{ "authors": [ "hebron-george", "route" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10334", "repo": "route/ferrum", "url": "https://github.com/route/ferrum/issues/35" }
gharchive/issue
Brave instead of Chrome? Is it possible to use Brave browser instead of Chrome? It's supposedly build off of Chromium, if that helps? Sorry I don't know much so I'm not even sure if this is a valid question to ask. Thanks! @hebron-george Sure if it's built on Chrome then it should work. Check out https://github.com/route/ferrum#customization and Ferrum::Browser.new(browser_path: "/path/to/browser") option and put a path to your browser bin there. Let me know if it works we can add it to README. @route Looks like that worked. I haven't done any thorough testing though so I'm not sure if there are some features that won't work but at the least - the browser loads up with the page I requested. Thanks!
2025-04-01T06:40:16.279289
2024-03-11T09:21:21
2178637460
{ "authors": [ "jstsddg", "rowanruseler" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10335", "repo": "rowanruseler/helm-charts", "url": "https://github.com/rowanruseler/helm-charts/pull/245" }
gharchive/pull-request
[pgadmin4] Bump appVersion to v8.4 What this PR does / why we need it: This update of pgAdmin4 fixes CVE-2024-2044. Which issue this PR fixes none Special notes for your reviewer: Checklist [Place an '[x]' (no spaces) in all applicable fields. Please remove unrelated fields.] [x] DCO signed [x] Chart Version bumped [ ] Variables are documented in the README.md [x] Title of the PR starts with chart name (e.g. [pgadmin4]) Note: it turns out that dpage/pgadmin4 does not have v8.5 available yet. Note: it turns out that dpage/pgadmin4 does not have v8.5 available yet. that is correct, cause 8.4 got pushed 4 days ago. So most likely in a month 8.5 will be released.
2025-04-01T06:40:16.333215
2020-11-16T22:47:32
744262558
{ "authors": [ "onnenon", "supcik" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10338", "repo": "rpi-ws281x/rpi-ws281x-go", "url": "https://github.com/rpi-ws281x/rpi-ws281x-go/pull/14" }
gharchive/pull-request
Support for arm64 Addresses #13 This PR modifies build tags for ws2811_sim.go and ws2811_arm.go (also renamed to ws2811_hw.go) to support building for arm64 targets like the Raspberry Pi 4B This has been tested, and works directly compiling on a Raspberry Pi 4B running Ubuntu 20.10 server for arm64. Thank you @onnenon for your contribution. Thank you! Happy I could contribute.
2025-04-01T06:40:16.377276
2020-04-03T17:17:27
593526753
{ "authors": [ "rpotter12" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10339", "repo": "rpotter12/whatsapp-play", "url": "https://github.com/rpotter12/whatsapp-play/pull/276" }
gharchive/pull-request
updates readme Issue that this pull request solves Solves: updates readme Proposed changes Updates readme - added new commands details, added wiki link Types of changes Put an x in the boxes that apply [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [x] Documentation update (Documentation content changed) [ ] Other (please describe): Checklist Put an x in the boxes that apply [x] My code follows the style guidelines of this project [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have made corresponding changes to the documentation [x] My changes generate no new warnings Screenshots changes link: https://github.com/rpotter12/whatsapp-play/blob/pull/README.md Here is an overview of what got changed by this pull request: Issues ====== + Solved 9 See the complete overview on Codacy
2025-04-01T06:40:16.388795
2023-10-31T18:08:20
1971039196
{ "authors": [ "rubgeax" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10341", "repo": "rrirower/DP2SceneGraphWizard", "url": "https://github.com/rrirower/DP2SceneGraphWizard/issues/49" }
gharchive/issue
Multiple errors on creation In the part where the images are chosen, an error appears, which I attach, and after that part where all the parameters that were configured are confirmed, when creating the files, more errors also appear, which I also attach. Last Version Version [e.g. <IP_ADDRESS>-alpha-8] It was a problem with my permissions, thank you, issue closed
2025-04-01T06:40:16.391512
2022-03-09T17:31:38
1164237843
{ "authors": [ "erabti", "rrousselGit" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10342", "repo": "rrousselGit/river_pod", "url": "https://github.com/rrousselGit/river_pod/issues/1266" }
gharchive/issue
ProviderOrFamily not exposed Could you please expose ProviderOrFamily so we can reference it? We need it in some implementation of writing custom state notifier provider builder, the generics can get too long so we are writing these helper to shorten that, but then we cannot declare the type of ' List? dependencies'. Could you expand on the reason why you need it? Let's say that we have a StateNotifier<Some<Really<Nested>>> that is used super frequently. When creating provider with: StateNotifierProvider<Notifier, State>, it's kind of unnecessary to repeat it if the state type and notifier are unified through out some major functionalities of the app. We do understand the Dart limitation of type inference and the need of two generics. But It's not possible instantiate a StateNotifierProvider fully if you can't add dependencies when needed to conform with riverpod API. I believe exposing this will allow more flexibility according to the user's architecture and needs. But how does this relate to ProviderOrFamily? Oh I see. I forgot that it's used by "dependencies". Fair enough then, it not being exposed is a mistake.
2025-04-01T06:40:16.419326
2023-07-14T21:26:04
1805601376
{ "authors": [ "DorisMai", "codecov-commenter", "kmdalton" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10343", "repo": "rs-station/careless", "url": "https://github.com/rs-station/careless/pull/121" }
gharchive/pull-request
Update README.md Updated installation instructions for GPU support. This definitely looks like an improvement to me. I approved the CI run and can merge when it completes. Codecov Report Merging #121 (a392641) into main (2ebe147) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## main #121 +/- ## ======================================= Coverage 78.82% 78.82% ======================================= Files 49 49 Lines 2135 2135 ======================================= Hits 1683 1683 Misses 452 452 Flag Coverage Δ unittests 78.82% <ø> (ø) Flags with carried forward coverage won't be shown. Click here to find out more. :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
2025-04-01T06:40:16.424906
2015-07-01T07:40:59
92303121
{ "authors": [ "bpoplauschi", "cbowns", "mythodeia", "wodetian" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10344", "repo": "rs/SDWebImage", "url": "https://github.com/rs/SDWebImage/issues/1202" }
gharchive/issue
My application crash in Iphone4s & IOS8.3 Filtered syslog: 2015-06-25 23:54:58.769 Meet[390]: <Error> objc[390]: Class SDImageCache is implemented in both /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Frameworks/SDWebImage.framework/SDWebImage and /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Meet. One of the two will be used. Which one is undefined. 2015-06-25 23:54:58.769 Meet[390]: <Error> objc[390]: Class SDWebImageDownloaderOperation is implemented in both /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Frameworks/SDWebImage.framework/SDWebImage and /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Meet. One of the two will be used. Which one is undefined. 2015-06-25 23:54:58.770 Meet[390]: <Error> objc[390]: Class SDWebImageManager is implemented in both /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Frameworks/SDWebImage.framework/SDWebImage and /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Meet. One of the two will be used. Which one is undefined. 2015-06-25 23:54:58.771 Meet[390]: <Error> objc[390]: Class SDWebImageCombinedOperation is implemented in both /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Frameworks/SDWebImage.framework/SDWebImage and /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Meet. One of the two will be used. Which one is undefined. 2015-06-25 23:54:58.771 Meet[390]: <Error> objc[390]: Class SDWebImagePrefetcher is implemented in both /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Frameworks/SDWebImage.framework/SDWebImage and /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Meet. One of the two will be used. Which one is undefined. 2015-06-25 23:54:58.771 Meet[390]: <Error> objc[390]: Class SDWebImageDownloader is implemented in both /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Frameworks/SDWebImage.framework/SDWebImage and /private/var/mobile/Containers/Bundle/Application/302D9935-E381-46AE-8D03-B58829D2CD24/Meet.app/Meet. One of the two will be used. Which one is undefined. 2015-06-25 23:54:59.184 Meet[390]: <Error> assertion failed: 12F70: libxpc.dylib + 51923 [89A85F22-6D73-33D8-9D7A-FCF58D58C58E]: 0x7d 2015-06-25 23:54:59.942 Meet[390]: <Error> SecTrustEvaluate [leaf AnchorTrusted] 2015-06-25 23:55:00.218 Meet[390]: <Warning> Could not load the " " image referenced from a nib in the bundle with identifier "com.yktx.Meets" 2015-06-25 23:55:05.282 Meet[390]: <Error> *** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: '*** -[__NSPlaceholderDictionary initWithObjects:forKeys:count:]: Does the app run fine on the simulator or an iPhone 5/6/6P ? Is it a swift project? How did you add the SDWebImage.framework in your project? Can you check if you have added it twice? Hello, I'm using swift, actually, I'm here is to run everything possible, but submit appStore, feedback to me is on the iphone 4 collapse, then, I saw the show my log is SwWeb have a problem. I am using CocoaPods added,I didn't add twice... I am also very upset So you only face this problem in the iPhone 4S device? I think it should be. At least I hand the 5S, touch, 6plus, 6 real machine running no problem. Here I no 4S, only the virtual machine, virtual machines running if no problem, so I think is not only real 4S will appear in this issue. I don't know why this is a problem, and I don't know what 4S is like. I intend to file a, have a look will not have this problem. Your app is directly linking SDWebImage as a static library, and using iOS 8's dynamic frameworks to link it at runtime. You'll want to remove one of those two. @wodetian any updates on this? @wodetian I'm going to close the issue. Please reopen when you can bring updates.
2025-04-01T06:40:16.427306
2019-10-24T08:30:56
511790330
{ "authors": [ "rsaihe", "sainnhe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10345", "repo": "rsaihe/gruvbox-material-kitty", "url": "https://github.com/rsaihe/gruvbox-material-kitty/pull/2" }
gharchive/pull-request
adjust aqua and blue in the light variant Aqua and blue are not distinguishable enough in the light variant, I'm tring to make them more distinguishable. Hmmm, I think it can be a little bit difficult to distinguish them in the light variant, but I think the original is decent enough as is. The new version seems a little bit harder to distinguish and I think the contrast looks good enough already. So, my vote is to stick to the original. Sorry for bothering, but I need other people's help. And don't worry, I don't mind helping out. It's not a bother at all! 😄 Thanks for your help! I decided to use the original color palette instead, but slightly increased the contrast of black(used in the Normal highlight group). I'm personally satisfied with this. What do you think? Sure, I think this looks good! I'll go ahead and merge.
2025-04-01T06:40:16.460173
2016-08-12T21:07:59
170959566
{ "authors": [ "mwcz", "rserota" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10346", "repo": "rserota/wad", "url": "https://github.com/rserota/wad/pull/69" }
gharchive/pull-request
point bower "main" at the build wad.js file From the current bower.json: "main": "wad.js", wad.js is inside build/ though, so tools like wiredep and bowerRequirejs can't find it. I suggest pointing it at build/wad.js. I could be missing something though! Thoughts? Thanks for contributing!
2025-04-01T06:40:16.500270
2019-04-26T07:19:51
437533585
{ "authors": [ "JonRowe", "benoittgt", "danielwaterworth", "mikegee" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10347", "repo": "rspec/rspec-core", "url": "https://github.com/rspec/rspec-core/issues/2620" }
gharchive/issue
Let! shadowing Here's the behaviour in question, this test fails. describe do let!(:foo) { 1 } context do let(:foo) { raise "boom" } it {} end end The inner foo overrides the outer foo's behaviour, but it doesn't turn foo into a let from a let!. This isn't what I would expect although I understand why it happens. I don't see this as a bug. 😊 So I am closing the issue. We can re-open it if needed. I'd say the example code's behavior is somewhat surprising. One might expect lets and let!s to be completely independent and not interact with each other. Perhaps a documentation adjustment is warranted? What do you think about it @JonRowe ? From: https://relishapp.com/rspec/rspec-core/v/3-8/docs/helper-methods/let-and-let Use let to define a memoized helper method. The value will be cached across multiple calls in the same example but not across examples. Note that let is lazy-evaluated: it is not evaluated until the first time the method it defines is invoked. You can use let! to force the method's invocation before each example. It should therefore be clear this is expected Ruby semantics for methods that have been defined. If we add to this I think it should be in the form of adding a note to the second paragraph that "This can not be undone and will apply to any future conflicting method definition". Or some such. I'd also support highlighting the fact that methods are defined, maybe we just add helper *method*
2025-04-01T06:40:16.504782
2011-11-18T23:56:55
2285981
{ "authors": [ "garybernhardt", "justinko" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10348", "repo": "rspec/rspec-core", "url": "https://github.com/rspec/rspec-core/issues/506" }
gharchive/issue
Modules included in example groups aren't visible in Ruby 1.9.3 In Ruby 1.9.3, when including a module in an example group, constants defined in the module aren't in scope: module X module Y def self.a_method 5 end end end describe X::Y do include X it "can access Y" do Y.a_method.should == 5 end end This fails with "NameError: uninitialized constant Y". This worked fine in 1.9.2, so it seemed like a Ruby change caused the regression. It happens both with rspec-core 2.6.0 and 2.7.1. Looks like a Ruby bug: module X module Y def self.a puts 'a' end end end class Foo end a = Class.new(Foo) do include X def self.b Y.a end end puts a.b # => uninitialized constant Y (NameError) Looks like 1.9.3 has a bug with "block evaluation" (Class.new, module_eval, etc.) I just tried it with 1.9.3-head, still no fix :( @garybernhardt - I'll let you upsert this bug to ruby core. Done: http://redmine.ruby-lang.org/issues/5657 Apparently this was never supposed to work in the first place; the fact that it worked on 1.9.2 was a bug. Shame. :/ http://redmine.ruby-lang.org/issues/5657
2025-04-01T06:40:16.508017
2023-02-27T23:39:09
1602123410
{ "authors": [ "JonRowe", "JunichiIto" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10349", "repo": "rspec/rspec-metagem", "url": "https://github.com/rspec/rspec-metagem/issues/71" }
gharchive/issue
https://relishapp.com/rspec shows application error Currently https://relishapp.com/rspec shows application error. Could you investigate it? Related issue: https://github.com/rspec/rspec-metagem/issues/70 , https://github.com/rspec/doc-site/issues/5 👋 There is an issue with relishapp at the moment, you can find our website at https://rspec.info which has our API documentation, or read the cucumber scenarios directly from repos. See: https://github.com/rspec/doc-site/issues/5#issuecomment-1446152910 Closing because this is not an rspec issue.
2025-04-01T06:40:16.511388
2024-03-08T13:00:38
2176000417
{ "authors": [ "JonRowe", "eugeneius" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10350", "repo": "rspec/rspec-rails", "url": "https://github.com/rspec/rspec-rails/pull/2742" }
gharchive/pull-request
Only load example group modules when they're used Currently the example group modules for all spec types are always loaded, regardless of which type of spec is being run. Some of those modules force parts of the user's application to load; for example, RSpec::Rails::HelperExampleGroup includes ActionView::TestCase::Behavior, and loading ActionView::TestCase loads all of the application's helpers. By autoloading these modules and including them lazily the first time the corresponding spec type is defined, we can avoid loading parts of the user's application unnecessarily. Co-authored by @ilianah 🍐 Anything we can do to load things only when needed gets a plus from me, however I'm slightly concerned the build is failing because our snippets are exposing the change in whats loaded by default (in the case of 3.2 its just minitest being missing) which makes me wonder if this is a breaking change from a users perspective but I'm willing to be convinced its not if its easy to fix the build. The suggested improvement to rspec-core is also fine, I think the helpers you are using were actually invented for rspec-rails ironically (for the define derived types options) Previously ActiveSupport::TestCase was always loaded, so minitest would be required here: https://github.com/rails/rails/blob/v<IP_ADDRESS>/activesupport/lib/active_support/test_case.rb#L3 The most straightforward fix for that is to continue to load minitest by requiring it explicitly; I've pushed a commit to do that.
2025-04-01T06:40:16.517571
2022-11-03T12:01:57
1434545908
{ "authors": [ "rssh", "wiltonlazary" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10351", "repo": "rssh/cps-async-connect", "url": "https://github.com/rssh/cps-async-connect/issues/4" }
gharchive/issue
Any clue how to implement zio2 FiberRef? Expression type: FiberRef[Map[String, Any]] Symbol signature: def await[F[$1], T, G[$2]](f: F[T])(using am: CpsAwaitable[F], ctx: CpsMonadContext[G]): T Pseudofunction, which can be used inside async block, to 'await' (i.e. receive value of t:T from ft:F[T]). Can't find cps.CpsMonadConversion[[A >: scala.Nothing <: scala.Any] => zio.ZIO[zio.Scope, scala.Nothing, A], zio.Task]: given instance zioToZio in package cps.monads.zio does not match type cps.CpsMonadConversion[[A] =>> zio.ZIO[zio.Scope, Nothing, A], zio.Task]sbt import zio.{given, *} import cps.monads.zio.{given, *} import cps.{given, *} object ctx: private val ref = zio.FiberRef.make(Map.empty[String, Any]) def withCtx[R](body: => Task[R]) = async[Task] { val ctx = await(ref) } Let-s try to deduce types by hand: We have zioToZio with the next signature: given zioToZio[R1,R2<:R1,E1,E2>:E1]: CpsMonadConversion[[T] =>> ZIO[R1,E1,T], [T]=>>ZIO[R2,E2,T]] Our failed conversion: // zio[Scope,Nothing,A] => zio.Task[A] // zio[Any, Throwable, A] // // . R1 = Scope, E1 = Nothing // R2 = Any E1 = Throwable // . R2 <: R1 = (Any <: Scope) - false. So, resolution is correct - this should not be compiled, because environment of Task (Any) is not a subset of Scope. For rechecking we can emulate same logic without async: def withCtxNoAwait[R](body: => Task[R]):Task[Unit] = { ref.map(_ => ()) } To have Scope in task we should have something like: def withCtx[R](body: => Task[R]) = async[[T]=>>ZIO[Scope,Throwable,T]] { val ctx = ref body } Btw, let's think, are R2<:R1 requirement is correct(?) R - is a set of requirements. Any - no requirements, buf FiberRef introduce Scope requirement, so I think it's necessory. Great, thanks very much.
2025-04-01T06:40:16.520340
2016-02-01T10:56:34
130329103
{ "authors": [ "JoaoGFarias", "rssh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10352", "repo": "rssh/scala-gopher", "url": "https://github.com/rssh/scala-gopher/issues/9" }
gharchive/issue
License Hello @rssh and @alisa-rend, I wasn't able to find a license in the project. scala-gopher was supposed to be open source or is it a proprietary software? What are the restriction in using and expanding it? Yes, it's open source. License default is Apache [http://www.apache.org/licenses/LICENSE-2.0]. If you prefer other OSS-compatible license, tell me. Thanks for notice, I will add reference to docs. Thanks @rssh
2025-04-01T06:40:16.610600
2017-02-08T06:12:08
206105577
{ "authors": [ "GitanjaliThete", "javierluraschi", "kevinykuo" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10353", "repo": "rstudio/sparklyr", "url": "https://github.com/rstudio/sparklyr/issues/475" }
gharchive/issue
Error while connecting with spark with sparkly using RStudio with Kerberos enabled machines I have 2 machine clustor. I am trying to connect to spark. When I try to connect using R shell, It is working fine. Getting following error from Rstudio. Connection: sc <- spark_connect(master="yarn-client",spark_home = "/usr/hdp/<IP_ADDRESS>-37/spark", config = list( default = list( spark.submit.deployMode= "client", spark.yarn.keytab= "path/keytab", spark.yarn.principal= "user", spark.executor.instances= 1, spark.executor.memory= "250M", spark.executor.cores= 4, spark.driver.memory= "250M"))) Error: Failed during initialize_connection: org.apache.hadoop.service.ServiceStateException: java.io.IOException: Failed on local exception: java.io.IOException: javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]; Host Details : local host is: "host1"; destination host is: "host2":8020; @GitanjaliThete are you using RStudio Server PRO? Kerberized sessions are only supported in the RStudio Server PRO. If that's the case, you would have to configure: /etc/pam.d/rstudio-original /etc/pam.d/rstudio-session /etc/rstudio/rserver.conf For details, see http://docs.rstudio.com/ide/server-pro/r-sessions.html#kerberos - Has this been configured already? Thank javierluraschi! Yes, I installed RStudio Server Pro 1.0.136 on CentOS. Also configured following files for PAM Still not able to connect. Please let me know in case made some mistake. /etc/pam.d/rstudio #%PAM-1.0 auth sufficient pam_rootok.so session required pam_env.so readenv=1 session required pam_env.so readenv=1 envfile=/etc/default/locale session required pam_limits.so @include common-auth @include common-account @include common-session auth include system-auth auth optional pam_mount.so use_first_pass account required pam_unix.so /etc/pam.d/rstudio-session auth sufficient pam_rootok.so session required pam_env.so readenv=1 session required pam_env.so readenv=1 envfile=/etc/default/locale session required pam_limits.so @include common-auth @include common-account @include common-session auth include system-auth auth optional pam_mount.so use_first_pass account required pam_unix.so /etc/rstudio/rserver.conf Server Configuration File auth-pam-sessions-profile=rstudio-session auth-pam-sessions-use-password=1 Try opening up a terminal from within the browser (tools -> shell?) and run kinit It is working after kinit. Thanks :)
2025-04-01T06:40:16.613363
2016-06-29T15:22:05
162953974
{ "authors": [ "dselivanov", "javierluraschi" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10354", "repo": "rstudio/sparklyr", "url": "https://github.com/rstudio/sparklyr/issues/50" }
gharchive/issue
Working with YARN cluster in client mode How to connect to yarn cluster in yarn-client mode? With SparkR my connection looks like: spark_env = list('spark.executor.memory' = '8g', 'spark.executor.instances' = '32', 'spark.executor.cores' = '4', 'spark.driver.memory' = '4g', 'spark.kryoserializer.buffer.max' = '1g', 'spark.yarn.executor.memoryOverhead' ='1000') sc <- sparkR.init(master = "yarn-client", appName = "sparkr", sparkEnvir = spark_env) sc <- spark_connect(master = "yarn-client") with the SPARK_HOME pointing to the right spark directory? To pass the config values you can use the config parameter in spark_connect or the config.yaml file based on the template included in inst/config Indeed, it works. When I tried to connect first time, for some reason it get stuck for several minutes. spark_log also showed nothing, so I decided to ask... But now it works fine. Thank you, closing.
2025-04-01T06:40:16.634681
2018-04-25T19:58:38
317772329
{ "authors": [ "davidfischer", "ericholscher" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10355", "repo": "rtfd/readthedocs.org", "url": "https://github.com/rtfd/readthedocs.org/pull/4022" }
gharchive/pull-request
Support sign in and sign up with GH/GL/BB While this was technically supported if you knew the URLs, this adds formal support for logging in and signing up with our providers (currently GitHub, BitBucket, and GitLab). How this works Let's say your user account already is connected to GitHub because you actually use Read the Docs with your GitHub account. The "Sign in with GitHub" button will just work for you. You can still use your old username and password if you like. If your GitHub account is not connected to anything and you click the "Sign in with GitHub", you will authenticate with GitHub, and then you will get a confirmation screen where you tie your GitHub account and email to a new Read the Docs username (see screenshot below). After confirming, you'll be logged in. The first and last name may be populated from the provider if available. BitBucket works exactly the same as GitHub. I didn't actually test GitLab but I assume it does. Note: There is no technical difference between the "Sign in with.." and "Sign up with.." buttons. If your social account is already connected to a user account, you will be logged in. If it isn't, you will be asked to confirm a new sign up. Weird edge cases Say your social account isn't connected, you click "Sign in with..." and go to the confirm new user screen. The email address already exists and you try to submit it again. You will see an error An account already exists with this e-mail address. Please sign in to that account first, then connect your Bitbucket account. This is strange a little bit at first but I believe it is correct functionality. It also won't let you create a duplicate username but that seems standard. Say a brand new user signs up with their GitHub account and then disconnects GitHub from their account (on https://readthedocs.org/accounts/social/connections/). The user won't be able to login with GitHub any longer (the accounts aren't connected) but they also can't login with a username and password (their account has no password). The only way to retrieve the account in this case is with a password reset to the email on file. I just tested the email reset and it worked. 👍 I don't believe I got an verification email though, which might make things not work in the future. I wonder if there's a setting or something about email verification that we need to turn on? The setting is SOCIALACCOUNT_EMAIL_VERIFICATION and it takes the same values as ACCOUNT_EMAIL_VERIFICATION. For some reason it is set to 'none' in development. I'll make the settings match in dev and open a PR for the production settings. 👍 looks good to me. I imagine we might get some support requests around this for edge cases, but we can take those as they come.
2025-04-01T06:40:16.667508
2016-08-23T14:44:04
172718470
{ "authors": [ "chillus", "rtrouton" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10356", "repo": "rtrouton/First_Boot_Package_Install_Generator", "url": "https://github.com/rtrouton/First_Boot_Package_Install_Generator/issues/3" }
gharchive/issue
Can't modif Hello When I do the following to bypass network check, get error "unable to execute script Applescript" Download/install the app, browse the package contents, and copy "installer_build_components.tgz" into a temporary directory somewhere. Extract the contents of "installer_build_components.tgz" Edit "private/user/firstbootpackageinstall_noswupdates.sh". Comment out or delete lines 84-114 (the section that checks for a network connection) Recompress that directory (e.g. "tar -cvzf installer_build_components.tgz installer_build_components") Replace the "installer_build_components.tgz" file in the package contents with the one you just created. Do you have an other way to bypass this check action ? Tx Nope. If you want to modify how this application works, the responsibility for making it work afterwards is on you as the person making the modifications.
2025-04-01T06:40:16.677673
2016-06-22T10:59:05
161651842
{ "authors": [ "Skywalker-11", "dawud-tan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10357", "repo": "rtyley/spongycastle", "url": "https://github.com/rtyley/spongycastle/pull/21" }
gharchive/pull-request
modify become-spongy.sh to apply changes for working mail package for android This changes the become-spongy.sh script to get mail packages working in android see #20 for the result after appying the script nice..
2025-04-01T06:40:16.679582
2017-04-24T21:52:11
223962072
{ "authors": [ "dhiru1990", "ruananswer" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10358", "repo": "ruananswer/twitter-anomalyDetection-java", "url": "https://github.com/ruananswer/twitter-anomalyDetection-java/issues/2" }
gharchive/issue
run problem and main file how to run this project..where is the main file of this program In the test path src/test/java/com/github/ruananswer/
2025-04-01T06:40:16.691658
2023-10-07T07:08:58
1931229600
{ "authors": [ "SuZhou-Joe", "codecov-commenter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10359", "repo": "ruanyl/OpenSearch-Dashboards", "url": "https://github.com/ruanyl/OpenSearch-Dashboards/pull/209" }
gharchive/pull-request
Store current workspaceId into hash Description Issues Resolved Screenshot Testing the changes Check List [ ] All tests pass [ ] yarn test:jest [ ] yarn test:jest_integration [ ] yarn test:ftr [ ] New functionality includes testing. [ ] New functionality has been documented. [ ] Update CHANGELOG.md [ ] Commits are signed per the DCO using --signoff Codecov Report :exclamation: No coverage uploaded for pull request base (main@9c192a3). Click here to learn what that means. The diff coverage is n/a. :exclamation: Current head ce64684 differs from pull request most recent head 05fda74. Consider uploading reports for the commit 05fda74 to get more accurate results @@ Coverage Diff @@ ## main #209 +/- ## ======================================= Coverage ? 35.35% ======================================= Files ? 1864 Lines ? 35774 Branches ? 6482 ======================================= Hits ? 12649 Misses ? 22300 Partials ? 825 Flag Coverage Δ Linux_4 35.35% <0.00%> (?) Flags with carried forward coverage won't be shown. Click here to find out more. :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
2025-04-01T06:40:16.709641
2020-09-15T21:26:21
702284792
{ "authors": [ "marcandre", "sergeev17" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10360", "repo": "rubocop-hq/rubocop-ast", "url": "https://github.com/rubocop-hq/rubocop-ast/pull/114" }
gharchive/pull-request
Make most constants private Constants for node.rb are left as is (other cops depend on them) Constants for node_pattern.rb are left as is (whole class is @api private)
2025-04-01T06:40:16.721667
2020-10-19T11:01:27
724506172
{ "authors": [ "bquorning", "chubchenko", "pirj" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10361", "repo": "rubocop-hq/rubocop-rspec", "url": "https://github.com/rubocop-hq/rubocop-rspec/issues/1044" }
gharchive/issue
FactoryBot/AttributeDefinedStatically doesn't work well When processing the following code: # spec/lib/application/configuration_spec.rb describe '.ignore' do context 'when the block is given' do it 'changes the count of the registered regex' do expect do described_class.ignore do |ignore| ignore << /(a|b)/ ignore.push(/.*/) end end.to change(described_class.ignore, :count).by(2) end end end I have the next recommendation: spec/lib/application/configuration_spec.rb:33:13: C: FactoryBot/AttributeDefinedStatically: Use a block to declare attribute values. ignore.push(/.*/) ^^^^^^^^^^^^ Also, I have the same recommendation in another file: # spec/dummy/config/initializers/seo.rb config.ignore do |ignore| ignore.push(%r{\A/admin.*}.freeze) ignore.push(%r{\A/\z}.freeze) end spec/dummy/config/initializers/seo.rb:28:5: C: FactoryBot/AttributeDefinedStatically: Use a block to declare attribute values. ignore.push(%r{\A/admin.*}.freeze) ^^^^^^^^^^^^^^^^^^^^^^^^^^ spec/dummy/config/initializers/seo.rb:29:5: C: FactoryBot/AttributeDefinedStatically: Use a block to declare attribute values. ignore.push(%r{\A/\z}.freeze) ^^^^^^^^^^^^^^^^^^^^^^ rubocop -V 0.93.1 (using Parser <IP_ADDRESS>, rubocop-ast 0.8.0, running on ruby 2.6.5 x86_64-darwin20) bundle info rubocop-rspec * rubocop-rspec (1.43.2) Summary: Code style checking for RSpec files Homepage: https://github.com/rubocop-hq/rubocop-rspec Path: /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-rspec-1.43.2 Wondering why this https://github.com/rubocop-hq/rubocop-rspec/blob/e11f6e3c4a539203185e0e11089c8fe2e8f05393/config/default.yml#L8 doesn't work for you. FactoryBot cops should only scan their directories. Would you like to debug this a little further, say, by running rubocop with --debug flag, or setting a breakpoint somewhere here https://github.com/rubocop-hq/rubocop-rspec/blob/0d5e03d686d569f9e6f2c00ddce0b04d6463cb3c/lib/rubocop/cop/rspec/base.rb#L39 ? Regarding the AllCops: RSpec/FactoryBot: Patterns: - spec/factories/**/*.rb this was the first thing I tried, but my attempt failed - the same behavior as before 😞 . Here is an example of output with the --debug option: bundle exec --gemfile gemfiles/rubocop.gemfile rubocop spec/dummy/config/initializers/seo.rb --debug For /Users/chubchenko/applications/seo: configuration from /Users/chubchenko/applications/seo/.rubocop.yml configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-performance-1.8.1/config/default.yml configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-performance-1.8.1/config/default.yml Default configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-0.93.1/config/default.yml configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-rails-2.8.1/config/default.yml configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-rails-2.8.1/config/default.yml configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-rspec-1.43.2/config/default.yml configuration from /Users/chubchenko/.rvm/gems/ruby-2.6.5/gems/rubocop-rspec-1.43.2/config/default.yml Inspecting 1 file Scanning /Users/chubchenko/Applications/private/cms-seo/spec/dummy/config/initializers/seo.rb Loading cache from /Users/chubchenko/.cache/rubocop_cache/6655e55d54761e1e00b5da5b9a104451067030e3/6d7a3b621ca1730e04accd938619e4bdab66cfb1/48dc38c32c72316fef6fe5bddb16f10a6b0fd8cc C Offenses: spec/dummy/config/initializers/seo.rb:28:5: C: FactoryBot/AttributeDefinedStatically: Use a block to declare attribute values. ignore.push(%r{\A/admin.*}.freeze) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ spec/dummy/config/initializers/seo.rb:29:5: C: FactoryBot/AttributeDefinedStatically: Use a block to declare attribute values. ignore.push(%r{\A/\z}.freeze) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 1 file inspected, 2 offenses detected, 2 offenses auto-correctable Finished in 0.7074670000001788 seconds Do you mind sharing the /Users/chubchenko/applications/seo/.rubocop.yml config? @bquorning Do you think we need to make some changes to the relevant_rubocop_rspec_file? in order to support RSpec/FactoryBot/Patterns setting? It seems that the Patterns setting is one of our own, not supported by the RuboCop config itself. And our logic on top of it it only respects RSpec/Patterns, not RSpec/FactoryBot/Patterns. Yes, no problem. Here is the configuration file: require: - "rubocop-performance" - "rubocop-rails" - "rubocop-rspec" AllCops: Exclude: - "bin/*" - "spec/dummy/bin/*" - "spec/dummy/db/**/*" TargetRubyVersion: 2.5 NewCops: enable ############### Layout ############### Layout/LineLength: Max: 120 ############### Naming ############### Naming/PredicateName: AllowedMethods: - "is_a?" - "have_redirect" ############### Style ############### Style/Documentation: Enabled: false ############### Metrics ############### Metrics/AbcSize: IgnoredMethods: - "change" - "redirectable" Metrics/BlockLength: ExcludedMethods: - "register" - "describe" - "shared_examples" Metrics/MethodLength: ExcludedMethods: - "change" ############### Rails ############### Rails/SkipsModelValidations: Exclude: - "app/models/seo_module/redirect.rb" - "app/services/seo_module/redirect/kill_cycling.rb" - "lib/seo_module/admin/redirect.rb" Rails/DynamicFindBy: Whitelist: - "find_by_page_type_and_slug" ############### RSpec ############### RSpec/DescribeClass: Exclude: - "spec/system/**/*" RSpec/ExampleLength: Exclude: - "spec/system/**/*" Max: 10 RSpec/MultipleExpectations: Exclude: - "spec/system/**/*" Max: 2 Do you think we need to make some changes to the relevant_rubocop_rspec_file? in order to support RSpec/FactoryBot/Patterns setting? Actually, I started thinking maybe it’s time to move the FactoryBot code into a separate gem (e.g. RuboCop-FactoryBot). Do the three FactoryBot cops rely on any RSpec cop specific code (the Language module, etc.) at all? I support the idea. Reasoning: no common code (that I could find with RSpec cops) do not overlap on scanned files (FactoryClassName, AttributeDefinedStatically) it's probably the best moment to extract at about the moment when RuboCop hits 1.0 it makes it possible to find misuses when using FactoryBot with Minitest as well ... I have a number of FactoryBot-related cop ideas based on my recent projects code assession Cons: CreateList really should watch both spec files and factories, since both may call create in a loop, or create_list. We'll probably have to keep it here Side note: we will still have to use über-departments since the whole problem was due to Rails/HttpStatus and RSpec/FilePath. But we can opt-out of this in rspec-factory_bot. Concerning file locations, as per FactoryBot doc, we should be looking in: test/factories.rb spec/factories.rb test/factories/*.rb spec/factories/*.rb Another wild idea would be to submit those cops directly to FactoryBot itself, but according to my recent experience, their maintainers are not too responsive. This should be fixed in #1063 that is released in 2.0. Please feel free to reopen if it still doesn't work properly.
2025-04-01T06:40:16.727837
2017-12-08T09:02:32
280419606
{ "authors": [ "Ana06" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10362", "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/issues/5199" }
gharchive/issue
todo file is not taken in to account if cop is in the configuration When regenerating the todo file, if there is a configuration in .rubocop.yml is duplicated and when running rubocop having a cop in both .rubocop.yml and .rubocop_todo.yml the cop only exclude the ones in .rubocop.yml. This affect Rails/SkipsModelValidations. I didn't try it with other cops. Steps to reproduce the problem & Actual behavior .rubocop.yml: Rails/SkipsModelValidations: Exclude: - 'src/api/db/migrate/*' - 'src/api/app/jobs/*' - 'src/api/spec/models/bs_request_action_spec.rb' - 'src/api/spec/models/review_spec.rb' I move the offenses from src/api/app/jobs/* to the file src/api/lib/tasks/project_log_rotate_manually.rake. If I regenerate the todo file, all this is added: - 'src/api/db/migrate/*' - 'src/api/app/jobs/*' - 'src/api/lib/tasks/project_log_rotate_manually.rake' - 'src/api/spec/models/bs_request_action_spec.rb' - 'src/api/spec/models/review_spec.rb' That's mean that the configuration in .rubocop.yml is somehow taken into account, but it is also being added to the .todo file. There are currently no even offenses in src/api/app/jobs/*, it appears because it is in .rubocop.yml. Also if with this you run rubocop, I get the following offense: src/api/lib/tasks/project_log_rotate_manually.rake:26:83: C: Rails/SkipsModelValidations: Avoid using update_all because it skips validations. Event::Base.where(project_logged: false).where(["created_at < ?", oldest_date]).update_all(project_logged: true) So it seems that the todo file is ignored if there is already a configuration in .rubocop.yml. This affect Rails/SkipsModelValidations. I didn't try it with other cops. Expected behavior When regenerating the todo file, if there is a configuration in .rubocop.yml, it is taking into account, in the way that only new offenses are excluded in the todo file. When running rubocop having a cop in both .rubocop.yml and .rubocop_todo.yml the cop exclude the files that appear in both files. RuboCop version $ rubocop -V 0.51.0 (using Parser <IP_ADDRESS>, running on ruby 2.4.2 x86_64-linux-gnu) It happens also for other cops, for example Style/DateTime so it seems unrelated from the cop. I think this is important and shouldn't be closed (not sure if it still happen in the current master version but it would be worthwhile to doble check)
2025-04-01T06:40:16.730465
2014-03-02T20:22:30
28587412
{ "authors": [ "georgemillo", "natakina", "nayefc" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10363", "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/issues/849" }
gharchive/issue
Ignore Rubocop Errors How can I ignore certain Rubocop errors? I am trying to ignore this one for example: Method has too many lines. [11/10] Is there a way to reference that error in some dot file? Note that I am running it in Emacs using Flycheck. Thanks, For further reference, the new link for "Disabling Cops Within Source Code" in @jonas054's post is here For reference, if anyone else stumbles upon this, the name of the cop is MethodLength. You can disable it in the code with: # rubocop:disable MethodLength def foo # ... end In the end of method we also need to add comment # rubocop:enable MethodLength
2025-04-01T06:40:16.735475
2022-10-10T13:15:11
1403161642
{ "authors": [ "klyonrad", "koic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10364", "repo": "rubocop/rubocop", "url": "https://github.com/rubocop/rubocop/issues/11061" }
gharchive/issue
Disallow method names with send and public_send Is your feature request related to a problem? Please describe. every so often ruby programmers need to write code that sends data somewhere Then they add a method that they name send another developer adds a TODO comment "don't overwrite that ruby method" Even though send is not exactly a keyword with methods named send one could the control for metaprogramming that relies on using this method Describe the solution you'd like a cop in the Lint namespace Describe alternatives you've considered None Additional context https://ruby-doc.org/core-3.1.2/Object.html#method-i-send Ruby provides __send__ considering that send is overridden. In other words, __send__ should not be overridden. I think metaprogramming like using send is used only in limited cases. Disabled by default, but Style/Send cop will already be there for that. Of course, I don't recommend overriding send, but I think Lint warning is too strict. Sorry, I didn't know that this cop existed already and didn't know that a standard library has a send method :) https://docs.rubocop.org/rubocop/cops_style.html#stylesend
2025-04-01T06:40:16.740560
2023-02-08T14:44:57
1576250039
{ "authors": [ "bbatsov", "koic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10365", "repo": "rubocop/rubocop", "url": "https://github.com/rubocop/rubocop/pull/11550" }
gharchive/pull-request
[Fix #11549] Fix an error for third party cops Fixes #11549. This PR fixes an error for third party cops when inheriting RuboCop::Cop::Cop. It makes the same changes to RuboCop::Cop::Cop as RuboCop::Cop::Base in #10839. Confirmed that rubocop-i18n tests pass again. Before submitting the PR make sure the following are checked: [x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences. [x] Wrote good commit messages. [x] Commit message starts with [Fix #issue-number] (if the related issue exists). [x] Feature branch is up-to-date with master (if not - rebase it). [x] Squashed related commits together. [ ] Added tests. [x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code. [x] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details. @bbatsov This seems to be a critical path for upgrading to RuboCop 1.45. It looks good that 1.45.1 will be released ASAP after the merge. Sure. I'll cut a new release right away. Thank you for the quick release!
2025-04-01T06:40:16.830806
2019-05-30T21:28:49
450514848
{ "authors": [ "clumnah", "rwyoung01" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10366", "repo": "rubrikinc/rubrik-scripts-for-powershell", "url": "https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91" }
gharchive/issue
Export-RubrikDatabasejob pass json file in SQL Server Agent Job When I use the following script in a SQL Agent job as a cmdshell script type powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json Nothing happens and I get the following message in the job history The step did not generate any output. Process Exit Code 0. The step succeeded. If I trying the script as a Powershell type D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json I get the following message A job step received an error at line 177 in a PowerShell script. The corresponding line is ' } while ($RubrikRequestInfo.status -notin $ExitList) '. Correct the script and reschedule the job. The error information returned by PowerShell is: 'You must provide a value expression on the right-hand side of the '-' operator. '. Process Exit Code -1. The step failed. The only way it will run is not passing the json file as a parm. Do you know what I am doing wrong Are you using the Powershell run type or the Operating system run type? From: rwyoung01<EMAIL_ADDRESS>Sent: Thursday, May 30, 2019 5:28 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Subscribed Subject: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) When I use the following script in a SQL Agent job as a cmdshell script type powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json Nothing happens and I get the following message in the job history The step did not generate any output. Process Exit Code 0. The step succeeded. If I trying the script as a Powershell type D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json I get the following message A job step received an error at line 177 in a PowerShell script. The corresponding line is ' } while ($RubrikRequestInfo.status -notin $ExitList) '. Correct the script and reschedule the job. The error information returned by PowerShell is: 'You must provide a value expression on the right-hand side of the '-' operator. '. Process Exit Code -1. The step failed. The only way it will run is not passing the json file as a parm. Do you know what I am doing wrong — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHubhttps://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSBJYXFMME3V3WYT67DPYBBJFA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4GW2J6QA&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324617384&sdata=2N3GHn3rOKqfuPObP%2FW1Ed0yrjw4ay0lKDk0F3JcK4U%3D&reserved=0, or mute the threadhttps://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSG45G3PG6LW6J5OJSDPYBBJFANCNFSM4HRL2OWQ&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324627395&sdata=uqrLTp%2FotYCmTVLFSqtB3FSX%2BcloTd1pMw8ANWL4wPs%3D&reserved=0. i have tried both. the first example is with the cmdexe and the second example is with the powershell type. Russell Young On Mon, Jun 3, 2019 at 7:44 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Are you using the Powershell run type or the Operating system run type? From: rwyoung01<EMAIL_ADDRESS>Sent: Thursday, May 30, 2019 5:28 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Subscribed Subject: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) When I use the following script in a SQL Agent job as a cmdshell script type powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json Nothing happens and I get the following message in the job history The step did not generate any output. Process Exit Code 0. The step succeeded. If I trying the script as a Powershell type D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json I get the following message A job step received an error at line 177 in a PowerShell script. The corresponding line is ' } while ($RubrikRequestInfo.status -notin $ExitList) '. Correct the script and reschedule the job. The error information returned by PowerShell is: 'You must provide a value expression on the right-hand side of the '-' operator. '. Process Exit Code -1. The step failed. The only way it will run is not passing the json file as a parm. Do you know what I am doing wrong — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSBJYXFMME3V3WYT67DPYBBJFA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4GW2J6QA&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324617384&sdata=2N3GHn3rOKqfuPObP%2FW1Ed0yrjw4ay0lKDk0F3JcK4U%3D&reserved=0>, or mute the thread< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSG45G3PG6LW6J5OJSDPYBBJFANCNFSM4HRL2OWQ&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324627395&sdata=uqrLTp%2FotYCmTVLFSqtB3FSX%2BcloTd1pMw8ANWL4wPs%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUHV5XWNRIQT2RNJELTPYUG3RA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZI3GQ#issuecomment-498240922, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUHVFK67A73U7VFY6ZTPYUG3RANCNFSM4HRL2OWQ . I am setting up a lab to test this and fix another unrelated problem with this script. I will work on both at the same time and provide a solution. From: rwyoung01<EMAIL_ADDRESS>Sent: Monday, June 3, 2019 9:14 AM To: rubrikinc/rubrik-scripts-for-powershell Cc: Chris Lumnah; Comment Subject: Re: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) i have tried both. the first example is with the cmdexe and the second example is with the powershell type. Russell Young On Mon, Jun 3, 2019 at 7:44 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Are you using the Powershell run type or the Operating system run type? From: rwyoung01<EMAIL_ADDRESS>Sent: Thursday, May 30, 2019 5:28 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Subscribed Subject: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) When I use the following script in a SQL Agent job as a cmdshell script type powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json Nothing happens and I get the following message in the job history The step did not generate any output. Process Exit Code 0. The step succeeded. If I trying the script as a Powershell type D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json I get the following message A job step received an error at line 177 in a PowerShell script. The corresponding line is ' } while ($RubrikRequestInfo.status -notin $ExitList) '. Correct the script and reschedule the job. The error information returned by PowerShell is: 'You must provide a value expression on the right-hand side of the '-' operator. '. Process Exit Code -1. The step failed. The only way it will run is not passing the json file as a parm. Do you know what I am doing wrong — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSBJYXFMME3V3WYT67DPYBBJFA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4GW2J6QA&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324617384&sdata=2N3GHn3rOKqfuPObP%2FW1Ed0yrjw4ay0lKDk0F3JcK4U%3D&reserved=0>, or mute the thread< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSG45G3PG6LW6J5OJSDPYBBJFANCNFSM4HRL2OWQ&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324627395&sdata=uqrLTp%2FotYCmTVLFSqtB3FSX%2BcloTd1pMw8ANWL4wPs%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUHV5XWNRIQT2RNJELTPYUG3RA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZI3GQ#issuecomment-498240922, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUHVFK67A73U7VFY6ZTPYUG3RANCNFSM4HRL2OWQ . — You are receiving this because you commented. Reply to this email directly, view it on GitHubhttps://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSGOA4KZ42GUETTJDCLPYUKMPA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZLNPQ%23issuecomment-498251454&data=02|01||d47c3165bee546b463de08d6e82573b5|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636951644880050253&sdata=L4G5TwBI8qLtSfCx2HErfmtGRsc0z9h%2F%2FsUvd0KeyZI%3D&reserved=0, or mute the threadhttps://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSDKKAE6UJC4RNF23ADPYUKMPANCNFSM4HRL2OWQ&data=02|01||d47c3165bee546b463de08d6e82573b5|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636951644880060258&sdata=gHo%2BueqF7rArhb%2B9LhXPJCXcyl%2FIPHh3X1VgQ8jVSOw%3D&reserved=0. I have learned that the Powershell type is running version 2.0 which does not support the 'notin' command because I am running SQL server 2012 R2. The cmdexe version is using powershell 3.0 but doesn't work. Hope this helps. Russell Young On Tue, Jun 4, 2019 at 8:22 AM Chris Lumnah<EMAIL_ADDRESS>wrote: I am setting up a lab to test this and fix another unrelated problem with this script. I will work on both at the same time and provide a solution. From: rwyoung01<EMAIL_ADDRESS>Sent: Monday, June 3, 2019 9:14 AM To: rubrikinc/rubrik-scripts-for-powershell Cc: Chris Lumnah; Comment Subject: Re: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) i have tried both. the first example is with the cmdexe and the second example is with the powershell type. Russell Young On Mon, Jun 3, 2019 at 7:44 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Are you using the Powershell run type or the Operating system run type? From: rwyoung01<EMAIL_ADDRESS>Sent: Thursday, May 30, 2019 5:28 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Subscribed Subject: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) When I use the following script in a SQL Agent job as a cmdshell script type powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json Nothing happens and I get the following message in the job history The step did not generate any output. Process Exit Code 0. The step succeeded. If I trying the script as a Powershell type D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json I get the following message A job step received an error at line 177 in a PowerShell script. The corresponding line is ' } while ($RubrikRequestInfo.status -notin $ExitList) '. Correct the script and reschedule the job. The error information returned by PowerShell is: 'You must provide a value expression on the right-hand side of the '-' operator. '. Process Exit Code -1. The step failed. The only way it will run is not passing the json file as a parm. Do you know what I am doing wrong — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSBJYXFMME3V3WYT67DPYBBJFA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4GW2J6QA&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324617384&sdata=2N3GHn3rOKqfuPObP%2FW1Ed0yrjw4ay0lKDk0F3JcK4U%3D&reserved=0 , or mute the thread< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSG45G3PG6LW6J5OJSDPYBBJFANCNFSM4HRL2OWQ&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324627395&sdata=uqrLTp%2FotYCmTVLFSqtB3FSX%2BcloTd1pMw8ANWL4wPs%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub < https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUHV5XWNRIQT2RNJELTPYUG3RA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZI3GQ#issuecomment-498240922 , or mute the thread < https://github.com/notifications/unsubscribe-auth/AFZFZUHVFK67A73U7VFY6ZTPYUG3RANCNFSM4HRL2OWQ . — You are receiving this because you commented. Reply to this email directly, view it on GitHub< https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSGOA4KZ42GUETTJDCLPYUKMPA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZLNPQ%23issuecomment-498251454&data=02|01||d47c3165bee546b463de08d6e82573b5|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636951644880050253&sdata=L4G5TwBI8qLtSfCx2HErfmtGRsc0z9h%2F%2FsUvd0KeyZI%3D&reserved=0>, or mute the thread< https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSDKKAE6UJC4RNF23ADPYUKMPANCNFSM4HRL2OWQ&data=02|01||d47c3165bee546b463de08d6e82573b5|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636951644880060258&sdata=gHo%2BueqF7rArhb%2B9LhXPJCXcyl%2FIPHh3X1VgQ8jVSOw%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUDYG7S7MOMPTGLXZW3PYZUBRA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODW4ROIY#issuecomment-498669347, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUG7MLWVFFJ37MOGWLTPYZUBRANCNFSM4HRL2OWQ . ok, I moved the script to a SQL Server 2014 and have the SQL Agent job run as a Powershell Type and am passing the json file. So the issue is with SQL 2012 which runs Powershell 2.0. Russell Young On Tue, Jun 4, 2019 at 8:22 AM Chris Lumnah<EMAIL_ADDRESS>wrote: I am setting up a lab to test this and fix another unrelated problem with this script. I will work on both at the same time and provide a solution. From: rwyoung01<EMAIL_ADDRESS>Sent: Monday, June 3, 2019 9:14 AM To: rubrikinc/rubrik-scripts-for-powershell Cc: Chris Lumnah; Comment Subject: Re: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) i have tried both. the first example is with the cmdexe and the second example is with the powershell type. Russell Young On Mon, Jun 3, 2019 at 7:44 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Are you using the Powershell run type or the Operating system run type? From: rwyoung01<EMAIL_ADDRESS>Sent: Thursday, May 30, 2019 5:28 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Subscribed Subject: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) When I use the following script in a SQL Agent job as a cmdshell script type powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json Nothing happens and I get the following message in the job history The step did not generate any output. Process Exit Code 0. The step succeeded. If I trying the script as a Powershell type D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -Jobfile Export-RubrikDatabaseJobFile.json I get the following message A job step received an error at line 177 in a PowerShell script. The corresponding line is ' } while ($RubrikRequestInfo.status -notin $ExitList) '. Correct the script and reschedule the job. The error information returned by PowerShell is: 'You must provide a value expression on the right-hand side of the '-' operator. '. Process Exit Code -1. The step failed. The only way it will run is not passing the json file as a parm. Do you know what I am doing wrong — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSBJYXFMME3V3WYT67DPYBBJFA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4GW2J6QA&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324617384&sdata=2N3GHn3rOKqfuPObP%2FW1Ed0yrjw4ay0lKDk0F3JcK4U%3D&reserved=0 , or mute the thread< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSG45G3PG6LW6J5OJSDPYBBJFANCNFSM4HRL2OWQ&data=02|01||b242996db0c04604043408d6e545cf95|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636948485324627395&sdata=uqrLTp%2FotYCmTVLFSqtB3FSX%2BcloTd1pMw8ANWL4wPs%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub < https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUHV5XWNRIQT2RNJELTPYUG3RA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZI3GQ#issuecomment-498240922 , or mute the thread < https://github.com/notifications/unsubscribe-auth/AFZFZUHVFK67A73U7VFY6ZTPYUG3RANCNFSM4HRL2OWQ . — You are receiving this because you commented. Reply to this email directly, view it on GitHub< https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSGOA4KZ42GUETTJDCLPYUKMPA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWZLNPQ%23issuecomment-498251454&data=02|01||d47c3165bee546b463de08d6e82573b5|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636951644880050253&sdata=L4G5TwBI8qLtSfCx2HErfmtGRsc0z9h%2F%2FsUvd0KeyZI%3D&reserved=0>, or mute the thread< https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSDKKAE6UJC4RNF23ADPYUKMPANCNFSM4HRL2OWQ&data=02|01||d47c3165bee546b463de08d6e82573b5|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636951644880060258&sdata=gHo%2BueqF7rArhb%2B9LhXPJCXcyl%2FIPHh3X1VgQ8jVSOw%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUDYG7S7MOMPTGLXZW3PYZUBRA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODW4ROIY#issuecomment-498669347, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUG7MLWVFFJ37MOGWLTPYZUBRANCNFSM4HRL2OWQ . Please see this article on running powershell with sql agent. https://www.sqlhammer.com/running-powershell-in-a-sql-agent-job/ For SQL Server 2012, i could get the script to run if i do the below as a job step powershell.exe -file E:\Scripts\Export-RubrikDatabasesJob.ps1 I had to modify the script to have this as the param section param( $JobFile = "E:\Scripts\Export-RubrikDatabasesJobFile.json" ) I am going to test the functionality on other versions of SQL Server. I was not able to run it on SQL 2012 because the Powershell type run version 2.0 and your code operator 'notin' is not supporter. But I was able to get it to run as CMDEXE but I had to code it this way powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -jobfile d:\powershell\rubrik\Export-RubrikDatabaseJobFile.json When I didn't specify the path to the json file it didn't do anything. So we can run it on 2012 when we upgrade powershell to 3.0 and run as CMDEXE. Russell Young On Wed, Jun 5, 2019 at 10:07 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Please see this article on running powershell with sql agent. https://www.sqlhammer.com/running-powershell-in-a-sql-agent-job/ For SQL Server 2012, i could get the script to run if i do the below as a job step powershell.exe -file E:\Scripts\Export-RubrikDatabasesJob.ps1 I had to modify the script to have this as the param section param( $JobFile = "E:\Scripts\Export-RubrikDatabasesJobFile.json" ) I am going to test the functionality on other versions of SQL Server. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUDDYEISBTLHX4ZSK2DPY7JCLA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXAANAY#issuecomment-499123843, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUCA5OIJXZSG3M4ZNQLPY7JCLANCNFSM4HRL2OWQ . After some testing, I was able to confirm your way of running things. For the version thing, there is not much I can do there due to the nature of what is supported as you can see from the blog post article. Now that you have the script running, shall I close the issue as resolved? From: rwyoung01<EMAIL_ADDRESS>Sent: Wednesday, June 5, 2019 2:36 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Chris Lumnah; Comment Subject: Re: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) I was not able to run it on SQL 2012 because the Powershell type run version 2.0 and your code operator 'notin' is not supporter. But I was able to get it to run as CMDEXE but I had to code it this way powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -jobfile d:\powershell\rubrik\Export-RubrikDatabaseJobFile.json When I didn't specify the path to the json file it didn't do anything. So we can run it on 2012 when we upgrade powershell to 3.0 and run as CMDEXE. Russell Young On Wed, Jun 5, 2019 at 10:07 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Please see this article on running powershell with sql agent. https://www.sqlhammer.com/running-powershell-in-a-sql-agent-job/ For SQL Server 2012, i could get the script to run if i do the below as a job step powershell.exe -file E:\Scripts\Export-RubrikDatabasesJob.ps1 I had to modify the script to have this as the param section param( $JobFile = "E:\Scripts\Export-RubrikDatabasesJobFile.json" ) I am going to test the functionality on other versions of SQL Server. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUDDYEISBTLHX4ZSK2DPY7JCLA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXAANAY#issuecomment-499123843, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUCA5OIJXZSG3M4ZNQLPY7JCLANCNFSM4HRL2OWQ . — You are receiving this because you commented. Reply to this email directly, view it on GitHubhttps://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSDZZ52Z7LXSXRRPJE3PZABT5A5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXAUCMY%23issuecomment-499204403&data=02|01||5eb3c95f89814f630fbd08d6e9e4c497|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636953566090470044&sdata=oKluql%2FegEdtB1PE7UB0n3M1iCIvkmKD6q%2Foko1Y5SA%3D&reserved=0, or mute the threadhttps://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSFVU2SZURALJ3OGKT3PZABT5ANCNFSM4HRL2OWQ&data=02|01||5eb3c95f89814f630fbd08d6e9e4c497|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636953566090480055&sdata=dqa0%2BFjuUKdyldlU7MdPufGnSdwYiMh6ynQ7FC%2BQHp8%3D&reserved=0. yes, thanks for helping Russell Young On Wed, Jun 5, 2019 at 9:55 PM Chris Lumnah<EMAIL_ADDRESS>wrote: After some testing, I was able to confirm your way of running things. For the version thing, there is not much I can do there due to the nature of what is supported as you can see from the blog post article. Now that you have the script running, shall I close the issue as resolved? From: rwyoung01<EMAIL_ADDRESS>Sent: Wednesday, June 5, 2019 2:36 PM To: rubrikinc/rubrik-scripts-for-powershell Cc: Chris Lumnah; Comment Subject: Re: [rubrikinc/rubrik-scripts-for-powershell] Export-RubrikDatabasejob pass json file in SQL Server Agent Job (#91) I was not able to run it on SQL 2012 because the Powershell type run version 2.0 and your code operator 'notin' is not supporter. But I was able to get it to run as CMDEXE but I had to code it this way powershell.exe D:\Powershell\Rubrik\Export-RubrikDatabasejob.ps1 -jobfile d:\powershell\rubrik\Export-RubrikDatabaseJobFile.json When I didn't specify the path to the json file it didn't do anything. So we can run it on 2012 when we upgrade powershell to 3.0 and run as CMDEXE. Russell Young On Wed, Jun 5, 2019 at 10:07 AM Chris Lumnah<EMAIL_ADDRESS>wrote: Please see this article on running powershell with sql agent. https://www.sqlhammer.com/running-powershell-in-a-sql-agent-job/ For SQL Server 2012, i could get the script to run if i do the below as a job step powershell.exe -file E:\Scripts\Export-RubrikDatabasesJob.ps1 I had to modify the script to have this as the param section param( $JobFile = "E:\Scripts\Export-RubrikDatabasesJobFile.json" ) I am going to test the functionality on other versions of SQL Server. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub < https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUDDYEISBTLHX4ZSK2DPY7JCLA5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXAANAY#issuecomment-499123843 , or mute the thread < https://github.com/notifications/unsubscribe-auth/AFZFZUCA5OIJXZSG3M4ZNQLPY7JCLANCNFSM4HRL2OWQ . — You are receiving this because you commented. Reply to this email directly, view it on GitHub< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frubrikinc%2Frubrik-scripts-for-powershell%2Fissues%2F91%3Femail_source%3Dnotifications%26email_token%3DAFKMBSDZZ52Z7LXSXRRPJE3PZABT5A5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXAUCMY%23issuecomment-499204403&data=02|01||5eb3c95f89814f630fbd08d6e9e4c497|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636953566090470044&sdata=oKluql%2FegEdtB1PE7UB0n3M1iCIvkmKD6q%2Foko1Y5SA%3D&reserved=0>, or mute the thread< https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAFKMBSFVU2SZURALJ3OGKT3PZABT5ANCNFSM4HRL2OWQ&data=02|01||5eb3c95f89814f630fbd08d6e9e4c497|84df9e7fe9f640afb435aaaaaaaaaaaa|1|0|636953566090480055&sdata=dqa0%2BFjuUKdyldlU7MdPufGnSdwYiMh6ynQ7FC%2BQHp8%3D&reserved=0 . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rubrikinc/rubrik-scripts-for-powershell/issues/91?email_source=notifications&email_token=AFZFZUCQC7RYENHKG7EHYQDPZB4D3A5CNFSM4HRL2OW2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXBSQ5Q#issuecomment-499329142, or mute the thread https://github.com/notifications/unsubscribe-auth/AFZFZUGLPQ7BXVTQFY2NAV3PZB4D3ANCNFSM4HRL2OWQ .
2025-04-01T06:40:16.869485
2021-09-09T09:42:50
992015522
{ "authors": [ "dhh", "fxn", "ko1", "rafaelfranca", "st0012" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10367", "repo": "ruby/debug", "url": "https://github.com/ruby/debug/issues/261" }
gharchive/issue
Integrating ruby/debug with Rails to replace byebug through a single "debugger" call Love the progress with this! We'd very much like to make ruby/debug the default debugger in Rails 7. This would replace byebug, which has been the default for 7 years. But to do that, and retain the same developer experience, we're looking for a way to retain a similar interface. That means being able to require ruby/debug either all the time or at least all the time in development, and then jumping into the debugger merely by adding a call to "debugger" anywhere in the code base (alternatively using "binding.break", but I find the "debugger" call more intuitive). Is this approach compatible with the aspirations of ruby/debug? If so, is there anything we can do to help it along, such that we can make the switch for Rails 7? ✌️❤️ That means being able to require ruby/debug either all the time or at least all the time in development, and then jumping into the debugger merely by adding a call to "debugger" anywhere in the code base Just to share my experience here: my project (Rails 6.1) has installed the debugger for 2 weeks and requires it by default in both development and test environment (even when running on CI). It doesn't break any test case, Rails/Rake task. It also doesn't slow the application in any obvious way. And whenever we need to debug, we just use binding.b to enter a debugging session. So I think this is possible already 🙂 However, the current implementation assumes that loading the gem means you want to start a session: % ruby -rdebug -e1 DEBUGGER: Session start (pid: 73510) In a Rails application, ideally you'd require the gem via Bundler as usual, and have binding.break (or whatever) start the session if not already started. I believe this alternative pattern fits better web applications in development. Could be behind a configuration point for backwards compat (provided that is possible). I see. This is the discussion that led to the current design: https://github.com/ruby/debug/issues/65 From my understanding, debugger session records every iseq sequence that's loaded once it's started. https://github.com/ruby/debug/blob/c5917c529b99dee8a7e27ec6abdc2c8e70d4e6f6/lib/debug/session.rb#L86-L89 (which eventually leads to here) https://github.com/ruby/debug/blob/c5917c529b99dee8a7e27ec6abdc2c8e70d4e6f6/lib/debug/session.rb#L1304-L1317 So if the debugger session only starts after binding.b is called, all the code loaded before it would not be stored into the debugger's SourceRepository. And later affect the source the debugger can retrieve https://github.com/ruby/debug/blob/c5917c529b99dee8a7e27ec6abdc2c8e70d4e6f6/lib/debug/session.rb#L265-L271 https://github.com/ruby/debug/blob/c5917c529b99dee8a7e27ec6abdc2c8e70d4e6f6/lib/debug/frame_info.rb#L62-L64 Ah, I just tested this now. It's actually doing the right thing. It doesn't halt execution on require, which is what I thought. So we can just remove require: false from the Gemfile. It does, however, output a noisy "DEBUGGER: Session start (pid: 48334)". Is it possible to turn that off? Also, still wish it was possible to just called "debugger" instead of "binding.b". But that's a minor point. I've turned it on by default for Rails 7 dev/test. So this isn't really a problem. Would be nice if there was a way not to get the output statement, and I'd still love "debugger" as an alias for "binding.b", but this is great in any case. I just investigated this a little bit more. If we require debug by default in the Rails framework, and you use the vscode extension (which will call rdbg) or use rdbg directly you get two debug sessions. The reason for that is that lib.debug.rb starts a session and rdbg will start another one requiring debug/open or debug/start. Would it possible to not start the session using ruby -r debug and always use ruby -r debug/start? and only use debug to providing the biding methods and setup tracepoint? @rafaelfranca sorry I couldn't catch the points. lib/debug.rb should not be used if (1) rdbg is used (2) require 'debug' with Gemfile and bundle exec. But lib/debug.rb can be required if (3) require 'debug' in Ruby 2.7 or before without Gemfile and bundle exec. Could you check the situation and make another issue with your reproduce process? Thanks. Ah, I can repro and make new issue.
2025-04-01T06:40:16.873531
2023-03-29T10:01:40
1645449209
{ "authors": [ "ko1", "st0012" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10368", "repo": "ruby/debug", "url": "https://github.com/ruby/debug/pull/950" }
gharchive/pull-request
Restart threads on DAP evaluation This makes sure DAP evaluation request also doesn't hang. See #947 for the original console implementation. @ono-max I tried implementing the same change for CDP as well but it still hangs. So I will leave that to you. [master]$ git diff diff --git a/lib/debug/server_dap.rb b/lib/debug/server_dap.rb index 8298428..bae1e4d 100644 --- a/lib/debug/server_dap.rb +++ b/lib/debug/server_dap.rb @@ -625,6 +625,7 @@ module DEBUGGER__ expr = req.dig('arguments', 'expression') if find_waiting_tc(tid) + restart_all_threads request_tc [:dap, :evaluate, req, fid, expr, context] else fail_response req @@ -701,6 +702,7 @@ module DEBUGGER__ register_vars result[:variables], tid @ui.respond req, result when :evaluate + stop_all_threads message = result.delete :message if message @ui.respond req, success: false, message: message could you add stop_all_threads? Ah I thought it's handled in Session#process_event too. I've updated code but will try to put up a test case for thread-stopping too. I think I found another issue: when I run the test, the result was not deterministic even without sleep and fails around 10% of the time on my machine. I initially thought it's because of my test, but then I found that it's because this early return isn't always correct. Sometimes it doesn't reflect the actual running thread count and thus the debugger doesn't stop the threads. If I comment it out, the test doesn't fail intermittently anymore. Should I address that in this PR too? Now current behavior is already non-deterministic (by design). Essentially this is hard to write a test for such cases. As the test could make CI unstable, I've dropped it.
2025-04-01T06:40:16.891772
2023-08-04T14:01:57
1836805411
{ "authors": [ "enebo", "eregon", "kddnewton" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10369", "repo": "ruby/yarp", "url": "https://github.com/ruby/yarp/pull/1204" }
gharchive/pull-request
Rename RescueNode#exception to RescueNode#reference Fixes #1160 Okay, let me know when it's ready. I'm closing out for the weekend so going to merge this. I really want to figure this out going forward, otherwise it's impossible to make breaking changes without waiting for y'all to make changes. Maybe we could mark the check as optional. @kddnewton The usual strategy for this is to wait that TruffleRuby adopts the change, and then when that's merged we can merge here and no CI breaks. So if this is not urgent I would wait for that. @eregon @kddnewton I can see the benefits of wanting people to see breaking changes on TR for bug fixes, but for development work this creates a backwards dependency. If I submit a PR here (X) to change something then if we followed your process and I was willing to submit a PR (Y) to your repo first (and let's just pretend that is no work but just something which just happens); then I think this assumes YARP development is some single linear process. So Y is landed, then someone in yarp fixes a different bug with a PR (Z), then we land X. In this case Z is red and has nothing to do with the PR you fixed with Y. That will be confusing to the author of Z. Other scenarios (WIP PRs, language level changes needing githubaction changes depending on language level TR has) but right now this one feels like it would be common. If we adopted the same process for JRuby, then we have to have two consuming repositories do PRs before the PR can land on YARP. That ignores some of YARPs CI will catch things which would be hard to test locally (so it would be nice to see that before two PRs land in other projects). I wonder if there is some other way of achieving the goals here. Correct me if I am wrong @eregon but you want to make sure if someone breaks TR from a fix there is a chance for them to notice that and reject the PR (or at least consult about what happened). I think the main issue is once it's red nobody will notice if there is another incompatible change or a bug slipping in. And so it might be hard to get back to green if that happens (e.g. if Java deserialization is broken and has multiple bugs it quickly becomes challenging to fix it, that already happened before the check was there). I think the existing process of merging in TruffleRuby first and in YARP right after works well for most cases. The main alternative I see is to avoid breaking that check, for instance in this case by keeping the old field or so. It often feels hacky or overhead though or it might not be reasonably possible. I also thought at some point to only do the "can it deserialize cleanly" check (maybe running it on JRuby if it doesn't depend on specific YARP node names and field names). But that's weaker in terms of testing than actually checking it works for a real usage of YARP, i.e., as TruffleRuby uses it. I think so far this usage is the only one that really checks the AST makes sense and has the relevant information. Nothing else in CI really uses YARP nodes for execution yet, e.g., if there was an incorrect node as the receiver of a CallNode, I think nothing would notice it except that check. If we do a bunch of field/node renames then it probably makes sense to group them, and similarly for other related breaking changes, so there is less often the need to merge things in a particular order.
2025-04-01T06:40:16.907250
2019-05-18T04:44:19
445678416
{ "authors": [ "armahillo", "jeduardo824" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10370", "repo": "rubyforgood/diaper", "url": "https://github.com/rubyforgood/diaper/issues/1003" }
gharchive/issue
DistributionsController#create should be cleaned up There are currently 3 different branches in that action that result in render :new. My gut says there is probably some redundancy here that could be reduced. working on this
2025-04-01T06:40:16.909329
2017-10-08T13:39:26
263718563
{ "authors": [ "gabteles", "leesharma" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10371", "repo": "rubyforgood/playtime", "url": "https://github.com/rubyforgood/playtime/pull/116" }
gharchive/pull-request
Developer login Hello, I've enabled development login as needed in issue #73. Now when you signin, you should see this screen: And you can choose which authentication to use in development by using the environment variable FORCE_AMAZON_LOGIN. (true/false) It disables csrf protection only in development mode and to the developer auth provider. Thanks for the PR! :tada: I'll take a look at it tomorrow. Hey @leesharma, did it! :D Awesome, thanks! 👍
2025-04-01T06:40:16.977667
2022-05-09T20:50:02
1230237338
{ "authors": [ "nerg4l", "rueian" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10372", "repo": "rueian/rueidis", "url": "https://github.com/rueian/rueidis/issues/34" }
gharchive/issue
Cannot auth with password only Redis can be configured with requirepass which makes AUTH command accept a password without username. Currently, there is no option for this. AUTH command can only initiated with username and password. Source: https://redis.io/commands/auth/ https://github.com/rueian/rueidis/blob/a525ad43300a4ea316a8ca1355a46752ee8be8ea/pipe.go#L73-L79 Hi @nerg4l, Thank you for pointing it out. In the new v0.0.45, it will use the default username if no username but password provided according to https://redis.io/commands/hello/ Thanks
2025-04-01T06:40:17.022235
2022-06-08T14:08:07
1264813393
{ "authors": [ "hqhtiantian520", "ruitian-olivia" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10373", "repo": "ruitian-olivia/STIC-model", "url": "https://github.com/ruitian-olivia/STIC-model/issues/3" }
gharchive/issue
question for image selected in every individual Dear author, Can you explain specifically which cross-section image each individual selected comes from, or was it selected manually? Your article is great, but I would like to know how to select CT cross-sections for specific individual images. Can you provide an image example? Hello, thank you for your question. For each individual, 2D axial slices with the largest tumor area along the z-axis from 3D CT images were selected as the input of diagnostic model. thanks for your response! .could you send me sample clinical data and i am trying to reproduce your paper. Here is my<EMAIL_ADDRESS>
2025-04-01T06:40:17.039072
2023-10-20T08:40:29
1953815124
{ "authors": [ "ppcololo", "runabol" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10374", "repo": "runabol/tork-web", "url": "https://github.com/runabol/tork-web/issues/11" }
gharchive/issue
[Bug] Wrong Runtime if task failed Hello, faced interesting issue: state: RUNNING createdAt: 2023-10-20T01:19:56.420301Z scheduledAt: 2023-10-20T01:19:56.424951Z startedAt: 2023-10-20T07:12:15.879177Z failedAt: 2023-10-20T01:19:56.252516Z So, if task was failed we can see incorrect Runtime in tork-web Will be nice to see some Fail counter near the task State When it completed time is correct: @ppcololo this should be fixed. could you confirm? I don't know what did you fix, but I still see this issue: coordinator: Can you try again using the latest release? I will try, thanks. But can't reproduce it right now. If I have an issue then write here createdAt: 2023-11-11T04:32:06.689706Z scheduledAt: 2023-11-11T04:32:06.700287Z startedAt: 2023-11-11T05:37:18.931082Z failedAt: 2023-11-11T04:32:06.675842Z``` New data: no time at all for running task state: RUNNING createdAt: 2023-11-15T11:46:35.067896Z scheduledAt: 2023-11-15T11:46:35.091199Z startedAt: 2023-11-15T11:46:35.214332Z failedAt: 2023-11-15T11:46:35.0151Z for completed state: COMPLETED createdAt: 2023-11-13T20:33:12.05211Z scheduledAt: 2023-11-13T20:36:03.056434Z startedAt: 2023-11-15T03:16:02.248571Z completedAt: 2023-11-15T03:16:03.156508Z Can you try again using the latest release? I didn't see any issues closing
2025-04-01T06:40:17.090078
2021-09-11T18:55:01
993913686
{ "authors": [ "Hydrox6", "MyopicHuman", "Whowhos" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10375", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/14127" }
gharchive/issue
Hi, could we PLEASE get a larger health bar? The current one is too short The problem is that I like to PvP and knowing what your health is, as exactly as possible, is important. This is why I would LOVE to have a longer health bar, so it's easier to see how much hp I have Numbers over the current health bar would also be amazing Thank you! You're welcome The healthbar above the player is limited, since the values used for it aren't representative of the actual health values; this is a Jagex limitation. A wider health bar would not make it more granular. Status Bars or the hp orb are much better ways of showing this.
2025-04-01T06:40:17.091033
2021-11-28T17:08:57
1065385000
{ "authors": [ "Adam-", "Jin-Jiyunsun" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10376", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/14424" }
gharchive/issue
Add combined option "Europe" to quick-hop region If you are based in Europe, the UK and Germany worlds usually have very similar ping, being able to utilize a larger pool of worlds for quick hopping while maintaining a decent ping would be handy. added in f6a3d222b4986ab7beab54b650797f2821f5852c
2025-04-01T06:40:17.093556
2018-05-04T00:26:49
320122522
{ "authors": [ "Adam-", "Eadgars-Ruse", "Jalopyy" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10377", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/2318" }
gharchive/issue
Hot/Cold Clue Solver https://twitter.com/bitterkoekjers/status/840219782372839424 I feel this would be hard to implement, but if done would be amazing. If you knew the exact ranges of what each temperature meant for how close you are to the clue (not sure if that is documented anywhere) then you could effectively have a plugin that narrows down the locations for you over time. I've got a working solution for this. I need to clean up my code and work on the UI a bit, but I should be able to do a PR soon. Here's some teasers: https://streamable.com/370tj https://streamable.com/hrgrq https://streamable.com/gb8bn Done in #2412
2025-04-01T06:40:17.095009
2019-02-02T03:07:21
405941443
{ "authors": [ "deathbeam", "michaelcubel" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10378", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/7661" }
gharchive/issue
"Reset others" in loot tab Is your feature request related to a problem? Please describe. Currently have to reset all manually when I want to track a single monsters loot after I have looted a few e.g. slayer tasks. Describe the solution you'd like A "Reset others" option like in the exp tracker. Its not that hard to simply click reset multiple times
2025-04-01T06:40:17.096061
2018-04-07T00:01:20
312152025
{ "authors": [ "deathbeam", "mrpker9", "sethtroll" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10379", "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/pull/1301" }
gharchive/pull-request
player indicators: Add clan caller adds a clan caller option to the player indicators which will show clan caller name as well as show clan callers target name Any news on this? My comments still apply, the options should be renamed to just highlight etc to be more universal
2025-04-01T06:40:17.116995
2017-09-04T19:36:52
255107090
{ "authors": [ "d3zd3z", "utzig" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10380", "repo": "runtimeco/mcuboot", "url": "https://github.com/runtimeco/mcuboot/pull/115" }
gharchive/pull-request
Add travis' validation of "Signed-off-by:" line This adds validation of "Signed-off-by:" line. Signed-off-by: Fabio Utzig<EMAIL_ADDRESS> Regarding the comment suggesting to add commiter checking as well as author: [utzig@inspiron mcuboot]$ git show -s --format="%cn <%ce>" cb1bb48 David Brown<EMAIL_ADDRESS>[utzig@inspiron mcuboot]$ git show -s --format="%an <%ae>" cb1bb48 David Brown<EMAIL_ADDRESS> I'm not sure it would be a great idea at the moment. I would prefer to work on it later in a new patch (if at all). You should only ever see the davidb.org commit after github rebases it. It is using my primary account for the rebase, as far as it is concerned. You should always be ok adding the committer check, and when added, it will usually be after the author Sob. Added check for commiter as well, let's see how it goes.
2025-04-01T06:40:17.120571
2024-05-25T10:51:56
2316926294
{ "authors": [ "bjivanovich", "rupeshs" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10381", "repo": "rupeshs/fastsdcpu", "url": "https://github.com/rupeshs/fastsdcpu/issues/190" }
gharchive/issue
fastsdcpu offline? I want to know if there's a way to use fastsdcpu offline. How can i use the models I have on A1111/models/StableDiffusion folder on my PC I will convert the models into OpenVINO models follown de scrits. Should I modify stable-diffusion-models.txt or settings.yaml? The models are located in D:/A1111/models/StableDiffusion I'm new on Stable Diffusion, need help and step by step instruction @bjivanovich Thanks for using FastSD CPU, enable this setting Checkout this https://github.com/rupeshs/fastsdcpu?tab=readme-ov-file#models @bjivanovich LCM-LoRA offline tutorial https://www.youtube.com/watch?v=T8kZL5l3K8c
2025-04-01T06:40:17.141138
2016-02-17T13:10:32
134275399
{ "authors": [ "novocodev", "ruslo" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10382", "repo": "ruslo/polly", "url": "https://github.com/ruslo/polly/issues/66" }
gharchive/issue
Setting public headers for Xcode framework target I have built the 02-library example for Xcode with the following command: build.py --clear --toolchain xcode --framework This built a Framework, but it does not have any public headers in the framework Headers folder. I added a public.h file to the example and added the following entry in the CMakeLists.txt set_target_properties( foo PROPERTIES PUBLIC_HEADER public.h) When building the public.h file is copied to '_install/xcode/lib/' But the build fails with the error: Expected only one lib in directory: .../_install/xcode/lib But found: ['.../_install/xcode/lib/libfood.a', '.../_install/xcode/lib/public.h'] What is the correct way to configure the public headers for a framework build. Try to install headers to include directory. Explanation: this is workaround script which expect only one file in lib folder (which should be the *.a library) so when there are some other files script can't decide what file is a library. Just for your information there are few fixes in latest CMake version that allow creating frameworks without hacks/workarounds (just using CMake). I haven't tried it yet, but if it works okay --framework option will be removed. Still using make 3.3 cmake_minimum_required(VERSION 3.3) I added the FRAMEWORK property set_target_properties( foo PROPERTIES FRAMEWORK TRUE PUBLIC_HEADER public.h) Then I built using the following command ./buildtools/polly/bin/build.py --clear --toolchain xcode This created a Framework with 'public.h' in the Headers folder located at: _builds/xcode/Debug/foo.framework But this only appears to be a MAC OS X framework not a fat Framework, I ran the file command: file _builds/xcode/Debug/foo.framework/foo result was: Mach-O 64-bit dynamically linked shared library x86_64 I also tried --toolchain ios-9-2 but the framework still appears to be created only for x86_64 But this only appears to be a MAC OS X framework not a fat Framework This is by design. Fat library created only for iOS. I also tried --toolchain ios-9-2 but the framework still appears to be created only for x86_64 I will test latest CMake version with few improvements, may be will create an example. What is the correct way to configure the public headers for a framework build Okay, I think I know what is the problem here. --framework option expect headers be located in directory <install-prefix>/include/<frameworkname>. I.e. if you have library libfoo.dylib installed, then headers should be located in <install-prefix>/include/foo and will be moved to foo.framework/Headers. The reason of this is that project which uses foo should do #include <foo/*.hpp> in both variants. I've added warning to build.py script, so you should see next message: Warning: no headers found for framework (dir: /.../_install/ios-9-1-armv7/include/boo) Which can be fixed by next CMake code: install(FILES boo.hpp DESTINATION include/boo) I also tried --toolchain ios-9-2 but the framework still appears to be created only for x86_64 Are you sure you're using patched CMake version? https://github.com/ruslo/polly/wiki/Toolchain-list#ios https://github.com/ruslo/hunter#notes-about-version-of-cmake Example with framework and native Xcode project updated: https://github.com/forexample/ios-dynamic-framework#headers Great! I have successfully built and linked the ios-dynamic-framework project. I did have to use patched CMake version, I tried with the latest 3.5.0-rc2 CMake but that failed to build the Fat framework bundle. I'm going to try and replicate the project setup to my own library repo. but that failed to build the Fat framework bundle With CMake 3.5+ you have to use CMAKE_IOS_INSTALL_COMBINED=YES and CMAKE_XCODE_ATTRIBUTE_ONLY_ACTIVE_ARCH=NO and run install (no need to use --framework, now it's --install). However dynamic framework will not be signed properly so the last time I've tried it I failed to run bundle on real device. I have replicated the demo project setup on my own repo, I did not copy over the custom jenkins.py as I am just trying to build a dynamic library component, I am calling build.py directly with: ./buildtools/polly/bin/build.py --clear --toolchain ios-9-2 --framework --config Release The build is successful and the lib binary now reports as a Fat lib and the public header is copied over to the framework/Headers folder. I have noticed that if I build with --config Debug, that 'd' gets appended to the project name for all folders and files and it also causes the public header not to get copied to framework/Headers (unless I install it to an include/'d' folder. Is there a way to prevent the 'd' suffix getting added and maybe create the framework under a debug or release folder depending on the --config value. Is there a way to prevent the 'd' suffix getting added Set it in command line -DCMAKE_DEBUG_POSTFIX="", in terms of build.py script it will be --fwd CMAKE_DEBUG_POSTFIX="" Both iOS and OSX framework builds are working, thanks.
2025-04-01T06:40:17.145015
2016-09-09T16:18:48
176049513
{ "authors": [ "matthewkmayer" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10383", "repo": "rusoto/rusoto", "url": "https://github.com/rusoto/rusoto/pull/375" }
gharchive/pull-request
Adding ec2 test autogen Builds on https://github.com/rusoto/rusoto/pull/373 and will need a rebase after that is merged. Found a few bugs! test result: FAILED. 75 passed; 15 failed; 0 ignored; 0 measured Trying to figure out how we should handle cases where the call output is a shape wrapped in a shape where the top shape is only used for the top level xml tag. Example test_parse_ec_2_request_spot_instances: <RequestSpotInstancesResponse xmlns="http://ec2.amazonaws.com/doc/2014-06-15/"> <spotInstanceRequestSet> ... </spotInstanceRequestSet> </RequestSpotInstancesResponse> The output shape looks like this: "RequestSpotInstancesResult":{ "type":"structure", "members":{ "SpotInstanceRequests":{ "shape":"SpotInstanceRequestList", "documentation":"<p>One or more Spot instance requests.</p>", "locationName":"spotInstanceRequestSet" } }, We want to generate code that ensures the top level is there then immediately calls the child shape deserializer. After working on this more I see while the tests pass nothing is correctly populated in the output object. Needs a bit deeper of a dive than expected. This is off in the weeds. I'm extracting a better solution of trimming code we generate by not populating serializers and deserializers if we know we don't need them. Then back to extracting the good bits from this into smaller scoped PRs. Closing this, got the work extracted from this mess.
2025-04-01T06:40:17.152998
2017-11-17T00:59:26
274717120
{ "authors": [ "SecurityInsanity", "bluejekyll", "matthewkmayer" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10384", "repo": "rusoto/rusoto", "url": "https://github.com/rusoto/rusoto/pull/871" }
gharchive/pull-request
add metadata to S3 put/copy/create_multipart The metadata field was already available on the S3 request objects, but it was not being serialized into the request headers. This fixes that. This automatically prepends the x-amz-meta- to all keys in the metadata hashmap. I haven't done anything on the get object side, which appears to properly have the z-amz-meta- stripped from the keys in the GetObjectOutput metadata field. I believe it is correct, and I've tested with my own setup. The failure appears to be transient: Too many open files (os error 24) Thanks for the PR! I've restarted the one Travis job that failed - hopefully it'll work this time. 😄 @bluejekyll - Have you run the s3 integration tests that we don't run on CI? If not I'd be happy to run them for you, but the code itself looks fine so running them is really just to make sure. I didn’t realize there were tests not run by CI, if you point me to the doc, or give me the command I’ll happily post the results. I can't find the docs on running integration tests, so I'll make sure that gets documented in the repo. To get going, going into the integration_tests folder and running cargo test --features all runs all integration tests. To run just S3, run cargo test --features s3. The OSX build has failed for two different reasons so far that appear to be transient so I'm restarting the failing one again. This looks good to me, thanks! Could you add an integration test to the S3 tests demonstrating this new behavior? The test goes in this file: https://github.com/rusoto/rusoto/blob/master/integration_tests/tests/s3.rs . The OSX build issues should be reduced or eliminated from https://github.com/rusoto/rusoto/pull/875 which is now in master. 😄 I'll work on the integration test today. Thanks! Some things I noticed while testing: unicode in the keys is not supported, if anyone wants that it will need some form of encoding unicode is also a problem in the values, in so far as unicode characters are encoded, but not decoded on the Output objects. Those seem like more general issues with the way headers are treated in the Rusoto library, and I'm not explicitly handling those cases in this PR, but it's something we may want to enhance in the future. Anyway, I've added GetObject and HeadObject tests for S3 to validate the metadata section. Let me know if there's anything else you'd like to see. @matthewkmayer let me know if anything else is needed for this change, Thanks! Planning on getting to it tonight. 👍 All good, thanks! Thank you!
2025-04-01T06:40:17.157383
2023-11-14T18:20:13
1993303049
{ "authors": [ "gwenn", "max-sixty" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10385", "repo": "rusqlite/rusqlite", "url": "https://github.com/rusqlite/rusqlite/issues/1415" }
gharchive/issue
Change in float behavior in 0.30.0 Hi there! Thanks for the project, it's awesome. We saw a change in float behavior from 0.29 to 0.30: https://github.com/PRQL/prql/actions/runs/6864485300/job/18666598362?pr=3797#step:13:101, from https://github.com/PRQL/prql/pull/3797 We test PRQL across a number of databases, and rusqlite==0.29.0 matched other databases on that query, while 0.30.0 doesn't. Over at PRQL we don't have particularly strong preferences about exact float behavior, though avoiding special-casing specific databases is good, and I thought it might be helpful for you to know about the change. Thanks! I am not sure but it may be related to this same regression in an Java / JDBC SQLite driver here: https://github.com/xerial/sqlite-jdbc/commit/8880c338290e6505bc85eec96917552f027a99bd Since version 3.43.0 the sqlite source doesn't round the value anymore, but returned the value including the 8-byte IEEE floating point number inaccuracies. Somehow different operating systems behave differently. On SQLite forum, only 32-bit platform should be impacted ? https://sqlite.org/forum/forumpost/b9511aa180bd1f4875ff20cf7f9953a216f5fe8092e95237321c39a0884ed930
2025-04-01T06:40:17.165939
2020-08-16T11:01:52
679734174
{ "authors": [ "bjorn3", "samhedin" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10386", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/issues/5771" }
gharchive/issue
Hover details show unexpected crate name/path Hey! I hope that this hasn't been posted before. I searched and could not find anything. I expect the first line of the hover details to show std::collections::HashMap. Instead, it shows std::collections::hash::map. std::collections::hash::map::HashMap is the full path of HashMap. std::collections::HashMap is just a re-export. https://github.com/rust-lang/rust/blob/de32266a1780aa4ef748ce7f6200a1554fad0aca/library/std/src/collections/hash/map.rs#L200 https://github.com/rust-lang/rust/blob/de32266a1780aa4ef748ce7f6200a1554fad0aca/library/std/src/collections/mod.rs#L425 I see - I figured that something like this would be the case. Is there a way to get the re-exported path instead? Rustc has the same problem with error messages: https://github.com/rust-lang/rust/issues/21934 I see. I understand that this isn't a priority for my people. I'll keep an eye on it though. Thank you :100:
2025-04-01T06:40:17.167601
2021-10-25T00:25:51
1034555515
{ "authors": [ "lnicola", "terrynsun" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10387", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/pull/10629" }
gharchive/pull-request
feat: Add assist for replacing turbofish with explicit type. Converts ::<_> to an explicit type assignment. let args = args.collect::<Vec<String>>(); -> let args: Vec<String> = args.collect(); Closes #10285
2025-04-01T06:40:17.168885
2021-04-07T15:50:47
852549806
{ "authors": [ "flodiebold" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10388", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/pull/8402" }
gharchive/pull-request
Remove Ty::substs{_mut} Almost all uses actually only care about ADT substs, so it's better to be explicit. The methods were a bad abstraction anyway since they already didn't include the inner types of e.g. TyKind::Ref anymore. bors r+ changelog skip bors r+
2025-04-01T06:40:17.171460
2021-07-03T00:01:19
936123499
{ "authors": [ "Veykril", "flodiebold" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:10389", "repo": "rust-analyzer/rust-analyzer", "url": "https://github.com/rust-analyzer/rust-analyzer/pull/9468" }
gharchive/pull-request
feat: Make inline_function work on methods Now called inline_call. There is a lot of improvement potential for the assist here still. For one better handling of the self param, which currently always emits ref (mut) for &(mut )self params even if not necessary. A problem we currently have is that function parameters can have coercions happen, which this assist will lose as it doesnt emit type ascriptions for the let statements. And there is the general potential for just not even generating some of the let statements and inline the expressions instead depending on the expression(I'll work on this next). bors r+ We should start recording adjustments (for coercions and self parameter autoref / deref) in type inference sooner rather than later so we can handle these cases more correctly! Definitely, I think this is the third issue/feature which wants coercions tracked. I'll open a new issue for this one to track them together.