added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:38:57.320937
2023-05-26T14:19:36
1727726017
{ "authors": [ "lydiapuric" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6770", "repo": "hlxsites/moleculardevices", "url": "https://github.com/hlxsites/moleculardevices/pull/445" }
gharchive/pull-request
Missing wave Please always provide the GitHub issue(s) your PR is for, as well as test URLs where your change can be observed (before and after): Fix #420 Test URLs: Before: https://main--moleculardevices--hlxsites.hlx.page/en/assets/tutorials-videos/br/automated-plate-loading After: https://missing-wave--moleculardevices--hlxsites.hlx.page/en/assets/tutorials-videos/br/automated-plate-loading AND Before: https://main--moleculardevices--hlxsites.hlx.page/en/assets/tutorials-videos/br/how-to-set-up-elisa-assay-and-basic-analysis-using-softmax-pro After: https://missing-wave--moleculardevices--hlxsites.hlx.page/en/assets/tutorials-videos/br/how-to-set-up-elisa-assay-and-basic-analysis-using-softmax-pro AND Before: https://main--moleculardevices--hlxsites.hlx.page/resources/citations/axon After: https://missing-wave--moleculardevices--hlxsites.hlx.page/resources/citations/axon Note: I added also 2 additional styles for videos-and-webinars theme to make sure there is some space between hero and some space to wave
2025-04-01T06:38:57.420273
2019-05-01T10:33:19
439125059
{ "authors": [ "ganganna-peddapothu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6771", "repo": "hmcts/ccd-definition-store-api", "url": "https://github.com/hmcts/ccd-definition-store-api/pull/393" }
gharchive/pull-request
Validate NOT and OR syntax on Definition store to allow user to import definition file https://tools.hmcts.net/jira/browse/RDM-4635 Include validation in definition store to allow definition (excel) file with NOT and OR show condition syntax Does this PR introduce a breaking change? (check one with "x") [ ] Yes [ X] No Added tests.
2025-04-01T06:38:57.426321
2024-03-08T14:05:45
2176111545
{ "authors": [ "miguelMolina3691" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6772", "repo": "hmcts/civil-service", "url": "https://github.com/hmcts/civil-service/pull/4178" }
gharchive/pull-request
CIV-12777 Part Admit Claimant Dashboard JIRA link (if applicable) https://tools.hmcts.net/jira/browse/CIV-12777 Change description Added: Handler for claimant response (claimant) Script for Part admit Accepted claimant response Method to return ordinal formatted date Unit and integration test Modified: I have extracted the utils ClaimantResponseUtils.java from civil/service/docmosis/utils to civil/utils because it is generally used. Consequently some imports have been modified Does this PR introduce a breaking change? (check one with "x") [ ] Yes [ x ] No Open for review before the end of the sprint Open to get reviews
2025-04-01T06:38:57.441963
2024-11-25T15:00:11
2691069767
{ "authors": [ "MartyFox" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6773", "repo": "hmcts/dtspo-daily-monitoring", "url": "https://github.com/hmcts/dtspo-daily-monitoring/pull/95" }
gharchive/pull-request
Updating Postgres checks, attempting to use pipeline variables to send messages to threads with each loops in pipeline Change description Updating Postgres checks Attempting to use pipeline variables to send messages to threads with each loops in pipeline. The first loop sets the header, subsequent loops should send to the thread of the same header example thread:
2025-04-01T06:38:57.487930
2022-11-07T19:20:20
1438855313
{ "authors": [ "hnaderi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6774", "repo": "hnaderi/lepus", "url": "https://github.com/hnaderi/lepus/pull/38" }
gharchive/pull-request
Update scala3-library to 3.2.1 Updates org.scala-lang:scala3-library from 3.2.0 to 3.2.1. GitHub Release Notes - Version Diff - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scala-lang", artifactId = "scala3-library" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "@monthly" }, dependency = { groupId = "org.scala-lang", artifactId = "scala3-library" } }] labels: library-update, early-semver-patch, semver-spec-patch, commit-count:n:2 Superseded by #56.
2025-04-01T06:38:57.490249
2020-06-07T17:20:49
633593613
{ "authors": [ "Idadelveloper", "Idss15", "Oluwatobiloba98", "Technelly", "adekniyi", "dnor-dev" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6775", "repo": "hng-teamfury/stage2", "url": "https://github.com/hng-teamfury/stage2/issues/6" }
gharchive/issue
Intern Selection Process We should start with the grouping so we can know who and who are working together Please what's supposed to go on here? What I would mention here is that the leaders of the group should give room for at least 50% newbies to learn from the experienced. hi, what is the selection process?, what anbout mobile development beginners?, or any beginner? There should be a mode for selection Check the slack channel the selection process is already stated there Really?
2025-04-01T06:38:57.503117
2024-07-21T11:47:11
2421363973
{ "authors": [ "NgBlaze", "Ruthiejayjay" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6776", "repo": "hngprojects/hng_boilerplate_php_laravel_web", "url": "https://github.com/hngprojects/hng_boilerplate_php_laravel_web/pull/67" }
gharchive/pull-request
Squeeze page Contributions Description Developed the API endpoint for the Squeeze Page to handle POST requests, collect user information, and send an email template to the user. Endpoints Created Endpoint: POST /api/v1/squeeze Type of Change [x] New feature (non-breaking change which adds functionality) [ ] Bug fix (non-breaking change which fixes an issue) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] This change requires a documentation update Implementation Details Created Endpoint: Implemented the POST /api/v1/squeeze endpoint to handle form submissions. Data Validation: Added validation for incoming data to ensure required fields are provided and the email format is correct. Response Handling: Implemented appropriate responses for both success and error scenarios: Success Response:{ "message": "Your request has been received. You will get a template shortly." } Error Response:{ "message": "Failed to submit your request", "status_code": 400 } Database Storage: Saved user information securely in the database. Testing Unit Tests: Written unit tests for backend validation and data handling. Testing Scenarios: Tested various scenarios for form submissions including successful, invalid data, and server errors. Challenges and Issues Database Table Issue: Problem: The database table provided by the DB maintainers lacked an email field when I forked and cloned the repo. Solution: Updated the table and migrations on my separate branch to include the necessary email field. Note: This issue should be addressed and noted during testing. Email Template Sending Bottleneck: Problem: The task required sending email templates to the user's email address, but emails could not be sent immediately after the POST request. This necessitated asynchronous processing. Solution: Utilized Laravel's job queue system to handle email sending asynchronously. However, the existing jobs table in the database was already configured for different purposes, such as job type, job name, and organization details. Issue: The email functionality does not work due to the jobs table configuration conflicting with the intended use for email queue processing. Checklist [x] My code follows the style guidelines of this project [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have made corresponding changes to the documentation [x] My changes generate no new warnings [x] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes You have merge conflicts and it doesn't link to your issue You have merge conflicts and it doesn't link to your issue I did not even touch that, maybe it's because i did not pull before pushing I worked on my create branch all through
2025-04-01T06:38:57.509325
2019-11-27T22:11:11
529594309
{ "authors": [ "ebauman", "hall", "wjimenez5271" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6777", "repo": "hobbyfarm/admin-ui", "url": "https://github.com/hobbyfarm/admin-ui/issues/24" }
gharchive/issue
Admin UI sometimes won't save Scenario step edits Not really sure how to reporduce at this time, the behavior seems to be that after some time of using the UI to edit scenario steps, it stops saving them. The green confirmation message won't show up after you click save and when you go back to the scenario later the changes are missing How long of time are we talking? The JWT token expires after a day. Check the response to the PUT requests in the network tab of your browser; like so: in the span of 10s of minutes, definitely not a day Its also intermittent, some moments later the save action will work again I haven't been able to reproduce this behavior. Can you post some more details about which versions of gargantua and admin-ui you're using and the environment in which you installed the app? Also post the HTTP response to a failed request and/or the corresponding logs when the edit fails to save. after further debugging, I suspect it might be because of latin charset in my scenario steps...specifically curly quotes vs straight. The error messages can be seen on the JS console, but the UI doesn't report the error, just silently fails. Closing in favor of #25
2025-04-01T06:38:57.571094
2017-11-16T22:48:43
274691554
{ "authors": [ "codecov-io", "holmgr" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6778", "repo": "holmgr/gemini", "url": "https://github.com/holmgr/gemini/pull/21" }
gharchive/pull-request
Add log and env_log packages, resolves #20 Description This PR introduces the log and env_logger packages to handle logging for the source code. These packages adds several useful macros such as info!, debug!, err! etc which can be used for logging of different kinds. To view the logs one must run test or the main program with the RUST_LOG flag set to the appropriate level, for instance RUST_LOG=gemini=info cargo run. Alternate Designs N/A Benefits Allows logging for debug etc without disturbing the user. Possible Drawbacks N/A Applicable Issues #20 Codecov Report Merging #21 into master will decrease coverage by 3.08%. The diff coverage is 50%. @@ Coverage Diff @@ ## master #21 +/- ## ========================================== - Coverage 98.87% 95.78% -3.09% ========================================== Files 3 3 Lines 89 95 +6 ========================================== + Hits 88 91 +3 - Misses 1 4 +3 Impacted Files Coverage Δ src/main.rs 100% <ø> (ø) :arrow_up: src/resources.rs 100% <100%> (ø) :arrow_up: src/generators/names.rs 94.66% <40%> (-3.91%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 29f1ec8...14af326. Read the comment docs.
2025-04-01T06:38:57.597527
2023-06-13T10:12:09
1754530146
{ "authors": [ "Hoxbro", "MarcSkovMadsen", "droumis", "ianthomas23", "jbednar", "maximlt", "richlysakowski" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6779", "repo": "holoviz/holoviz", "url": "https://github.com/holoviz/holoviz/issues/366" }
gharchive/issue
HoloViz Roadmap 2023 This issue is meant to be commented by the HoloViz contributors, maintainers and steering committee members. In order to define the HoloViz roadmap for the next 6-12 months we would like to collect what would be your personal roadmap, i.e. what you think should be the most important goals for HoloViz and what you would like to work on? Please insert only 1 comment (edit it later if you need to) before the 3rd of July 2023. No matter which website or repo a user has landed on, guide them quickly and easily to the right tool for their job, i.e. the right library and the right API for that library. Will be discussed July 3rd Put HoloViz more to the front: making it clear on each project website that they are part of HoloViz, linking more to holoviz.org, improving the holoviz.org story of what can be done with the intersection of the HoloViz packages (panel+hvplot/geo/holoviews+datashader(+bokeh)), improving the interoperability between the HoloViz packages (aka viz dashboards made easy). Following the work on Panel, keeping on improving the docs by completing Panel's work (API docs) and proceeding to hvPlot's overhaul On a personal level, I'd like to spend a lot more time maintaining hvPlot (which fits with 1. and 2.) Improve the structure of Panel's technical reference material (primarily the API reference and component gallery sections-#4305) Improve guidance on tool selection as part of the getting started experience on HoloViz.org (#351) Reduce energy barriers to first-time contributions. For instance, improving/coordinating contributor guides or getting the complete docs build time down to seconds. IMO, dramatically prioritizing this point would benefit all aspects of HoloViz. Start applying the Panel docs format to hvPlot and HoloViews Datashader: Complete the matrix of support for all Reduction classes and as many of the following as make sense: work on GPU, work using Dask, antialiasing, contained within a by reduction, contained within a where reduction. Ref https://datashader.org/api.html#reductions Simplify the Reduction code which currently has many very similar functions in append and combine functions. This should reduce the code complexity and the number of numba functions. May be able to cache the numba functions. Improve the documentation is line with recent changes to Panel, in particular by adding smaller How To Guides for specific tasks. HoloViews: Code: Keep on improving the code. I would like to look at the code in-depth and make plans to remove some part that is not tested or is no longer used. Test: Update the test to use pytest with fixtures, not unittest.TestCase. Test: Support running the tests in random order so it is possible to parallelize the tests. Test: To improve or at least better understand how to change the plotting backend in the test suite. It could be related to the previous point. Type hints: Begin adding type hints to HoloViews. DEV: Update development infrastructure to use Hatch. I want to begin this transition as soon as possible. DEV: Look into CI's build, docs, and tests concerning the warnings emitted but often ignored. So, in general, improvements behind the scene. I have been following HoloViz and its component libraries for several years and was originally drawn back to use it, specifically because of the Panel Tabulator component. I need a FREE OSS highly-functional, high-performance Python-interactive table widget for Pandas to replace QGrid. I liked Qgrid because it was highly functional, fast & big data scalable for 1M+ rows, preserved state after user interaction, was fully addressable as a Python object and many other features (sort columns, filters, groupby and many other Pandas). Qgrid code ran in Flask, Jupyter Notebook and Django and other Python full-stack web frameworks. QGrid was developed by Quantopian on top of SlickGrid. However, Quantopian ran out of money and closed down; eventually Quantopian followers gave up maintaining Qgrid. The R community has had HTMLtable for many years, offering high-quality, high-functionality table widgets "that just work". Aggrid is a popular proprietary commercial alternative to Qgrid that is being heavily promoted by Streamlit (now a Snowflake property). Streamlit is freemium software. Aggrid is expensive (Starting at...$$$999 Per Developer). Aggrid is still not very mature v0. for the open-source version. Aggrid is a Javascript / Angular framework, very functional. Many features have not been surfaced to Python. The most important Aggrid features (row selection, groupby, and others) are locked to monthly or enterprise subscriptions. The Streamlit community promotes Aggrid heavily through YouTube demos. You can see more about the JavaScript hacks required to add free functionality to aggrid here: Streamlit Ag-Grid "Since AgGrid is a JavaScript-based grid, the background properties of columns or cells can be changed by adding JavaScript code to the Python code. I am adding the JsCode library to be able to use JS code in the Grid." If Panel Tabulator offers a FOSS full-function alternative to Aggrid, it will lead to much greater adoption of the rest of Holoviz. My vote is to focus on enhancing Tabulator to be a good strong FREE alternative ("competitor") to aggrid, and get fully functional as a simple-to-use component in Jupyter Notebook, Flask, Django and other tools. With demos like PyGWalker from Kanaries. PyGWalker is pretty cool. Tabulator with other Holoviz widgets could equal or surpass PyGWalker. I am willing to help with applications of Tabulator, but we have to get it wrapped into a Streamlit component to jump on the Streamlit bandwagon. I found a Chinese port of Streamlit Component wrapper for Tabulator but it does not work completely. I will post a link to it as soon as I re-find it. For me the most important is removing the friction of getting started with Panel, hvPlot, param, HoloViews, ... in that order. It is primarily about improving Documentation Fixing bugs Support on Discourse and Discord Communication on social media. Speed/ Performance
2025-04-01T06:38:57.601551
2023-05-26T17:28:42
1727992227
{ "authors": [ "nicolaskruchten", "philippjfr" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6780", "repo": "holoviz/hvplot", "url": "https://github.com/holoviz/hvplot/issues/1072" }
gharchive/issue
Grouped/dodged bar charts not working as documented ALL software version info MacOS Big Sur, M1 Python 3.10.10 JupyterLab v3.5.2 hvplot 0.8.3 bokeh 3.1.1 Description of expected behavior and the observed behavior I'm trying to make a dodged bar chart like the one in https://hvplot.holoviz.org/reference/pandas/bar.html#basic-bar-plots but the output is always stacked, never dodged Complete, minimal, self-contained example code that reproduces the issue I just downloaded this notebook https://hvplot.holoviz.org/reference/pandas/bar.html#basic-bar-plots and did run-all: No errors in the console. Indeed, seems to be a regression in Bokeh 3.x. See also https://github.com/holoviz/holoviews/issues/5726
2025-04-01T06:38:57.618812
2024-01-10T14:56:48
2074554277
{ "authors": [ "Mikescotland", "TechHummel", "cnecrea" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6781", "repo": "home-assistant-HomeWhiz/home-assistant-HomeWhiz", "url": "https://github.com/home-assistant-HomeWhiz/home-assistant-HomeWhiz/issues/185" }
gharchive/issue
Unknown state Hi there, I've also discovered that my BEKO B5WFU78435WB<PHONE_NUMBER>) it has one new task that HomeWhiz doesn't not recognize. Right now the washing machine is "washing" but Home Assistant shows "unknown" task How can I get the "tasks" from HomeWhiz phone app and see what's wrong? Hi 👋, could you share some debug logs? Thanks! Hi there, Yes, here we go: homewhizz.log Pretty much something isn't right in my washer_substate too. I've got washer-dryer and old versions of your Addon worked fine (before the breaking change). Version 0.6 randomly was changing washer_substate but mainly stayed in "remove the laundry". The latest version is constantly stuck on "washing finished, remove laundry" while all other options and sensors are flawless. See the screenshot F999 9226 1840 7970 4793 37 It works correctly if I unplug the washing machine from the mains and plug it in again. For the following washing then the substate is reported fine, another washing after this fine and then after a day I do another washing and the substate is always "remove laundry". I haven't used the app for a year and I am unable to test if the app reads the same status as.. For some reasons the app doesn't let you see the progress of washing unless you chose "download program". It used to do it before, I'm sure. @cnecrea I decoded all the messages from your logs. And those should be, according to the app, washing? I'll close this issue for now as there haven't been any updates in some time.
2025-04-01T06:38:57.636267
2021-03-05T06:31:35
822792199
{ "authors": [ "MartinHjelmare", "jcam", "kpine", "raman325" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6782", "repo": "home-assistant-libs/zwave-js-server-python", "url": "https://github.com/home-assistant-libs/zwave-js-server-python/pull/156" }
gharchive/pull-request
Switch order of ValueNotification and Value update so that the notification includes the latest value In #154 the order of operations for creating the ValueNotification was wrong and led to the value that came in the notification being returned as the value that we was previously in the state (see https://github.com/home-assistant/core/issues/47432). This fixes the order of operations. I could use some help writing some good test cases for this one, I don't think we have enough given all the issues that have occurred as we try to fix this @kpine can you take a look at this and see if it makes sense since you spotted this immediately? I was wondering, why is value modified at all? Isn't it needed just to get the metadata? These notifications are supposed to be events or stateless values, not value updates. To test this I'd create a Value instance on the node and then create an event coming in with new data. The ValueNotification data should then be created with the Value instance data as base but be updated with the new data from the event. So we need to make sure that we can assert that difference when creating the initial Value and event data I was wondering, why is value modified at all? Isn't it needed just to get the metadata? These notifications are supposed to be events or stateless values, not value updates. Yes, it doesn't sound like the Value state should be updated when reading the docs about ValueNotification event: https://zwave-js.github.io/node-zwave-js/#/api/node?id=quotvalue-notificationquot Is there any pre-existing method to do this? value_notification.update(value.metadata) I imagine value.metadata doesn't exist yet... ah, perhaps it does... so maybe this value_notification.metadata.update(value.metadata) value_notification.metadata.update(value.metadata.data) Just tested on my prod setup, this works, and resolves https://github.com/home-assistant/core/issues/47432 value_notification.metadata.update(value.metadata.data) To test this I'd create a Value instance on the node and then create an event coming in with new data, that should trigger a ValueNotification event. The ValueNotification data should then be created with the Value instance data as base but be updated with the new data from the event. So we need to make sure that we can assert that difference when creating the initial Value and event data So it looks like my tests are missing a value in the args part of the value notification event, and I don't test the value explicitly. If I were to add it with a different value from the Values value, and then check it, would that be sufficient? To test this I'd create a Value instance on the node and then create an event coming in with new data, that should trigger a ValueNotification event. The ValueNotification data should then be created with the Value instance data as base but be updated with the new data from the event. So we need to make sure that we can assert that difference when creating the initial Value and event data So it looks like my tests are missing a value in the args part of the value notification event, and I don't test the value explicitly. If I were to add it with a different value from the Values value, and then check it, would that be sufficient? actually, the Value in my test data doesn't even have a value, so if I add a value to the notification, we can guarantee it's coming from the notification event if the ValueNotification has a value value_notification.metadata.update(value.metadata.data) The problem with this approach is that ValueNotification.endpoint is None. There are two ways to solve this: Update the entire value_notification.data. This may have unintended side effects but not sure. Update the endpoint property to return 0 by default. I think this is safe because every value has to be on an endpoint, doesn't it? I think the endpoint is not included from the ValueNotification, so the endpoint shouldn't be included in the HA event? I think the endpoint is not included from the ValueNotification, so the endpoint shouldn't be included in the HA event? The notification is for a value that does have an endpoint. There's a third option, see my upcoming commit Looks good! FYI, my scene activation and basic set events do include endpoint from the server.
2025-04-01T06:38:58.017586
2017-10-27T10:58:35
269067873
{ "authors": [ "alexbarcelo", "jonkristian", "pvizeli" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6783", "repo": "home-assistant/hassio-addons", "url": "https://github.com/home-assistant/hassio-addons/pull/196" }
gharchive/pull-request
Adding support for deployment_key for git_pull addon The main idea is to allow the user to provide a private key which will be used typically as "deployment_key". The json will be something like: { (...) "deployment_key": [ "-----BEGIN RSA PRIVATE KEY-----", "MIIEpAIBAAKCAQEArNTRDaU1nQ/Fcb5VttyxTuEEpf+W3tNe6QMVswAXAirDh4Do", "WTC1/IFae89jId5FckmFOszskAqoCLjI6ZrnxP3ZLW7yVsDn7TkI0iR21DdJ/fSU", (...) "-----END RSA PRIVATE KEY-----" ], "deployment_key_protocol": "rsa" } The multiline is a bit messy, but seems the more comfortable way to make copy&paste. The alternative is to add \n for each line and remove the lines and the json becomes unreadable. Note that deployment_key_protocol is optional and defaults to rsa. The idea is to allow the user to allow any of the typically supported ssh protocols for identities, which are: dsa ecdsa ed25519 rsa I don't consider the ~/.ssh/identity, which appears in the man pages, but I have never seen it nowadays. I assume most people will do ssh-keygen and just forget about the protocol (rsa seems to be the default in all the ssh packages I have ever used). All the aforementioned protocols are suffix to ~/.ssh/id_<suffix> so that is simply used as the file name. If this PR sounds good, I am ready to add the documentation for it. We can also use /share/keys if you think that make more sense and add a options to set the name, like we use on google_assistant or snips? I didn't like the idea to require Samba in order to use that. Because that is the only method I have seen that can be used in order to upload a file like a key. But if you think is better I can change. The deployment key doesn't change (typically is set and forgotten) so a bit messy initial setting may not be so bad. For me, it dosn't matter. It should be user friendly as it is possible. So if you prefere this, it is okay. You can also use scp/sftp with ssh add-on to access to share. If you know a nice webinterface for data handling (like file station from synology) we can add it as add-on. Anyway, you are free to handle that as file or as options. For me is both okay :+1: Do we neef also the openssh client? I think that the current workaround for multiline-private-key-into-config-json is good enough. It should be enough for my scenario, and I assume that it will be a common case for people having private repos somewhere (either cloud or "in premises"). Last two commits should address your comments. Please update also the website "repository": "match((?:git|ssh|https?|git@[-\\w.]+):(\/\/)?(.*?)(\\.git)(\/?|\\#[-\\d\\w._]+?))", This needs to improve, because some of us aren't using github, for instance mine is gitlab@git.domain.com:group/repo.git I'd suggest a regexp which can match anything word before @. @jonkristian I think that you meant to comment on home-assistant/hassio-addons#198 Because that line that you are quoting is from that PR, not this one. I was very confused at first... @alexbarcelo So sorry about the confusion. Should I move my comment?
2025-04-01T06:38:58.019056
2017-09-28T06:44:50
261209404
{ "authors": [ "pschmitt", "pvizeli" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6784", "repo": "home-assistant/hassio-build", "url": "https://github.com/home-assistant/hassio-build/pull/37" }
gharchive/pull-request
Fix typo: BUILT_TYPE -> BUILD_TYPE Sidenote: have issues been disabled intentionally for this repo? I build a new release in 10min
2025-04-01T06:38:58.021086
2019-12-03T23:31:44
532338231
{ "authors": [ "balloob", "gmatyszczak" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6785", "repo": "home-assistant/home-assistant-android", "url": "https://github.com/home-assistant/home-assistant-android/issues/90" }
gharchive/issue
Alert window doesn't has correct title or button When the frontend uses the JavaScript alert(), it shows an alert to the user with the wrong title and a missing caption for the OK button (button does work). @gmatyszczak does this also fix the %s or should I open a new issue for that? I couldn't reproduce the issue with %s. Could you provide steps? Or retest it on your side? Cannot reproduce either anymore. Perfect !
2025-04-01T06:38:58.179756
2018-03-29T09:51:05
309690484
{ "authors": [ "JIOB", "OttoWinter", "Roemer", "Shadex12", "Spartan-II-117", "Wonkavator", "accelle17", "ajfriesen", "aliaghil", "allmoney", "awarecan", "backcountrymountains", "balloob", "benjamin1492", "bryanyork", "darylsargent", "dshokouhi", "e1miran", "edif30", "eeepsylon", "emontnemery", "fanaticDavid", "fillwe", "frankhommers", "gezonthenet", "grangemd", "harry19023", "iorifly", "j0barr", "josephkiser04", "joydashy", "kalkih", "kdjordjev", "kristoffernolgren", "masterz666", "psfales", "rgruebel", "smoke007", "soldag", "ttaidapos" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6786", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/13530" }
gharchive/issue
Cast groups unavailable Home Assistant release with the issue: 0.66.0.beta1 Last working Home Assistant release (if known): 0.65.6 Operating environment (Hass.io/Docker/Windows/etc.): Raspbian GNU/Linux 9 (stretch) Component/platform: media_player.cast Description of problem: Since update to the latest beta version, my cast groups are unavailable now and then. However, the single devices of the groups can be controlled, but not the groups itself. I encountered this issues typically in the morning. When restarting HASS, all cast devices and groups are available, again. Problem-relevant configuration.yaml entries and (fill out even if it seems unimportant): discovery: media_player: - platform: cast Traceback (if applicable): Additional information: Hmmm, thanks for reporting this 👍 Seems a lot like https://github.com/home-assistant/home-assistant/issues/13483 and probably due to the groups changing their "elected leader". Could you also enable debug logs as mentioned in the comment in there? It would really help. Maybe back to the original elected leader? :) Hmmm, that might be it, but I have not idea how that could affect the socket client. Oh gosh, this is getting interesting... 😂 @dshokouhi The problem is that not only the host (or IP address) of the device that is the "elected leader" or the device that accepts all commands changes, but also the port number. And that dynamically, so there's no way that a user can list the cast members in their configuration. btw, from what I know now, the issue is not related to the discovery aspect of the cast integration. In fact, we discover elected leader changes just fine. The issue seems to be somewhere in the disconnect from old leader + connect to new leader code. Oh wow nice to hear you guys seem to have pinned down the issue :) if you need any help testing I am sure myself and others would be glad to help. I'm still trying to find out what exactly caused this, but If you'd like to help 😬, it would be good to know whether with the following configuration the issue still happens (with discovery: disabled). media_player: - platform: cast scan_interval:<PHONE_NUMBER> I personally am not seeing the issue with my setup with that config, but it would be good to get some data from others too. This configuration should not really affect anything and the cast integration should still work (without the new semi-polling mode I introduced in the linked PR). If the error doesn't appear with that configuration, it would mean that we know the code change that caused this, but I still have no idea why the connections drop. Ok I have disabled discovery for google_cast, and applied the same config for the media_player component. Will reply back if I notice the issue. @OttoWinter the issue came back overnight, one of my groups is currently unavailable and the other one went unavailable over night and became available again. @dshokouhi Ok. At least that rules out one possible explanation... Thanks very much! I'd be happy to help test as well if there's are any new updates, I just closed my issue after troubleshooting and determining that cast groups are causing my problems, and that this issue probably relates to mine. If I include cast groups in global or cast discovery, then I lose the ability to cast to all but a few of my cast-enabled devices when Home Assistant starts up. If I explicitly define just the individual devices and exclude cast groups, then everything works fine. As a point of clarification, when I say "lose the ability to cast", I mean that when I hit the cast icon in Pandora, Spotify, or Google Play Music, I only get the devices with very old firmwares as options (Sony Speaker with Chromecast Built-In and an Insignia Alarm Clock with Chromecast Built-in). None of the Chromecast Audios, Google Homes, Nvidia Shields, nor any of the associated groups show up. I've attached a log. home-assistant test2.log I'm also seeing this behaviour today, running Home Assistant 0.67.1: The more-info dialog shows the details of the bottom group, but the state of the top group is identical. I also have the individual Chromecast Audio units in my Home Assistant setup and it seems one of them (Garoeda I) was unavailable for about half a minute. This caused the two groups (that include that specific Chromecast Audio) to become and remain unavailable. This is what syslog logged on my Raspberry Pi 3: pi@Ra:~ $ cat /var/log/syslog | grep unavailable Apr 22 17:13:41 Ra hass[13460]: #033[32m2018-04-22 17:13:41 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.garoeda_i, old_state=<state media_player.garoeda_i=off; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-22T04:38:21.736327+02:00>, new_state=<state media_player.garoeda_i=unavailable; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-22T17:13:41.211439+02:00>>#033[0m Apr 22 17:13:42 Ra hass[13460]: #033[32m2018-04-22 17:13:42 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio, old_state=<state media_player.home_audio=off; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T04:38:21.754878+02:00>, new_state=<state media_player.home_audio=unavailable; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T17:13:42.196000+02:00>>#033[0m Apr 22 17:13:47 Ra hass[13460]: #033[32m2018-04-22 17:13:47 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio_with_bathroom, old_state=<state media_player.home_audio_with_bathroom=unavailable; friendly_name=Home Audio with bathroom, supported_features=21437 @ 2018-04-22T15:47:52.989243+02:00>, new_state=<state media_player.home_audio_with_bathroom=off; friendly_name=Home Audio with bathroom, supported_features=21437 @ 2018-04-22T17:13:47.477800+02:00>>#033[0m Apr 22 17:13:47 Ra hass[13460]: #033[32m2018-04-22 17:13:47 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio, old_state=<state media_player.home_audio=unavailable; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T17:13:42.196000+02:00>, new_state=<state media_player.home_audio=off; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T17:13:47.500569+02:00>>#033[0m Apr 22 17:14:16 Ra hass[13460]: #033[32m2018-04-22 17:14:16 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.garoeda_i, old_state=<state media_player.garoeda_i=unavailable; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-22T17:13:41.211439+02:00>, new_state=<state media_player.garoeda_i=off; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-22T17:14:16.308174+02:00>>#033[0m Apr 22 17:14:16 Ra hass[13460]: #033[32m2018-04-22 17:14:16 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio_with_bathroom, old_state=<state media_player.home_audio_with_bathroom=off; friendly_name=Home Audio with bathroom, supported_features=21437 @ 2018-04-22T17:13:47.477800+02:00>, new_state=<state media_player.home_audio_with_bathroom=unavailable; friendly_name=Home Audio with bathroom, supported_features=21437 @ 2018-04-22T17:14:16.640050+02:00>>#033[0m Apr 22 17:14:16 Ra hass[13460]: #033[32m2018-04-22 17:14:16 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio, old_state=<state media_player.home_audio=off; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T17:13:47.500569+02:00>, new_state=<state media_player.home_audio=unavailable; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T17:14:16.659293+02:00>>#033[0m And this is what happened according to Home Assistant: 2018-04-22 17:13:41 WARNING (Thread-6) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-04-22 17:13:42 WARNING (Thread-8) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-04-22 17:13:51 ERROR (Thread-10) [pychromecast.socket_client] Failed to connect. No retries. 2018-04-22 17:14:11 ERROR (Thread-6) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-04-22 17:14:12 ERROR (Thread-8) [pychromecast.socket_client] Failed to connect: timed out. aborting due to stop signal. 2018-04-22 17:14:16 ERROR (Thread-11) [pychromecast.socket_client] Error reading from socket. 2018-04-22 17:14:16 ERROR (Thread-12) [pychromecast.socket_client] Error reading from socket. 2018-04-22 17:14:16 WARNING (Thread-11) [pychromecast.socket_client] Error communicating with socket, resetting connection 2018-04-22 17:14:16 WARNING (Thread-12) [pychromecast.socket_client] Error communicating with socket, resetting connection 2018-04-22 17:14:16 ERROR (Thread-11) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-04-22 17:14:16 ERROR (Thread-12) [pychromecast.socket_client] Failed to connect, retrying in 5.0s My configuration for the Cast component is basic: - platform: cast Did some further testing today and have included the logs and some screenshots. The first screenshot is my casting choices when cast groups are turned off in Home Assistant, but all other cast devices are turned on (meaning I'm selecting them individually with their host ip address). Everything shows up, including the groups. The "home assistant log test6" is when I turned groups on by: platform: cast While Home Assistant was running, I also restarted the router and then restarted every single cast device individually. All that should be evident in the log. None of that worked as you can see in screenshot 2, which is my list of casting choices. The Sony speaker and the Insignia Alarm clock, both running much older versions of chromecast software are the only devices present. The groups you see are groups that contain one or both of those two devices. I then commented out platform: cast to remove the groups, but kept my individual selections by host, ie: ###platform: cast platform: cast host: <IP_ADDRESS> name: Living Room Just like screenshot 1, all my device and groups now show up as cast targets. I included "home assistant log test7" with these conditions to compare against test6. home-assistant test6.log home-assistant test7.log Can you cast to your devices from, say, your phone, when they show as unavailable in Home Assistant? While I didn't post any screenshots, I too am seeing devices and groups become unavailable. I've always assumed the two issues were related. However, if you're still able to cast to a device / group that's unavailable in Home Assistant, then I agree I may have a different issue. I will have to test that when my Cast groups are unavailable the next time, as they have suddenly gone back to their off state almost 2 hours ago: Syslog: pi@Ra:~ $ cat /var/log/syslog | grep unavailable Apr 23 02:17:23 Ra hass[13460]: #033[32m2018-04-23 02:17:23 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio, old_state=<state media_player.home_audio=unavailable; friendly_name=Home Audio, supported_features=21437 @ 2018-04-22T17:14:16.659293+02:00>, new_state=<state media_player.home_audio=off; friendly_name=Home Audio, supported_features=21437 @ 2018-04-23T02:17:23.181650+02:00>>#033[0m Apr 23 02:17:23 Ra hass[13460]: #033[32m2018-04-23 02:17:23 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home_audio_with_bathroom, old_state=<state media_player.home_audio_with_bathroom=unavailable; friendly_name=Home Audio with bathroom, supported_features=21437 @ 2018-04-22T17:14:16.640050+02:00>, new_state=<state media_player.home_audio_with_bathroom=off; friendly_name=Home Audio with bathroom, supported_features=21437 @ 2018-04-23T02:17:23.219506+02:00>>#033[0m Apr 23 02:17:43 Ra hass[13460]: #033[32m2018-04-23 02:17:43 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.garoeda_i, old_state=<state media_player.garoeda_i=off; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-22T17:14:16.308174+02:00>, new_state=<state media_player.garoeda_i=unavailable; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-23T02:17:43.591443+02:00>>#033[0m Apr 23 02:17:59 Ra hass[13460]: #033[32m2018-04-23 02:17:59 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.garoeda_i, old_state=<state media_player.garoeda_i=unavailable; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-23T02:17:43.591443+02:00>, new_state=<state media_player.garoeda_i=off; friendly_name=Garoeda I, supported_features=21437 @ 2018-04-23T02:17:59.393369+02:00>>#033[0m Home Assistant log: 2018-04-23 02:17:43 WARNING (Thread-6) [pychromecast.socket_client] Heartbeat timeout, resetting connection Please test to see if they are available as cast targets regardless of status. What I've found is that whether they are "off" or "unavailable", they will not show up as available to cast to from Pandora, Google Play Music, or any other app that has cast capability. I'd love to have validation of that behavior. Please test to see if they are available as cast targets regardless of status. What I've found is that whether they are "off" or "unavailable", they will not show up as available to cast to from Pandora, Google Play Music, or any other app that has cast capability. I'd love to have validation of that behavior. It's very weird. Regardless of the media player status within Home Assistant, I can tell a Google Home to play a song on any of the individual Chromecast Audios or Google Homes and it will do so. If I try tell it to play to a cast group, however, only one set of speakers will play, which I assume is the leader of the group. And if I open Pandora or Google Play Music and hit the Cast icon, none of the individual devices or groups show up as options (except - as noted in an earlier comment - the older firmware devices with Chromecast built-in from Sony and Insignia. Those are solid). As a reminder, this behavior is only when I have Home Assistant running with platform: cast discovering the groups. If I allow it to only discover the individual cast players, then everything works fine both in Home Assistant and in Pandora, Google Home, etc. In addition to unavailable I just noticed unknown as well. @dshokouhi Are you using cast groups in Home Assistant? I've found if I remove them, then the unavailable issue goes away (unfortunately, so too does a bunch of functionality). Could you test by removing cast groups and see if you get to: @benjamin1492 this bug is only about cast groups so removing them to test is not an option. They need to work which is why we are troubleshooting this. Understood. I'll wait on the resolution to this one, and if it doesn't fix my problem, I'll reopen my other case. For now, I'll leave ya'll alone to get the cast groups fixed. I am having an issue where none of my cast components show up in home assistant anymore even if I use discovery. So my problem is definitely different. Running the beta, I was able to get all the devices to A) show as available in Hass, and B) show as cast targets. But that only lasted for about 15 minutes before I was back to A) only the leader playing when cast group selected, and B) having no cast targets for GPM or Pandora. The 15 minutes to failure is an improvement...used to be instantaneous as soon as I started Hass. @benjamin1492 have you tried removing/unplugging devices one by one to see if one of them is a factor? Just chiming in because I'm seeing the same thing. My group will randomly go "unavailable" a few times a day, and then a few hours later it goes back to available again. I only have two devices in the group, so it seems plausible that when the leader selection is changed, it toggles between the two devices. At this moment, it's showing as unavailable in HA, but I'm able to use the group with Pandora Hi, so I had another look at this problem today to see what could be causing this. One potential problem is that the disconnection logic is currently not perfect and we don't wait until the previous socket has truly disconnected when re-establishing a connection. This doesn't seem like something that could cause this error to me but I'm running out of potential causes. I've update the cast integration a bit to resolve above issue. If someone wants to help, copy this file to <config_dir>/custom_components/media_player/cast.py. I'm not sure if it would solve the issue but it would at least be worth a try. I've been running that 4/28 version for about a day and it hasn't solved the problem, but subjectively, I think it may a little different. It went unavailable three times today, but both were only for a minute or so (rather than hours like I think it was before) @psfales Thanks for testing! I think 30 seconds to 60 seconds are more or less expected and there's not much we can do to cut this time down: It takes 10-15 seconds until we receive the mDNS message with the new host. Then we need to disconnect from the old chromecast. Can take up to 5-10 seconds depending on the amount of data waiting to be transferred and so on. Finally, we need to connect to the new host. Can take 10+ seconds. Especially while the new host is still setting up to receive connections. Unfortunately, it hasn't stayed so good. It went unavailable at 5:30 this morning and stayed unavailable for 2.5 hours until I rebooted Oh noes :/ Reopening then... Unfortunately, I won't be able to work on this for the next week and a half though because I'm in the process of studying for my final exams 📚 Certainly understand about your schedule. When you have time, I'm curious.... Is it possible to tell when the elected leaders changes, or which device it is at any given time? I think it would be interesting to confirm that the group going unvailable (or return to available) is correlated with a leader change. There's no real visual way to see that from the front-end. But if you have debug logs enabled you can see a Disconnecting from previous chromecast message in there when a leader change is detected. I experience this problem I'm pretty new to Homeassistant but am a developer. Let me know if you need anything! I also have this problem. Not a developer but happy to test/help where I can. I just realized that I have another chromecast devices (not group) that became unavailable prior to the chromecast group, 3 mins after the single device became available again, the group went unavailable. The kitchen devices unavailablitiy is not restricted to chromecast, when it's unavailable in chromecast, it's not reachable from any other apps either. "All"s problems is howeber only limited to chromecast. Let me know if I can helt in any way :) Same goes for me and the time a group takes to become unavailable seems totally random. I went away for a week so no music was played via other devices and on my return all groups were available. Unsure if a total fluke or due to not casting from other devices there was no need to elect a new group "leader". I presume Home Assistant was casting TTS weather alerts etc because i forgot to switch them off. @JIOB, did you check your history? i'd be really curious to kno if not interacting with it caused it to hold stable On Fri, May 18, 2018 at 5:15 AM JIOB<EMAIL_ADDRESS>wrote: Same goes for me and the time a group takes to become unavailable seems totally random. I went away for a week so no music was played via other devices and on my return all groups were available. Unsure if a total fluke or due to not casting from other devices there was no need to elect a new group "leader". I presume Home Assistant was casting TTS weather alerts etc because i forgot to switch them off. — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/home-assistant/home-assistant/issues/13530#issuecomment-390188171, or mute the thread https://github.com/notifications/unsubscribe-auth/AEuJQHrfPVcCjQ3uWSxKLDI_pSNvpWk1ks5tzrtNgaJpZM4TADhw . I didn't check at the time and its since been overwritten. If the same situation occurs again prior to a fix ill ensure i've got a log to share. Not my proudest work, but I just did this as a cludge: - alias: 'All unavailable for 3 minutes' trigger: platform: state entity_id: media_player.all to: 'unavailable' for: minutes: 15 action: service: homeassistant.restart homeassistant reboots whenever the group is unavialable for more than 15 minutes. I have a feeling this will come back and bite me in the ass... I am actually wondering if maybe the cast platform can create a service call that allows us to basically "reload" the groups like what happens during a home assistant restart. A hass restart is the only way to bring the groups back maybe there is a way to trigger the rediscovery of just the cast groups when they become unavailable? I think I saw some platforms use unloading like hue maybe something similar can be triggered here? Sorry if I am shooting in the dark just trying to help offer ideas/suggestions :) I have been thinking the same, but not sure if it has to be a service, it should just do it if the groups lose connection in that case. It's because of mess with configuration. I had same problem, I checked all my configuration an there was some parts that I correct them. Then I removed the database and restart the machine. It's been about 10 days that I don't have that problem anymore. I agree with @dshokouhi and @kristoffernolgren it would be nice if some logic could be built in to restart the service when a group/groups become unavailable. @aliaghil which parts of your configuration did you correct? My configuration now only contains.. - platform: cast In my configuration I had this code in sensors: ` sensor: platform: systemmonitor resources: type: since_last_boot type: processor_use type: memory_free type: disk_free arg: / platform: template sensors: time_template: friendly_name: Time value_template: '{{ as_timestamp(now()) | timestamp_custom("%d/%m %H:%M", True) }}' since_last_boot_template: friendly_name: Up Time value_template: > {%- set s = states.sensor.since_last_boot.state.split('.')[0] -%} {{ s }} platform: version name: version and I changed it to: sensor: platform: systemmonitor resources: type: since_last_boot type: processor_use type: memory_free type: disk_free arg: / platform: template sensors: time_template: friendly_name: "Time" value_template: '{{ as_timestamp(now()) | timestamp_custom("%d/%m %H:%M", True) }}' since_last_boot_template: friendly_name: "Up Time" value_template: > {%- set s = states.sensor.since_last_boot.state.split('.')[0] -%} {{ s }} platform: version name: "Version" ` In my sensors wherever there was friendly_name: I used quotation mark for it. After version 0.60 developers add cloud component, and I add it to my configuration, but didn't use it, even I didn't login to it. I think the problem was because of this, after I confined it, and logged in to it, the problem gone. in configuration.yaml i had this: ` homeassistant: customize_glob: !include config/customize_glob.yaml packages: !include_dir_named config/packages/ frontend: extra_html_url: /local/custom_ui/state-card-custom-ui.html I disabled that two. homeassistant: customize_glob: !include config/customize_glob.yaml packages: !include_dir_named config/packages/ frontend: extra_html_url: - /local/custom_ui/state-card-custom-ui.html ` I was using separated config I found it out, these files shouldn't go in the config folder config/config automations.yaml entity_registry.yaml known_devices.yaml scripts.yaml secrets.yaml I bring them back to root config folder config the last one I update all my automation with Automation editor. @aliaghil are you saying that those steps have permanently fixed your instance of the cast component? @JIOB yes. I update hassio to 0.71.0 and unfortunately the problem come back again :( I'm going to downgrade it to v0.70.1 and see what's going on... Just updated HomeAssistant to 0.72.0 and it forced me to reconfigure my Google cast devices. Anybody know if the change has the ability to fix this issue? Will post back if i find anything. I also updated to 0.72.0 and got the dialog about forcing me to reconfigure Google Cast. But I don't understand what this means. It apparently created a file in the config directory called ".config_entries.json" which has an entry for "Google Cast." But other than that one-time interaction I don't see any change in behavior. "Google Cast" does not appear anywhere in the UI and all the Chromecast, Chromecast Audio, and audio groups that were there before are still lthere. On Sun, Jun 24, 2018 at 6:28 AM, JIOB<EMAIL_ADDRESS>wrote: Just updated HomeAssistant to 0.72.0 and it forced me to reconfigure my Google cast devices. Anybody know if the change has the ability to fix this issue? Will post back if i find anything. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/home-assistant/home-assistant/issues/13530#issuecomment-399749490, or mute the thread https://github.com/notifications/unsubscribe-auth/AKyoCsrJK8dEx3GKnh19WWJhmAk19vjBks5t_3fbgaJpZM4TADhw . After updating to 0.72.0 and using the new UI config method, I am still experiencing this issue. So it doesn't look like the new config method fixed the issue. I am having the same issue here with 0.72. Individual devices work, by group gets dropped. I do not recall having this issue before The HA log shows the issue happening first two nights ago @2:38am, which seems to somehow recovered (stopped reporting errors in the log) , and similar errors last nigh @2:27am , which have not recovered yet 2018-07-05 02:38:17 ERROR (Thread-10) [pychromecast.socket_client] Error reading from socket. 2018-07-05 02:38:17 WARNING (Thread-10) [pychromecast.socket_client] Error communicating with socket, resetting connection 2018-07-05 02:38:17 ERROR (Thread-10) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-05 02:38:22 ERROR (Thread-10) [pychromecast.socket_client] Failed to connect. No retries. 2018-07-05 02:38:39 WARNING (Thread-8) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-05 02:38:54 ERROR (Thread-8) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-05 02:41:58 WARNING (Thread-9) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-05 02:42:05 ERROR (Thread-9) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-05 03:05:35 WARNING (Thread-7) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-05 03:05:38 ERROR (Thread-7) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-05 03:29:43 ERROR (MainThread) [homeassistant.core] Error doing job: Task exception was never retrieved 2018-07-06 02:27:17 ERROR (Thread-191485) [pychromecast.socket_client] Error reading from socket. 2018-07-06 02:27:17 WARNING (Thread-191485) [pychromecast.socket_client] Error communicating with socket, resetting connection 2018-07-06 02:27:17 ERROR (Thread-191485) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-06 02:27:22 ERROR (Thread-191485) [pychromecast.socket_client] Failed to connect. No retries. 2018-07-06 02:27:45 WARNING (Thread-8) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-06 02:27:52 ERROR (Thread-8) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-06 02:51:57 WARNING (Thread-7) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-06 02:52:04 ERROR (Thread-7) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-06 02:52:15 ERROR (Thread-481327) [pychromecast.socket_client] Error reading from socket. 2018-07-06 02:52:15 WARNING (Thread-481327) [pychromecast.socket_client] Error communicating with socket, resetting connection 2018-07-06 02:52:15 ERROR (Thread-481327) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-06 04:35:58 WARNING (Thread-9) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-06 04:36:01 ERROR (Thread-9) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-06 07:41:15 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting I use Chromecast Audio groups extensively and use playing/off state changes to trigger many automations. (volume setting, turning off and on connected amps etc). Here's a log from last night, which shows the groups becoming unavailable at 22:36, coming back 61s later, then going unavailable again 1s later. The groups recover during the night. There are two groups, "All" and "Downstairs". "sony" is the A/V receiver with cast built in. 'sony_2' is the songpal component of same device. I doubt that's relevant, other than to show that songpal and pychromecast both report unavailable at roughly the same times. Casting GPM to the CCA 'bedroom' begins at 22:23 and ends at 23:22. 2018-07-11 22:23:36 INFO bedroom -> 'idle' 2018-07-11 22:23:36 INFO (MainThread) [homeassistant.core] Bus:Handling <Event call_service[L]: service_data=entity_id=media_player.bedroom, volume_level=0.15, service_call_id=1975828368-6972, service=volume_set, domain=media_player> 2018-07-11 22:23:36 INFO (Thread-21) [pychromecast.controllers] Receiver:setting volume to 0.1 2018-07-11 22:23:36 INFO bedroom -> 'playing' ... 2018-07-11 22:34:33 INFO bedroom -> 'playing' 2018-07-11 22:36:01 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='afb73d06-77f7-39df-d206-0b98b58be99d', model_name='STR-DN860', friendly_name='Sony') 2018-07-11 22:36:04 WARNING (Thread-53) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-11 22:36:06 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 22:36:06 DEBUG (Thread-53) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-11 22:36:06 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 22:36:06 WARNING (Thread-52) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-11 22:36:06 DEBUG (Thread-52) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-11 22:36:06 INFO downstairs -> 'unavailable' 2018-07-11 22:36:06 INFO all -> 'unavailable' 2018-07-11 22:36:11 WARNING (Thread-24) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-11 22:36:11 DEBUG (Thread-24) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-11 22:36:11 INFO sony -> 'unavailable' 2018-07-11 22:36:36 ERROR (Thread-53) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-11 22:36:36 ERROR (Thread-52) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-11 22:36:41 ERROR (Thread-24) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-11 22:37:19 INFO sony_2 -> 'unavailable' 2018-07-11 22:37:55 INFO sony_2 -> 'off' 2018-07-11 22:37:55 DEBUG (Thread-52) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-11 22:37:55 INFO all -> 'off' 2018-07-11 22:37:55 DEBUG (Thread-24) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-11 22:37:55 DEBUG (Thread-53) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-11 22:37:55 INFO sony -> 'off' 2018-07-11 22:37:55 INFO downstairs -> 'off' 2018-07-11 22:37:56 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='afb73d06-77f7-39df-d206-0b98b58be99d', model_name='STR-DN860', friendly_name='Sony') 2018-07-11 22:37:56 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 22:37:56 DEBUG (MainThread) [homeassistant.components.media_player.cast] Discovered chromecast with same UUID: ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 22:37:56 DEBUG (MainThread) [homeassistant.components.media_player.cast] Disconnecting from chromecast socket. 2018-07-11 22:37:56 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 22:37:56 INFO all -> 'unavailable' 2018-07-11 22:37:56 DEBUG (MainThread) [homeassistant.components.media_player.cast] Discovered chromecast with same UUID: ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 22:37:56 DEBUG (MainThread) [homeassistant.components.media_player.cast] Disconnecting from chromecast socket. 2018-07-11 22:37:56 INFO downstairs -> 'unavailable' 2018-07-11 22:38:01 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connecting to cast device ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 22:38:01 INFO (Thread-20) [pychromecast] Querying device status 2018-07-11 22:38:01 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connecting to cast device ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 22:38:01 ERROR (Thread-20) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-11 22:38:01 INFO (Thread-6) [pychromecast] Querying device status 2018-07-11 22:38:01 ERROR (Thread-6) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-11 22:39:20 INFO bedroom -> 'idle' 2018-07-11 22:39:20 INFO bedroom -> 'playing' ... 2018-07-11 23:22:37 INFO bedroom -> 'playing' 2018-07-11 23:22:51 INFO (Thread-11) [pychromecast] Quiting current app 2018-07-11 23:22:51 INFO (Thread-11) [pychromecast.controllers] Receiver:Stopping current app '2872939A' 2018-07-11 23:22:51 INFO bedroom -> 'off' 2018-07-11 23:37:14 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 23:37:14 DEBUG (MainThread) [homeassistant.components.media_player.cast] Discovered chromecast with same UUID: ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 23:37:14 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connecting to cast device ChromecastInfo(host='<IP_ADDRESS>', port=42904, uuid='02005ffe-7f5c-4512-b118-50e1d4f87cdc', model_name='Google Cast Group', friendly_name='All') 2018-07-11 23:37:14 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 23:37:14 INFO (Thread-3) [pychromecast] Querying device status 2018-07-11 23:37:14 DEBUG (MainThread) [homeassistant.components.media_player.cast] Discovered chromecast with same UUID: ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 23:37:14 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connecting to cast device ChromecastInfo(host='<IP_ADDRESS>', port=42760, uuid='c790e2bc-dae8-439b-8ab9-408266d10422', model_name='Google Cast Group', friendly_name='Downstairs') 2018-07-11 23:37:14 INFO (Thread-4) [pychromecast] Querying device status 2018-07-11 23:37:15 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connection successful! 2018-07-11 23:37:15 INFO all -> 'off' 2018-07-11 23:37:15 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connection successful! 2018-07-11 23:37:15 INFO downstairs -> 'off' 2018-07-12 02:23:21 WARNING (Thread-30) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 02:23:21 DEBUG (Thread-30) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 02:23:21 INFO lounge_mini -> 'unavailable' 2018-07-12 02:23:36 DEBUG (Thread-30) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-12 02:23:36 INFO lounge_mini -> 'off' 2018-07-12 02:23:38 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='e957ad8e-8b83-0f38-97ae-89c501ce6e34', model_name='Google Home Mini', friendly_name='Lounge Mini') 2018-07-12 02:45:48 WARNING (Thread-29) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 02:45:48 DEBUG (Thread-29) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 02:45:48 INFO office -> 'unavailable' 2018-07-12 02:45:55 ERROR (Thread-29) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-12 02:46:00 DEBUG (Thread-29) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-12 02:46:00 INFO office -> 'off' 2018-07-12 02:46:03 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='a0721def-786f-a924-c5cd-75cd5a7f569d', model_name='Chromecast Audio', friendly_name='Office') 2018-07-12 03:15:51 WARNING (Thread-28) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 03:15:51 WARNING (Thread-27) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 03:15:51 DEBUG (Thread-28) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 03:15:51 DEBUG (Thread-27) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 03:15:51 INFO chromecast_lounge -> 'unavailable' 2018-07-12 03:16:07 DEBUG (Thread-28) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-12 03:16:07 INFO chromecast_lounge -> 'off' 2018-07-12 03:16:10 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='dd9ff914-24cf-86ad-5ed1-85ce9ac1af8e', model_name='Chromecast', friendly_name='Chromecast Lounge') 2018-07-12 04:05:28 WARNING (Thread-26) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 04:05:28 DEBUG (Thread-26) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 04:05:28 INFO bedroom_home -> 'unavailable' 2018-07-12 04:05:35 ERROR (Thread-26) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-12 04:05:41 DEBUG (Thread-26) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-12 04:05:41 INFO bedroom_home -> 'off' 2018-07-12 04:05:42 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='5e2fc371-10e9-ea49-23b4-216a51072fc7', model_name='Google Home', friendly_name='Bedroom Home') 2018-07-12 04:51:47 WARNING (Thread-32) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 04:51:47 DEBUG (Thread-32) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 04:51:47 INFO bedroom -> 'unavailable' 2018-07-12 04:51:55 ERROR (Thread-32) [pychromecast.socket_client] Failed to connect, retrying in 5.0s 2018-07-12 04:52:00 DEBUG (Thread-32) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-12 04:52:00 INFO bedroom -> 'off' 2018-07-12 04:52:04 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='b5fda076-4ba4-1565-9f3d-8f98ddfcb78c', model_name='Chromecast Audio', friendly_name='Bedroom') 2018-07-12 04:56:25 WARNING (Thread-25) [pychromecast.socket_client] Heartbeat timeout, resetting connection 2018-07-12 04:56:25 DEBUG (Thread-25) [homeassistant.components.media_player.cast] Cast device availability changed: LOST 2018-07-12 04:56:25 INFO kitchen -> 'unavailable' 2018-07-12 04:56:39 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered previous chromecast ChromecastInfo(host='<IP_ADDRESS>', port=8009, uuid='24ddbab5-3d2c-aebf-805c-e866fa3c2458', model_name='Chromecast Audio', friendly_name='Kitchen') 2018-07-12 04:56:41 DEBUG (Thread-25) [homeassistant.components.media_player.cast] Cast device availability changed: CONNECTED 2018-07-12 04:56:41 INFO kitchen -> 'off' Since switching to hassio I have the problem too: Any news on this issue? Still not working for me with the latest version :( @OttoWinter @balloob - Is anyone actively working on this issue? This has been broken for months. I used to have automations to play music and announcements to certain groups. But casting to groups is anything but reliable. What can we do to help test/debug? Yeah, absolutely nothing has changed since this was first reported (neither better nor worse). :( I have a far from ideal workaround for triggering automations while casting to speaker groups. I'm using the Unifi WAP component to monitor the bytes flowing to Chromecast Audio's. If one spikes above 20000, then the amp supplying power to the speakers it feeds turns on. There's usually a delay of 5-20 seconds given the polling frequency, but it's otherwise tolerable. Also, if there are no spikes on any of the Chromecasts for 10 minutes, then the corresponding amp is (hopefully) turned off. This is only for cast groups...casting to individual Chromecast Audio's is much more straightforward. Also, to make this work, I had to convert all of the Chromecast Audio's from wired to wireless because the Unifi controller doesn't track bytes for wired devices. I haven't noticed any issues going to wireless, but I'd really prefer to keep them wired. While this is a passable work around, it is only about 60% effective. The speakers always turn on, but only turn off after 10 minutes about half the time. I haven't figured out why yet, so I have an automation that turns off all the amps at midnight. Thanks, I'll check that out then. I like the idea of the broadcast not killing any music that was playing. As since cast groups stopped working, I just listed out my cast devices within my script/automation as a work around for now. alias: TTS Test sequence: data: entity_id: media_player.basement_speakers, media_player.basement_assistant, media_player.garage_assistant, media_player.pole_barn_assistant volume_level: 0.6 service: media_player.volume_set data: entity_id: media_player.basement_speakers, media_player.basement_assistant, media_player.garage_assistant, media_player.pole_barn_assistant message: This is a test message. Is the volume loud enough? service: tts.google_say I too would like to know where things are at with this issue. I used to have automations fire when certain groups were playing etc. They've been useless for months. Agree this is still an issue. Now I am even starting to get this with certain cast devices. I have the same problem Wish there was a way to vote on issues, this is such a pain, having to reboot HA nightly to keep Chromecast groups somewhat operational. My server shuts down over night and starts up the next morning. The groups are often already non-functional by the time I get home from work! On Thu, Sep 20, 2018 at 11:01 AM joydashy<EMAIL_ADDRESS>wrote: Wish there was a way to vote on issues, this is such a pain, having to reboot HA nightly to keep Chromecast groups somewhat operational. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/home-assistant/home-assistant/issues/13530#issuecomment-423102075, or mute the thread https://github.com/notifications/unsubscribe-auth/AQNwbRzGtM6FyPjMn9UnszZ3-4qDxLq0ks5uc1l1gaJpZM4TADhw . My server shuts down over night and starts up the next morning. The groups are often already non-functional by the time I get home from work! … On Thu, Sep 20, 2018 at 11:01 AM joydashy @.***> wrote: Wish there was a way to vote on issues, this is such a pain, having to reboot HA nightly to keep Chromecast groups somewhat operational. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub <#13530 (comment)>, or mute the thread https://github.com/notifications/unsubscribe-auth/AQNwbRzGtM6FyPjMn9UnszZ3-4qDxLq0ks5uc1l1gaJpZM4TADhw . My HA also reboots when it needs to use Chromecast groups and detects they are Unavailable. This of course introduces a delay in the functionality. It's a sucky workaround! Give 0.78.1 a try Give 0.78.1 a try Looks promising! Will update soon. already on 0.78.1 and looks like it fixed my issue. got instances of unavailable issues on my 3 cast groups and they went from off to unavailable to off in seconds. Updated to 0.78.1 yesterday evening. This morning woke up to 2 of my 4 groups Unavailable... and one of my individual chromecast audios, as well. Had to restart HA to re-connect to them. I also still have the issues with 0.78.1. Also still having the same problem with 0.78.1 @awarecan Thanks for your help with this. If you need anything in addition to what I've attached, please let me know. I'd be glad to help any way I can in getting this resolved. Here's my config: # Chromecast Audios - platform: cast name: Den host: <IP_ADDRESS> - platform: cast name: Kitchen host: <IP_ADDRESS> - platform: cast name: Dining Room host: <IP_ADDRESS> - platform: cast name: Living Room host: <IP_ADDRESS> - platform: cast name: Patio host: <IP_ADDRESS> - platform: cast name: Master Bedroom host: <IP_ADDRESS> - platform: cast name: Master Bathroom host: <IP_ADDRESS> - platform: cast name: Master Closet host: <IP_ADDRESS> - platform: cast name: Playroom host: <IP_ADDRESS> - platform: cast name: Office host: <IP_ADDRESS> - platform: cast name: "Sloane's Speaker" host: <IP_ADDRESS> - platform: cast #Google Home - platform: cast name: Office Home host: <IP_ADDRESS> - platform: cast name: Master Bedroom Home host: <IP_ADDRESS> - platform: cast name: "Sloane's Alarm Clock" host: <IP_ADDRESS> - platform: cast name: Kitchen Home host: <IP_ADDRESS> - platform: cast name: Den Home host: <IP_ADDRESS> #Chromecast Videos - platform: cast name: Den Chromecast host: <IP_ADDRESS> - platform: cast name: Den Shield host: <IP_ADDRESS> - platform: cast name: Playroom Shield host: <IP_ADDRESS> And here's the output from logger with focus on logger: home-assistant.log I can also confirm that https://github.com/home-assistant/home-assistant/pull/16732 has not fixed this. Which makes sense, as that fix refers to it being a problem for 15 days, but I have had this problem for much longer. Current issue is also much older. @benjamin1492, 0.78.1 would not resolve your issue since you are configuring host IP address manually. However, the #16786 may help to fix your problem, I rewrote the device connection logic for your case. 78.1 did not help me either. It lasted about a day, but now I have both my group and my cast devices showing as unavailable. I created a new PR #16804, not sure if it will help, but you can try. #16786 caused more problem than fixed, I closed it. 0.78.3 - same problem - group unavailable Running 0.78.3 this is my log when the cast group 'Home' got unavailable. The cast devices are setup by the discover feature, no IP component configuration: 018-09-24 13:00:48 DEBUG (zeroconf-ServiceBrowser__googlecast._tcp.local.) [homeassistant.components.media_player.cast] Discovered chromecast ChromecastInfo(host='<IP_ADDRESS>', port=42673, uuid='a1883456-632f-4b0b-a2c6-fa6a246f1010', manufacturer='', model_name='Google Cast Group', friendly_name='Home') 2018-09-24 13:00:48 DEBUG (MainThread) [homeassistant.components.media_player.cast] Discovered chromecast with same UUID: ChromecastInfo(host='<IP_ADDRESS>', port=42673, uuid='a1883456-632f-4b0b-a2c6-fa6a246f1010', manufacturer='', model_name='Google Cast Group', friendly_name='Home') 2018-09-24 13:00:48 DEBUG (MainThread) [homeassistant.components.media_player.cast] Disconnecting from chromecast socket. 2018-09-24 13:00:48 INFO (MainThread) [homeassistant.core] Bus:Handling <Event state_changed[L]: entity_id=media_player.home, old_state=<state media_player.home=off; friendly_name=Home, supported_features=21437 @ 2018-09-24T08:56:49.289521+02:00>, new_state=<state media_player.home=unavailable; friendly_name=Home, supported_features=21437 @ 2018-09-24T13:00:48.276523+02:00>> 2018-09-24 13:00:49 DEBUG (Thread-15) [pychromecast.socket_client] Sending: Message urn:x-cast:com.google.cast.tp.connection from sender-0 to receiver-0: {'type': 'CLOSE', 'origin': {}} 2018-09-24 13:00:49 DEBUG (Thread-15) [pychromecast.socket_client] connection listener: 70432210 (CastStatusListener) 2018-09-24 13:00:49 DEBUG (Thread-15) [homeassistant.components.media_player.cast] Received cast device connection status: DISCONNECTED 2018-09-24 13:00:49 DEBUG (MainThread) [homeassistant.components.media_player.cast] Connecting to cast device ChromecastInfo(host='<IP_ADDRESS>', port=42673, uuid='a1883456-632f-4b0b-a2c6-fa6a246f1010', manufacturer='', model_name='Google Cast Group', friendly_name='Home') @awarecan under new PR #16804 the groups still don't come back as per this issue. Confirmed here, as well. Since #16804 only addressed issue of start up hanging in manual configure scenario, rest logic reverted back to 0.77. So if you have group unavailable issue before, you still have it now. I don't have "group" myself, so I won't be able to reproduce or debug that issue, so I am afraid I won't be able to help much. @awarecan this is where you suggest people to buy you two Google Homes so you can create a group 👍 If a donation towards a Ghome Mini or so would help get this bug fixed, I'm in. This is by far my biggest issue with HA right now. If a donation towards a Ghome Mini or so would help get this bug fixed, I'm in. This is by far my biggest issue with HA right now. I would be in as well. I'm in as well, but want to add my two cents on how to proceed. As background, I have a very robust Google cast implementation: 10 Chromacast Audios, 4 Google Homes, an Insignia alarm clock, and a Sony speaker with Chromecast built-in. Ever since Google updated the firmware in Nov-17, this has been a problem. However, none of the Chromecast Built-in devices that remain on old firmware such as the Sony speaker or Insignia alarm clock are impacted. My suggestion is that each of us willing to send equipment to awarecan for this project send something different: a full size Google Home, a Google Home Mini, a Chromecast Audio, and at least 2 Chromecast built-in devices. This would allow awarecan to build a variety of different groups to test the code - group of all built-ins, group of Google devices, and a mixed group. This integration has been a graveyard for many a capable Home Assistant developer just because of it's overall complexity. This approach will give us the best shot at finally getting this thing resolved. I'd even be willing to throw a $100 bounty out there for a final fix, in addition in addition to a Google Home. Awarecan: You mission should you choose to accept it is to fix the Cast platform for good. Shoot me an address and I'll get a Home en route. I have only one google home mini and it regularly (after a few hours) goes to unabailable until I restart HA. Is this the same issue or is this fixed with #16804? Still having this issue on Home Assistant 0.79.3 @awarecan Is there anything we can do to help? Donations, devices, etc? I agree with most people here this desperately needs to be corrected once and for all and I know many of us are willing to help in anyway we can. Dear, someone working on this problem ??? Same problem in Home Assistant 0.80.2 Yeah still experiencing this on 0.80.2 Has no one tried reverting whatever component changed back in March? Nothing has been fixed regarding this issue since it cropped up back then. @j0barr this issue has nothing to do with any changes. The issue has been present since day 1 of Google Cast Groups being available and integrated into HA. It has to do with how the leader of the group (which is comprised of all the devices in the group) is elected. If the leader changes (happens relatively frequent), the group goes unavailable. You can see this with netdisco. More than likely a dev will have to do a big rewrite in order to address this issue. That's odd if nothing has changed because I swear it was working for me back with version 65. You can easily revert back to that version and test. Or set up a dev environment and only add the cast devices. You will see that the groups go unavailable. Trust me, I've been involved in this issue since day 1. Issue goes back way further than that. https://github.com/home-assistant/home-assistant/issues/7782 is when I first started tracking the problem. Looked into this for a bit, going to see if the latest version of zeroconf would fix this. Zeroconf is used for pychromecast which is used for the Cast component. Two fixes from their changelog interest me: Switched from netifaces to ifaddr (pure Python) ServiceBrowser entries are being refreshed when 'stale' now https://github.com/jstasiak/python-zeroconf/blob/master/README.rst The current version used by pychromecast is 0.17.7 @bryanyork Nope, that is unrelated. The issue is that when the cast group changes its elected leader we are currently not re-connecting to the new elected leader in the way the chromecast expects. We do get the callback from zeroconf that it has changed the leader. And we currently do try to disconnect from the old leader and re-connect to the new one. However, something in that process is behaving in a way that the chromecast doesn't expect, and though many hours of debugging I haven't been able to resolve the problem yet. This issue is particularly hard to diagnose because it just happens randomly every few hours, so if you test something you always have to wait that amount of time for the change to trigger again. I've noticed that restarting HA always fixes this problem. What is it about restarting HA that causes this to be fixed, at least temporarily? Is it possible we could make the software mimic whatever a reboot does to get around the issue? I'd tried adding an automation to reboot if cast groups become unavailable but that became unreliable too as sometimes HA wouldn't come back, I've noticed that restarting HA always fixes this problem. What is it about restarting HA that causes this to be fixed, at least temporarily? I'm with Shadex12, a restart will fix it. I was using cast groups as a alarm warning, but with them disappearing (going unavailable) I have temporaily reverted to a single speaker. Can't add much to the discussion except to add my encouragement for a fix. I am on Hassio 0.82.1 and my groups were also "unavailable" after some time in the middle of the night. I have two groups. Each of them were "off" from 22:47 to 02:19. (2018.12.04) Then they were "unavailable" from 02:19 to 03:08. (2018.12.05) Then again they went to status "off" from 03:08 to 04.57. (2018.12.05) Then they went unavailable until now: 04:57 to 22:04. Logs on hassio: https://paste.ubuntu.com/p/J6Nt3pSzH9/ If I can help in any way let me known. Same issue here on HA 0.84.2. Groups change to unavailable every other day and refuses to come back, restarting HA always does the trick. Non here either (0.86.2) I'm affected by the issue as well. I'm looking for a workaround until the issue is resolved. I've seen that restarting Home Assistant does solve the issue, so something is capable of healing this 'unavailable' status in HA. Does anyone know how to reproduce this without a complete restart ? My idea would be to create an automation which does something like: If no cast devices are running and at least 1 cast group is 'unavailable' then reload the chomecast integration (-> heal) Any idea on which service/function to use ? This should be fixed in beta 0.88.0b0 Please give it a try! To provoke errors, try restarting casts that are members of an audio group, either through the Google Home app or by simply power cycling. Note: It's normal that the group goes offline and comes back a few times as group leadership may be handed over several times if the restarted unit is the preferred group leader. In case it's still not working, a complete HA log - from starting HA until group is no longer - is needed for debugging. Logger options in configuration.yaml logger: default: info logs: homeassistant.components.cast: debug homeassistant.components.cast.media_player: debug pychromecast: debug pychromecast.discovery: debug pychromecast.socket_client: debug zeroconf: debug Hi, I pulled Home assistant Dev 0.89 in a docker container (fresh config, host networking) and used discovery to autoconfigure my chromecasts and groups. They are discovered but all 'unavailable'. The attached log shows "Failed to connect" errors. https://paste.ubuntu.com/p/4BRQ3hhQBC/ @masterz666 That should be fixed in beta 0.88.0b2 @emontnemery I'm on 0.88.0b0 since Thursday and didn't had this issue since then. Seems like it really fixed it (at least for me) 🎉 Thanks for you work! 👍 on 0.88b2 I sometime have this in the log. is this expected ? [Chromecast salon:8009] Heartbeat timeout, resetting connection I don't want to jump the gun but I think the new pychromecast update may have solved this issue! And mainly due to a experimental commit. https://github.com/balloob/pychromecast/pull/268. Crossing fingers but this is very promising! @edif30, @soldag Thanks for the feedback! balloob/pychromecast#268 is included in HA 0.88 via #21097 so I'll go ahead and close this issue. @masterz666 The print means HA lost the connection to a chromecast. At least for me, chromecasts are not extremely stable and reboot now and then, so yes, I'd say its expected. Does anyone on this thread have a Sony or Insignia speaker with Chromecast Built-in? If so, have you noticed them causing any issues after updating to the new pychromecast? The new version still isn't working for me, which I think makes me the exception based on the recent feedback. I'm now trying to troubleshoot my cast install, and starting with the less common devices. semi-unrelated: @ttaidapos can you post your yaml for those media player cards that you posted pics of? Thanks. For what it's worth. I'm on 88.2 and the group casting still going strong! Still have a plethora of errors though. Line 718: 2019-03-03 04:58:09 WARNING (Thread-12) [pychromecast.socket_client] [Baby display:8009] Heartbeat timeout, resetting connection Line 1807: 2019-03-04 02:01:09 WARNING (Thread-2) [pychromecast.socket_client] [Living Room Speaker:8009] Heartbeat timeout, resetting connection Line 1828: 2019-03-04 02:22:58 WARNING (Thread-7) [pychromecast.socket_client] [Office display:8009] Heartbeat timeout, resetting connection Line 1829: 2019-03-04 02:28:16 WARNING (Thread-9) [pychromecast.socket_client] [Kitchen Speaker:8009] Heartbeat timeout, resetting connection Line 1830: 2019-03-04 02:28:23 WARNING (Thread-6) [pychromecast.socket_client] [Bedroom Speaker:8009] Heartbeat timeout, resetting connection Line 1836: 2019-03-04 02:34:31 WARNING (Thread-15) [pychromecast.socket_client] [Bathroom Speaker:8009] Heartbeat timeout, resetting connection Line 1873: 2019-03-04 03:17:51 WARNING (Thread-12) [pychromecast.socket_client] [Baby display:8009] Heartbeat timeout, resetting connection Line 1894: 2019-03-04 03:42:43 WARNING (Thread-5) [pychromecast.socket_client] [Soundbar Speaker:8009] Heartbeat timeout, resetting connection Line 1944: 2019-03-04 04:36:27 WARNING (Thread-10) [pychromecast.socket_client] [Dining Room Speaker:8009] Heartbeat timeout, resetting connection Line 1988: 2019-03-04 05:26:33 WARNING (Thread-14) [pychromecast.socket_client] [Kitchen display:8009] Heartbeat timeout, resetting connection semi-unrelated: @ttaidapos can you post your yaml for those media player cards that you posted pics of? Thanks. It's a lot and I take no credit for it. I have modified it the yaml a lot. here it is and best of luck! https://pastebin.com/RbWJVNhH I lose cast groups still with 0.97.2...
2025-04-01T06:38:58.192616
2018-04-13T08:03:34
314006938
{ "authors": [ "elbogi", "jimbob1001" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6787", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/13846" }
gharchive/issue
0.67.0b1 - BT mac addresses get prefixed with double underscore in known_devices.yaml Home Assistant release with the issue: 0.67.0b1 Last working Home Assistant release (if known): Unknown Operating environment (Hass.io/Docker/Windows/etc.): Docker Component/platform: Bluetooth device tracker Description of problem: BT devices are getting a double underscore in the mac address in known_devices.yaml mac: BT__00:11:22:AA:BB:CC This causes an error as it thinks that the second underscore is part of the mac address Correct mac address... mac: BT_00:11:22:AA:BB:CC Problem-relevant configuration.yaml entries and (fill out even if it seems unimportant): Traceback (if applicable): Additional information: When removing the rogue underscore, or making any other changes in known_devices.yaml the device is rediscovered again, with a double underscore again Confirming the same behaviour running 0.67b1 in hassio. When a device is added as a new one, it shows up in known_devices.yaml as: my_iphone: hide_if_away: false icon: mac: BT__40:xx:xx:xx:xx:DB name: my iPhone picture: track: true If I change my existing device but suffix BT with an extra _ the bluetooth device tracker gives the following error: [homeassistant.components.device_tracker.bluetooth_tracker] Error looking up Bluetooth device Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/homeassistant/components/device_tracker/bluetooth_tracker.py", line 92, in update_bluetooth result = bluetooth.lookup_name(mac, timeout=5) File "/usr/lib/python3.6/site-packages/bluetooth/bluez.py", line 59, in lookup_name raise BluetoothError ("%s is not a valid Bluetooth address" % address) bluetooth.btcommon.BluetoothError: _40:xx:xx:xx:xx:DB is not a valid Bluetooth address Can, however confirm, that bluetooth tracker works if you delete your previously added BT-devices and let it recreate them in known_devices.yaml. This issue is directly related to this PR: https://github.com/home-assistant/home-assistant/pull/12458 Appears to be working correctly again in 0.67.0
2025-04-01T06:38:58.194919
2018-10-03T19:16:39
366487676
{ "authors": [ "fabaff", "simonvanderveldt" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6788", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/17112" }
gharchive/issue
Tap on slider on mobile triggers two state changes/events/messages Home Assistant release with the issue: v 0.78.3 Description of problem: When using home assistant on a mobile (probably also on a tablet) tapping a slider, for example for brightness, triggers two times the same state change/event/message. In my specific case I'm using MQTT and it results in 2x the same MQTT message being published. When moving the slider and then releasing it only a single state change/event/message is created/sent. This sounds like a frontend issue. /move to home-assistant/home-assistant-polymer
2025-04-01T06:38:58.200878
2019-09-05T08:13:26
489597139
{ "authors": [ "simonszu", "springstan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6789", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/26445" }
gharchive/issue
Entity property configuration: The following attributes are already set in customize.yaml Home Assistant release with the issue 0.98.2 Last working Home Assistant release (if known): unknown Operating environment (Hass.io/Docker/Windows/etc.): Docker on Debian Busterb Component/platform: Entity configuration in "Settings -> Configuration" Description of problem: See https://community.home-assistant.io/t/entity-property-configuration-the-following-attributes-are-already-set-in-customize-yaml/135444 Problem-relevant configuration.yaml entries and (fill out even if it seems unimportant): See the linked forum thread above, i have provided some snippets there. Traceback (if applicable): not applicable Anyone? This is still a problem on 0.100 for me. Is this still an issue? I have read through your forum thread and would suggest that you try out the given solution: homeassistant: customize: !include customize.yaml You may be right about YAML indentation, however I would always follow the official documentation. In this case: customize entities I have solved it by moving the "customize" dictionary key to the beginning of the dict. It's strange since yaml dicts shouldn't be ordered, but it works for me now. Oh okay that is great to hear 👍 Please mark this issue as solved :)
2025-04-01T06:38:58.213980
2017-03-03T00:49:12
211563790
{ "authors": [ "370gt", "KptnKMan", "balloob", "iambdud", "markferry", "philhawthorne" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6790", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/6366" }
gharchive/issue
hass becomes unresponsive after a few hours Make sure you are running the latest version of Home Assistant before reporting an issue. You should only file an issue if you found a bug. Feature and enhancement requests should go in the Feature Requests section of our community forum: Home Assistant release (hass --version): 0.39.2 Python release (python3 --version): 3.5.3 Component/platform: HASS Docker image Description of problem: after a few hours, the docker container and hass are unresponsive. The front end shows disconnected. the log no longer reports anything. The log is full of errors like this: 17-03-02 02:27:39 ERROR (Thread-3) [homeassistant.components.sensor.command_line] Timeout for command: python3 /config/sensor/getBIProfile.py 17-03-02 02:27:39 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting 17-03-02 02:27:44 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting 17-03-02 02:27:48 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting 17-03-02 02:27:52 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting 17-03-02 02:28:05 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting 17-03-02 02:28:11 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting 17-03-02 02:28:32 WARNING (Thread-13) [pychromecast.socket_client] Heartbeat timeout, resetting connection 17-03-02 02:28:50 WARNING (Thread-17) [pychromecast.socket_client] Error communicating with socket, resetting connection Full log here: https://hastebin.com/gabikeyayo.sql Expected: Problem-relevant configuration.yaml entries and steps to reproduce: Traceback (if applicable): Additional info: Also noticed this running on Docker as well. Think it is related to #6253 with too many events causing things to slow down to a crawl, and eventually unresponsive. This has nothing to do with too many events. I think that something is causing high CPU usage, not sure what it could be. @balloob When #6369 goes live, are there any specific logs/experiments you'd like if I can replicate? I'm not using the Z-wave component... so it can't be that for me. I am using nmap device tracker and/or asus router device tracker. I've tried using only one or the other and the issue persists. Could it be either of those? Anything you want to know from my setup or anything you want me to do? I turned off both nmap and Asus device trackers and I haven't seen the timer out of sync error in a while. I spoke too soon. It crashed overnight :( Same issue. Previously running 0.35.3 which was solid. Reverting now to 0.39.2 to reproduce. Reproduced within hours. Back to 0.38.4. @markferry did you use pip to downgrade? Nope, setup.py. Working from git. In gdb: #0 0x76f51b30 in __pthread_cond_timedwait (cond=0x3c3f68, mutex=0x3c3f98, abstime=0x7ef0d4a4) at pthread_cond_timedwait.c:199 which I'll admit is not exactly surprising. Backtrace for all threads: bt.txt Almost all waiting on OpenZWave. 0.37.0dev0 lasted longer but still died. Should probably also drop my DB tables before the next run. Back to 0.36.0dev0 (64800fd). Died again. Bisected back to 0.36.0.dev0 (c14a5fa) now. If this proves fruitless there's also bisecting the config. What joy. I'm getting the same issue in a docker as well! It is driving me nuts! These issues have been resolved. Don't want to create a catch all issue so closing this.
2025-04-01T06:38:58.222963
2017-07-17T10:26:46
243352611
{ "authors": [ "amelchio", "rossffs" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6791", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/8520" }
gharchive/issue
Flux LED - Scene with multiple platforms cannot be called while Flux controller is powered off Make sure you are running the latest version of Home Assistant before reporting an issue. You should only file an issue if you found a bug. Feature and enhancement requests should go in the Feature Requests section of our community forum: Home Assistant release (hass --version): 0.49.0 Python release (python3 --version): 3.5.2 Component/platform: Flux LED Description of problem: When HA starts and the Flux LED device controller is powered off, the error: ERROR (Thread-1) [homeassistant.components.light.flux_led] Failed to connect to bulb <IP_ADDRESS>, TV LEDs is shown in the info panel and the light entity is not created. If this entity is part of a HA Scene with other platforms/entities the scene can no longer be called. WARNING (MainThread) [homeassistant.helpers.state] reproduce_state: Unable to find entity light.tv_leds ERROR (MainThread) [homeassistant.core] Invalid service data for light.turn_on: two or more values in the same group of exclusion 'Color descriptors' @ data[<Color descriptors>]. Got None In my configuration the scene in question contains a Hue bulb and a Flux LED strip. Expected: The Hue bulb to respond to the scene call while the Flux LED controller is powered off. Problem-relevant configuration.yaml entries and steps to reproduce: - platform: hue host: <IP_ADDRESS> allow_unreachable: true - platform: flux_led devices: <IP_ADDRESS>: name: TV LEDs scene.yaml - name: Warm entities: light.lamp: state: on brightness: 223 color_temp: 443 xy_color: [ 0.5015, 0.4153 ] light.tv_leds: state: on brightness: 60 rgb_color: [ 255, 123, 9 ] Start/restart HA with the Flux LED controller powered off. Call the scene "Warm" either from an automation or the web interface. Traceback (if applicable): Additional info: Prior to version 0.44.0 (approx) the Scene worked as expected with the Flux LED controller powered off. The Yeelight platform creates the light entity with a state of "unavailable" if HA is started while the controller is powered off. This allows scenes containing other platforms to be successfully used. You can no longer have color_temp and xy_color in the same setting so try removing one of them from light.lamp. I removed color_temp which allows the scene to work. Not quite what I expected because my original config above works when the Flux LED controller is powered on and available in HA. That does sound unexpected. Would you mind double-checking that the color_temp/ xy_color combination really does work in that scenario (with a current HA)? If so, I will try replicating your configuration. I just double checked and I can confirm that when HA starts with the Flux LED Controller powered off, the Scene with both color_temp and xy_color does not work. When HA starts with the Flux LED Controller powered on the the Scene does work with both keys. By removing either of the keys the Scene works with the Flux LED Controller powered off. I have tried your scene now, it always fails for me (as expected). However, I only have LIFX to test with. As I cannot reproduce the inconsistent behavior, I will not try fixing it :)
2025-04-01T06:38:58.231507
2017-10-23T20:35:51
267808589
{ "authors": [ "fabaff", "fanaticDavid", "homeassistant", "ufulu" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6792", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/10092" }
gharchive/pull-request
Added website authentication Description: Added website authentication so you can also scrape data which is behind a login. Related issue (if applicable): fixes # Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here> Example entry for configuration.yaml (if applicable): - platform: scrape resource: http://example.com name: Website-auth-scrape username: user password: password authentication: basic select: 'h1' Checklist: If user exposed functionality or configuration variables are added/changed: [x ] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [ ] New dependencies have been added to the REQUIREMENTS variable (example). [ ] New dependencies are only imported inside functions that use them (example). [ ] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [ ] New files were added to .coveragerc. If the code does not interact with devices: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [ ] Tests have been added to verify that the new code works. Hi @ufulu, It seems you haven't yet signed a CLA. Please do so here. Once you do that we will be able to review and accept this pull request. Thanks! Isn't this a duplicate of #10079 ? Duplicate of #10079. To avoid exactly this, did I link the Pull Request in the forum thread.
2025-04-01T06:38:58.240988
2017-11-20T06:48:15
275252577
{ "authors": [ "andrey-git", "cgtobi", "pvizeli" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6793", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/10699" }
gharchive/pull-request
Cast return values and add unit tests. Description: Cast return values according to their type returned from yahoo. Add unit test for yahoo weather component. Related issue (if applicable): fixes # Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here> Example entry for configuration.yaml (if applicable): Checklist: If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [ ] New dependencies have been added to the REQUIREMENTS variable (example). [ ] New dependencies are only imported inside functions that use them (example). [ ] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [ ] New files were added to .coveragerc. If the code does not interact with devices: [x] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [x] Tests have been added to verify that the new code works. Could you write the tests in a way that mocks out the library instead of mocking out the network response and using the library? @andrey-git I can give it a try. What would be the benefit of that? Everyone who runs test won't need to install the library. Also the tests will be more hermetic - if library implementation would change without changing the providing API - the tests won't break. I think you no longer need the json file and can just return whatever values needed in the mock. @pvizeli maybe the lib should return numeric values in the first place? @andrey-git what is wrong with have a whole set of data is a separate file instead of small chunks of data inline? Is there a special reason? The json could eventually be reused for testing the yahoo weather sensor component. I don't feel strongly about this. Lets's wait for @pvizeli as lib author @fabaff ? @pvizeli Are you talking about the lib returning numeric values rather than strings? I added a temporary fix to work around the yahoo weather API issue (#10640).
2025-04-01T06:38:58.251373
2017-01-29T14:51:36
203874203
{ "authors": [ "Danielhiversen", "FREImedia", "homeassistant", "robbiet480" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6794", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/5622" }
gharchive/pull-request
Restore input_* states on startup Description: This PR enables HASS to restore the last state of an input_boolean, an input_slider and an input_select on a restart. This is pretty useful if you have e.g. input booleans or sliders for dynamic settings on your HASS. Example: Let's say you have an alarm clock setup with hass. The target time ist set by an input_slider and will now survive a reboot. The feature is optional and can be set by restore: startup for each entity of the named domains above. I have chosen 'restore' as the attribute, because you could re-use it for further functionalities/implementations later on (e.g. "reset value after 24hours") Related issue (if applicable): fixes # Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#1911 Example entry for configuration.yaml (if applicable): input_boolean: is_home: name: 'Is Home' initial: on nightmode: name: 'Alarm Clock enabled' initial: off restore: startup input_select: radio_channel: name: 'Radio Channel' options: - 'a' - 'b' initial: 'a' icon: 'mdi:import' restore: startup input_slider: alarm_hour: name: 'Alarm Hour' icon: 'mdi:timer' initial: 8 min: 0 max: 23 step: 1 restore: startup Checklist: If user exposed functionality or configuration variables are added/changed: [x] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [ ] New dependencies have been added to the REQUIREMENTS variable (example). [ ] New dependencies are only imported inside functions that use them (example). [ ] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [ ] New files were added to .coveragerc. If the code does not interact with devices: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [ ] Tests have been added to verify that the new code works. Hi @FREImedia, It seems you haven't yet signed a CLA. Please do so here. Once you do that we will be able to review and accept this pull request. Thanks! Duplicate of https://github.com/home-assistant/home-assistant/pull/4614 Hey, sorry the CLA link was wrong, it's now been fixed. Please sign the CLA here. I am closing this in favor of #4614
2025-04-01T06:38:58.288355
2017-02-13T14:22:19
207231759
{ "authors": [ "aronsky", "balloob", "fabaff", "homeassistant", "pvizeli", "rpitera" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6795", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/5948" }
gharchive/pull-request
Add 'entity_picture' property to DarkSkySensor component Description: Related issue (if applicable): fixes # Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here> Example entry for configuration.yaml (if applicable): Checklist: If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [x] New dependencies have been added to the REQUIREMENTS variable (example). [x] New dependencies are only imported inside functions that use them (example). [x] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [x] New files were added to .coveragerc. If the code does not interact with devices: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [ ] Tests have been added to verify that the new code works. Hi @aronsky, It seems you haven't yet signed a CLA. Please do so here. Once you do that we will be able to review and accept this pull request. Thanks! I attempted to run tox tests, and had one failure (in the test_darksky.py module, suspicious!) - but this failure seems to happen prior to my changes, as well. Looks like something about a mock load_fixture, which has nothing to do with the changes in this pull request. I don't think that is a good idea to add base64 icons to code. But I can also not say what the alternative sultion is. I'm not a frontend developer. CC @balloob @pvizeli I agree that it's not perfect, far from it. I just wanted it to be both self-contained, and not dependent on 3-rd party repositories, which might change and/or disappear at any moment, breaking the module. I looked around on the Dark Sky website, hoping to find some sort of permalinks to the icons, but their official icons are dynamic and use Javascript - I don't know how to cleanly integrate those into Home Assistant. If there's a more elegant solution, I'll be happy to do it. One idea of the weather component was to be able to display the conditions in a visual way. Unfortunately the frontend part is not done yet. @aronsky Could you use the Skycons library? It's self contained JS, public domain and created for DarkSky. Plus the icons are animated. https://github.com/darkskyapp/skycons @rpitera - yep, I saw those icons. They definitely look cool - but integrating them requires changes in the front-end, and without that - I have no idea how a specific module can influence the front-end extensively (and this is required, since the front-end doesn't currently support Javascript-based icons) I'm btw a big fan of the skycons. We should consider adding those to the frontend. OK. So, updating the frontend proved to be more complicated than I thought. I have, in fact, made a small patch to the frontend, that would show the state's icon property (instead of the currently used entity_picture). But this behavior is only desirable for weather forecast sensors. Unfortunately, by the time the state arrives to the frontend, there's nothing to differentiate it from other sensors, even other monitored conditions from the weather sensor (such as the temperature, which should display the value, not the icon). As a temporary hack, I enabled this only for sensors that have an icon containing the word weather, but obviously it's not a proper solution. I ended up doing what @balloob suggested for now, and will try to come up with a more robust solution to the frontend in the future. Speaking of the frontend, what are the guidelines to making changes to that component, since it compiles into files that are part of this repo? Do I make a pull request in the frontend's repository, and once it is accepted, compile and make a pull request here? You just make PRs against the frontend repo. I add new builds from frontend to this repo. Please add your images in a subfolder under images. Also a heads up that this is the last change that we will allow for the darksky sensor. Future efforts will have to be focussed on the weather component. This sensor will be removed in a future release. I had no idea there's a weather component in the works - I didn't see anything in the documentation. I'll look into it (it does make sense to have those unified somehow, as opposed to different sensors with different images and weather states). I'll move the images to a subfolder. And finally, regarding the frontend repo. I understand that this might be a headache to keep track of, but any chance you can tag frontend commits with versions matching the home-assistant tags that use those commits? It's a lot less clear on which version to work on in that repo. Every Home Assistant commit already contains a version of Home Assistant Polymer via a submodule: https://github.com/home-assistant/home-assistant/tree/dev/homeassistant/components/frontend/www_static That being said, you still always have to work on the latest version no matter what commit it is based off. Can you remove the png versions since they are not being used ? You also broke the tests. Hello @aronsky, When attempting to inspect the commits of your pull request for CLA signature status among all authors we encountered commit(s) which were not linked to a GitHub account, thus not allowing us to determine their status(es). The commits that are missing a linked GitHub account are the following: 01f346ccee1f62f354b0ee239ee9c42c38a43437 - No email found attached to the commit. b2cbde31945d9bf0e9ccc87c3e9259da781cde1a - No email found attached to the commit. Unfortunately, we are unable to accept this pull request until this situation is corrected. Here are your options: If you had an email address set for the commit that simply wasn't linked to your GitHub account you can link that email now and it will retroactively apply to your commits. The simplest way to do this is to click the link to one of the above commits and look for a blue question mark in a blue circle in the top left. Hovering over that bubble will show you what email address you used. Clicking on that button will take you to your email address settings on GitHub. Just add the email address on that page and you're all set. GitHub has more information about this option in their help center. If you didn't use an email address at all, it was an invalid email, or it's one you can't link to your GitHub, you will need to change the authorship information of the commit and your global Git settings so this doesn't happen again going forward. GitHub provides some great instructions on how to change your authorship information in their help center. If you only made a single commit you should be able to rungit commit --amend --author="Author Name<EMAIL_ADDRESS> (substituting Author Name and<EMAIL_ADDRESS>for your actual information) to set the authorship information. If you made more than one commit and the commit with the missing authorship information is not the most recent one you have two options: You can re-create all commits missing authorship information. This is going to be the easiest solution for developers that aren't extremely confident in their Git and command line skills. You can use this script that GitHub provides to rewrite history. Please note: this should be used only if you are very confident in your abilities and understand its impacts. Whichever method you choose, I will come by to re-check the pull request once you push the fixes to this branch. We apologize for this inconvenience, especially since it usually bites new contributors to Home Assistant. We hope you understand the need for us to protect ourselves and the great community we all have built legally. The best thing to come out of this is that you only need to fix this once and it benefits the entire Home Assistant and GitHub community. Thanks, I look forward to checking this PR again soon! :heart: Hello @aronsky, When attempting to inspect the commits of your pull request for CLA signature status among all authors we encountered commit(s) which were not linked to a GitHub account, thus not allowing us to determine their status(es). The commits that are missing a linked GitHub account are the following: 01f346ccee1f62f354b0ee239ee9c42c38a43437 - No email found attached to the commit. b2cbde31945d9bf0e9ccc87c3e9259da781cde1a - No email found attached to the commit. Unfortunately, we are unable to accept this pull request until this situation is corrected. Here are your options: If you had an email address set for the commit that simply wasn't linked to your GitHub account you can link that email now and it will retroactively apply to your commits. The simplest way to do this is to click the link to one of the above commits and look for a blue question mark in a blue circle in the top left. Hovering over that bubble will show you what email address you used. Clicking on that button will take you to your email address settings on GitHub. Just add the email address on that page and you're all set. GitHub has more information about this option in their help center. If you didn't use an email address at all, it was an invalid email, or it's one you can't link to your GitHub, you will need to change the authorship information of the commit and your global Git settings so this doesn't happen again going forward. GitHub provides some great instructions on how to change your authorship information in their help center. If you only made a single commit you should be able to rungit commit --amend --author="Author Name<EMAIL_ADDRESS> (substituting Author Name and<EMAIL_ADDRESS>for your actual information) to set the authorship information. If you made more than one commit and the commit with the missing authorship information is not the most recent one you have two options: You can re-create all commits missing authorship information. This is going to be the easiest solution for developers that aren't extremely confident in their Git and command line skills. You can use this script that GitHub provides to rewrite history. Please note: this should be used only if you are very confident in your abilities and understand its impacts. Whichever method you choose, I will come by to re-check the pull request once you push the fixes to this branch. We apologize for this inconvenience, especially since it usually bites new contributors to Home Assistant. We hope you understand the need for us to protect ourselves and the great community we all have built legally. The best thing to come out of this is that you only need to fix this once and it benefits the entire Home Assistant and GitHub community. Thanks, I look forward to checking this PR again soon! :heart: I'm not sure what broke the build... Any chance it has been broken in the dev branch that I merged into my code? It seems like it, because that's the first commit that seems to break the build. I merged the current dev branch, and hopefully it'll work - I'm not on my work computer, and can't run tox from here. If it still fails, it'll have to wait until tomorrow. P.S. Sorry for the mess with the commits without e-mail addresses, I haven't used my home computer for development before, and didn't notice that git hasn't had an e-mail set up. Seeing that this is using a subfolder of SVG icons, would this support putting in your own SVG icons as long as the naming matched the DarkSky naming conventions? Also, if you replaced them, would they get overwritten on subsequent HA updates? I'm assuming no, since this will be the final update (barring hot-fixes of course) allowed for this platform. @rpitera yes, if this pull request is accepted, you'll be able to switch to your own SVGs, as long as you replace the files (i.e., the names match). Your assumption about the subsequent updates sounds correct to me, but your guess is as good as mine in that regard. Thanks! It seems that you have added some commits to your branch unrelated to this change. Please clean it up. Hi @balloob, I'm not sure which commits you're referring to. There were a couple of commits I've pushed without properly signing them with my email, and I had to do some git history rewriting to fix that. Other than that, there are only commits relevant to the change, and the merges from dev I did from time to time to stay up to date with the code. If that's not something I should've done, I'll avoid it in the future, but at this point I don't know how to unravel those merges - and not sure how critical it is, since those are all commits accepted into the dev branch, anyway. Do you prefer that I redo my changes in a separate, clean pull request, and abandon this one? Have a look at commits or files changed to see what I mean: Yup, that's the mess with the authoring I referred to previously, sorry about that. I'm not sure how to fix that properly - if I start afresh, rebase my dev branch on the original commit where it diverged from the upstream repo, cherry-pick the commits with my changes, and then push - will it clean everything up and avoid the need for a new pull request? Again, sorry about the mess. Yes, that will work. You will have to force-push to overwrite the bad commits in your current branch. To be safe, you can also create a new branch that branches off latest dev and cherry-pick your changes. That sounds like the best idea. Will there be an option to update this pull request to point to the new branch, or do I check its correctness and then force-push it onto the current branch in question? Can't change the source branch. It will be easiest to just close this PR and open a new one. I'll do that. Once again, sorry about the mess with the branches and stuff. The new pull request, #6141, is just one clean commit :)
2025-04-01T06:38:58.291219
2017-06-01T20:31:46
233001470
{ "authors": [ "balloob", "danielperna84" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6796", "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/7861" }
gharchive/pull-request
Updated pyhomematic Description: Updating pyhomematic dependency to fix wired devices. Related issue (if applicable): fixes #7708 @danielperna84 you are allowed to merge dependency upgrade PRs yourself 👍
2025-04-01T06:38:58.293601
2024-08-01T20:09:00
2443320405
{ "authors": [ "TheFes", "tetele" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6797", "repo": "home-assistant/intents", "url": "https://github.com/home-assistant/intents/pull/2362" }
gharchive/pull-request
Add some more light control sentences Fix #2319 Summary by CodeRabbit New Features Introduced new commands for controlling lights, allowing users to specify actions like turning lights off or on in designated areas, as well as flexible phrasing without area specifications. Tests Added test cases to ensure the system recognizes new phrases for turning lights off and on, enhancing the robustness of the natural language processing for lighting commands. Looks good
2025-04-01T06:38:58.345769
2022-06-17T20:31:21
1275431083
{ "authors": [ "calintat", "jamievicary", "regular-citizen" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6798", "repo": "homotopy-io/homotopy-rs", "url": "https://github.com/homotopy-io/homotopy-rs/issues/509" }
gharchive/issue
Compile a fast LP solver to Wasm The best LP solvers such as CBC or Highs are written in C/C++. So if we want to use them in our project (#508), we need to compile them in Wasm using emscripten. Nick made some progress on this. This is great. Has it improved the performance of the layout algorithm? To be determined, there are some issues to address with layout tests that currently depend on minilp to address before we can move to good_lp+highs. Thanks Chiara. So is this new solver currently being used in the production layout code? On Fri, 29 Jul 2022, 14:38 Chiara Sarti, @.***> wrote: To be determined, there are some issues to address with layout tests that currently depend on minilp to address before we can move to good_lp+highs. — Reply to this email directly, view it on GitHub https://github.com/homotopy-io/homotopy-rs/issues/509#issuecomment-1199297520, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACQ4OHUO6U3OZYEJLXGFCMTVWPNFTANCNFSM5ZDK2OMQ . You are receiving this because you commented.Message ID: @.***> Not yet, but the idea is to either use this solver only for production or to just bin the tests altogether. I think @calintat is the best position to figure out what to do on this front, but once that is sorted moving to highs should be relatively straightforward.
2025-04-01T06:38:58.356593
2021-08-05T01:49:44
961332621
{ "authors": [ "coda1997", "honjow" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6799", "repo": "honjow/FEhViewer", "url": "https://github.com/honjow/FEhViewer/issues/7" }
gharchive/issue
Flutter run error: Could not complete submission of dSYM I meet this issue when I am trying running the code to deploy in my iPhone. Here is the detail info: Project /Users/dadachen/workspace/FEhViewer built and packaged successfully. Uploading dsym file error: Could not complete submission of dSYM at /Users/dadachen/workspace/FEhViewer/build/ios/Debug-iphoneos/fehviewer.app.d SYM: Error Domain=com.crashlytics.mac.error-domain.process-dsym Code=2 "File no longer exists at (null)" UserInfo={NSLocalizedFailureReason=File no longer exists at (null)} Command PhaseScriptExecution failed with a nonzero exit code note: Using new build system note: Building targets in parallel note: Planning build note: Analyzing workspace note: Constructing build description note: Build preparation complete Could not build the precompiled application for the device. I do not have any ideas of that. Could you give me a help? I have replaced the google firebase plist with mine. So can I do something to solve this? You can try to configure Runner.xcodeproj in xcode. left click on TARGETS, Runner, right select Build Phases , expand Run Script, comment out this line "${PODS_ROOT}/FirebaseCrashlytics/upload-symbols" -gsp ${GOOGLESERVICE_INFO_LOCATION} -p ios "${DWARF_DSYM_FOLDER_PATH}/${DWARF_DSYM_ FILE_NAME}" There may be something wrong with this script, I'll try to fix it It works for me. Thank you so much! ☺️
2025-04-01T06:38:58.359265
2024-06-29T14:16:12
2381775463
{ "authors": [ "3w36zj6", "chrisyalamov", "yusukebe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6800", "repo": "honojs/honox", "url": "https://github.com/honojs/honox/issues/197" }
gharchive/issue
Add documentation on how to use HonoX outside of Cloudflare What is the feature you are proposing? Projects created by selecting x-basic with npm create hono@latest support Cloudflare Workers & Pages, but it would be beneficial to have documentation on how to adapt the code to run on other platforms. I don't think this is currently possible as the Vite plugin built into HonoX is only adapted for Cloudflare Pages. There was some discussion on this in #42 where someone built their own plugin. If not, I'm not sure if it would be helpful, but Cloudflare have open-sourced the runtime for their Workers (https://github.com/cloudflare/workerd) so if you're looking to self-host, that might be helpful. Hi. You can run the HonoX app on other runtimes like Node.js and Bun instead of Cloudflare/workerd. To do it, you must build your app with the proper Vite settings. It would be good if the setting were written in the README.
2025-04-01T06:38:58.368405
2021-05-30T09:20:22
906692824
{ "authors": [ "dragondove", "ful1e5", "oblitum", "shadmansaleh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6801", "repo": "hoob3rt/lualine.nvim", "url": "https://github.com/hoob3rt/lualine.nvim/issues/268" }
gharchive/issue
Bug: lualine goes blank after calling NvimTreeToggle Self Checks [X] I'm using the latest lualine and neovim nightly. [X] I didn't find the issue in exsisting issues or prs. How to reproduce the problem install both nvim-tree and lualine plugins. open a file in nvim. call :NvimTreeToggle. lualine gets into the tree buffer. call :NvimTreeToggle again. lualine shows blank in current buffer. Actual behaviour lualine goes blank. Expected behaviour lualine should appears again after toggle nvimtree. Aditional information Minimal config to reproduce the issue Config This happens if any buffer where lualine is deactivated is focused and gets closed while focused. @dragondove Which colorscheme are you using? @dragondove Which colorscheme are you using? I'm using zephyr . Does it relate to this issue? The config of lualine I used is from evil_lualine. what's most likely happening is the active status isn't being enabled due to none of the event we look for being triggured We probably need aditional event to catch this case here: https://github.com/hoob3rt/lualine.nvim/blob/9726824f1dcc8907632bc7c32f9882f26340f815/lua/lualine/init.lua#L173 @dragondove evil_lualine doesn't configure inactive_status that's why you're probably seeing lualine disapear . Does leaving the window and entering it again bring back lualine ? what's most likely happening is the active status isn't being enabled due to none of the event we look for being triggured We probably need aditional event to catch this case here: https://github.com/hoob3rt/lualine.nvim/blob/9726824f1dcc8907632bc7c32f9882f26340f815/lua/lualine/init.lua#L173 @dragondove evil_lualine doesn't configure inactive_status that's why you're probably seeing lualine disapear . Does leaving the window and entering it again bring back lualine ? yes, If I firstly switch to the editing buffer(By C-w w), lualine will be there. It seems this bug disapears, but I don't know why, maybe it is caused by a specific version of neovim. Anyway, it works fine now.
2025-04-01T06:38:58.431489
2018-11-06T17:39:38
377958740
{ "authors": [ "HopsworksJenkins", "berthoug", "maismail" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6803", "repo": "hopshadoop/hops", "url": "https://github.com/hopshadoop/hops/pull/568" }
gharchive/pull-request
[HOPS-808] Fix performance degradation due to safemode and enable nam… …enode resource monitoring Make sure there is no duplicate PR for this issue Please check if the PR fulfills these requirements [ ] Tests for the changes have been added and passed (for bug fixes / features) [x] HOPS JIRA issue has been opened for this PR [x] All commits have been squashed down to a single commit [x] The commit message has the following format: [HOPS-XXX] message Post a link to the associated JIRA issue https://hopshadoop.atlassian.net/browse/HOPS-808 What kind of change does this PR introduce? (Bug fix, feature, docs update, ...) What is the new behavior (if this is a feature change)? Does this PR introduce a breaking change? (What changes might users need to make in their application due to this PR?) Other information: https://github.com/hopshadoop/hops-metadata-dal/pull/137 https://github.com/hopshadoop/hops-metadata-dal-impl-ndb/pull/170 Can one of the admins verify this patch? @berthoug i was benchmarking hopsFS with 4 namenodes and i found that my solution for avoiding reading the safemode variable from the database at every request only works for one namenode. So i fixed this one, and pushed it. There is some errors in this pull request (ex: "<<<<HEAD" ;) ) and I think I identified some potential bugs. I fixed them in the following pull request #580.
2025-04-01T06:38:58.444030
2020-09-08T14:21:25
695915734
{ "authors": [ "adank92", "hopsoft", "leastbad" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6804", "repo": "hopsoft/cable_ready", "url": "https://github.com/hopsoft/cable_ready/pull/71" }
gharchive/pull-request
Removed isEqualNode comparison Proposal to fix https://github.com/hopsoft/cable_ready/issues/70 I think we should retain the isEqualNode check for performance reasons; however, I'd be ok to skip this check for SELECT and OPTION elements. @adank92 Do you think this compromise would address the problems you faced in #70? That is a clever suggestion, if the issue is restricted to just select/option. @hopsoft @leastbad that would work for the first issue, but for the second one we would also need to skip INPUT Talk to me like I'm distracted and less confident... what is it about form elements that they need special care? Is it because the user can change their state? Should TEXTAREA be added to the list? Are there others we're not thinking about because "nobody" ever uses them? What are the gotchas? eg. what about the input type="numeric" that is forced to display the value abc, putting it in violation of the half-baked pattern validation built into some elements? is "equal but not valid" something that matters in this context? I'm just trying to brainstorm and still feeling haunted by the abstract thinking required to get to an outcome on https://github.com/patrick-steele-idem/morphdom/pull/206 Talk to me like I'm distracted and less confident... what is it about form elements that they need special care? Is it because the user can change their state? SELECT elements need to have one option with the selected attribute set in order for morphdom to work as expected. The library will ensure this by running code to sync the attribute (syncBooleanAttrProp(fromEl, toEl, "selected");). Using isEqualNode prevents this from happening. This is explained more in detail in the original issue #70 INPUT elements do not update their value attribute in the HTML definition when users type in it. If the page is loaded with an empty input, the user types in something, and then the backend returns an empty input again, ìsEqualNode will always return true because both HTML elements are still the same. Should TEXTAREA be added to the list? Yes, TEXTAREA has the same issue as INPUT. If we want to retain isEqualNode, it should be skipped too. Are there others we're not thinking about because "nobody" ever uses them? There might be others but I can only confirm these 3 elements. I don't think the other form elements in this list would have issues https://www.w3schools.com/html/html_form_elements.asp. But then again, there's a reason why morphdom doesn't use isEqualNode. What are the gotchas? eg. what about the input type="numeric" that is forced to display the value abc, putting it in violation of the half-baked pattern validation built into some elements? is "equal but not valid" something that matters in this context? I'm just trying to brainstorm and still feeling haunted by the abstract thinking required to get to an outcome on patrick-steele-idem/morphdom#206 Numeric inputs suffer from the same issue, their values are not updated if changed in the backend. I've tested it without isEqualNode and it works fine. If the backend clears the value, the UI clears it. If the backend changes the number, the UI changes it. And if the backend changes the value to a string, the UI just clears the value out. Interestingly, in my testing with a draft PR that doesn't use isEqualNode if the tagName is an input/select/textarea... I am finding that it's not select that is the problem, but option. So far as I can tell, @adank92's first MCVE works fine if I filter for option but still breaks if I filter for select. Does that make sense to you, Adan? Works: if (fromEl.tagName != 'OPTION' && fromEl.isEqualNode(toEl)) return false Doesn't work: if (fromEl.tagName != 'SELECT' && fromEl.isEqualNode(toEl)) return false I was/am worried that this bug will impact DATALIST elements as well, but if indeed it's the OPTION that is relevant, this could kill two birds with one stone. Yes, you are right. select itself is not a problem, it's option. Since morphdom can't find an option with a selected attribute, it ends up resetting the select. Thanks for taking care of it in a new PR.
2025-04-01T06:38:58.467999
2021-12-01T14:35:51
1068483091
{ "authors": [ "lajosrodek", "satyabollineni" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6805", "repo": "hortonworks/cloudbreak", "url": "https://github.com/hortonworks/cloudbreak/pull/11842" }
gharchive/pull-request
CB-15245:Fix NPE in create-aws-cluster when encryption type set to null Master PR: https://github.com/hortonworks/cloudbreak/pull/11840/files 2.49 PR: https://github.com/hortonworks/cloudbreak/pull/11841 I think this PR should be closed, and so should #11841. The right solution is to Repurpose #11840 for CB-2.49.0 (rebase the branch, force push, then change the target branch with the Edit button). Once #11840 gets merged, open a merge PR for CB-2.49.0 -> CB-2.50.0. Once the latter PR is merged, open another merge PR for CB-2.50.0 -> master (2.51). Please follow the merge process mentioned (here)[https://github.com/hortonworks/cloudbreak/pull/11809#pullrequestreview-814882747] and here. I think this PR should be closed, and so should #11841. The right solution is to Repurpose CB-15245:Fix NPE in create-aws-cluster when encryption type set to null #11840 for CB-2.49.0 (rebase the branch, force push, then change the target branch with the Edit button). Once CB-15245:Fix NPE in create-aws-cluster when encryption type set to null #11840 gets merged, open a merge PR for CB-2.49.0 -> CB-2.50.0, using a feature branch based off CB-2.50.0. Once the latter PR is merged, open another merge PR for CB-2.50.0 -> master (2.51), using a feature branch based off master. Please follow the merge process mentioned here and here. Changed the target branch for master PR to CB-2.50.0, and we are skipping the merge to 2.49 as discussed in https://jira.cloudera.com/browse/CB-15245?page=com.atlassian.jira.plugin.system.issuetabpanels%3Aall-tabpanel. Once the 2.50 PR is merged , will port that change to master.
2025-04-01T06:38:58.469772
2022-03-22T15:37:16
1176970855
{ "authors": [ "horadla23", "topolyai5" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6806", "repo": "hortonworks/cloudbreak", "url": "https://github.com/hortonworks/cloudbreak/pull/12476" }
gharchive/pull-request
CB-16388 RDSConfig calculation take long time, refactor the RdsConfig select by clusterId The rds config fetch is too expensive because the hibernate fetch all of the attached clusters as references. This is a backport from 2.55.0 @topolyai5 can you check why CI jobs seem stucked? @lajosrodek @topolyai5 what's the status of this PR? is this still targeted to 2.54? if yes, please either resolve comments or modify the change accordingly
2025-04-01T06:38:58.471837
2020-05-19T20:14:36
621255694
{ "authors": [ "bergerdenes", "cegganesh84" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6807", "repo": "hortonworks/cloudbreak", "url": "https://github.com/hortonworks/cloudbreak/pull/8100" }
gharchive/pull-request
DISTX-421 DISTX-427 Recommendation for host groups' scalability Added recommendations for auto-scalability and manual scalability. Implemented both CM service based recommendations and fallback recommendations. ./gradlew build ping @keyki Thanks @bergerdenes for taking a look. I have addressed those comments. @bergerdenes could you please merge this? Merged
2025-04-01T06:38:58.502174
2024-03-13T10:40:03
2183665664
{ "authors": [ "khushalidube-hwsys", "ravilodhi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6808", "repo": "hotwax/fulfillment-pwa", "url": "https://github.com/hotwax/fulfillment-pwa/issues/454" }
gharchive/issue
Pick and Save Not Enabled & Order Items Invisible Current behavior Currently, there are two critical issues affecting our system: Pick and Save Buttons Not Enabled: The Pick and Save buttons are not enabled in the app. Order Items Invisible in Fulfilment App: Additionally, the order items are not visible in the fulfilment application. As a result, we are unable to fulfill orders promptly. We tested this issue for order ID : SVC13094 in adoc-sv-oms Expected behavior The Pick and Save buttons should be enabled, allowing our team to efficiently manage orders. Order items should be visible within the fulfilment application to facilitate accurate order processing. This issues is due the dependency on OMS. The recent changes related to internationalInvoiceUrl (Custom document generation) were part of OMS 5.5.1 release. So in order to get it fix on adoc-sv-oms, need to be upgraded to 5.5.1.
2025-04-01T06:38:58.509239
2020-12-28T01:56:22
775150352
{ "authors": [ "dhh", "seanpdoyle" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6809", "repo": "hotwired/stimulus-rails", "url": "https://github.com/hotwired/stimulus-rails/pull/11" }
gharchive/pull-request
Make Autoloading more robust Duplicate of #3 Modifies the stimulus/loaders/autoloader module to account for module importing in 4 different scenarios: Immediately when parsed Whenever elements that declare [data-controller] are added, deleted, or modified (via a MutationObserver) Whenever the DOMContentLoaded event fires (for applications that don't depend on turbo Whenever the turbo:load event fires (for applications that do depend on turbo In addition to autoloading in more scenarios, this commit also changes the autoloader to read controller tokens in a spacing-agnostic manner, and to transform dasherized controller names (like message-rendering) to underscored file names (like message_rendering_controller.js). The changes made to test/dummy were done so manually, and are being checked into version control. In an ideal set up, the generator would be run for every System Test, and then reverted. Testing Add System Test level coverage to ensure that the configurations, helpers, and autoloading work together in concert to dynamically load Stimulus Controllers. To do so, add a Gemfile-level dependency on capybara, webdrivers, and selenium-webdrivers. Co-authored-by: Justin Malčić<EMAIL_ADDRESS> Thanks for the review @MarkoZabcic! I've added you as a co-author to the commit. Thanks for the review @MarkoZabcic! I've added you as a co-author to the commit. Plan is still to extract the bulk of this into stimulus, but at least we're not broken in the meanwhile! Plan is still to extract the bulk of this into stimulus, but at least we're not broken in the meanwhile!
2025-04-01T06:38:58.520079
2024-05-23T16:15:10
2313333700
{ "authors": [ "TJNOV", "howeyc" ], "license": "ISC", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6810", "repo": "howeyc/crocgui", "url": "https://github.com/howeyc/crocgui/issues/46" }
gharchive/issue
Upgrade the croc version croc is not compatible with older versions after upgrading to version 10.0 v9.6.16 was a breaking change so I have re-released it as v10.0.0 to signal that. clients must be >=v9.6.16 or >=v10 to be compatible. I updated it. Usually takes a week to show up in fdroid. Can a standalone apk file be provided? I added an APK to the release.
2025-04-01T06:38:58.547349
2023-07-21T02:27:46
1815058057
{ "authors": [ "Fridge003" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6811", "repo": "hpcaitech/ColossalAI", "url": "https://github.com/hpcaitech/ColossalAI/pull/4300" }
gharchive/pull-request
[pipeline] add pipeline support for T5Stack/T5EncoderModel 📌 Checklist before creating the PR [x] I have created an issue for this PR for traceability [x] The title follows the standard format: [doc/gemini/tensor/...]: A concise description [x] I have added relevant tags if possible for us to better distinguish different PRs 🚨 Issue number Link this PR to your issue with words like fixed to automatically close the linked issue upon merge e.g. fixed #1234, closed #1234, resolved #1234 #4247 📝 What does this PR do? Summarize your work here. if you have any plots/diagrams/screenshots/tables, please attach them here. Design special algorithm for distributing layers to pipeline stages, since the structure of T5(Encoder/Decoder) is more complex than other transformer models. Modify the logic of forward for T5Stack/T5EncoderModel, so that pipeline parallel can be smoothly run. Add relevant tests for layer distribution & pipeline forward. 💥 Checklist before requesting a review [x] I have linked my PR to an issue (instruction) [x] My issue clearly describes the problem/feature/proposal, with diagrams/charts/table/code if possible [x] I have performed a self-review of my code [x] I have added thorough tests. [x] I have added docstrings for all the functions/methods I implemented ⭐️ Do you enjoy contributing to Colossal-AI? [x] 🌝 Yes, I do. [ ] 🌚 No, I don't. Tell us more if you don't enjoy contributing to Colossal-AI.
2025-04-01T06:38:58.776898
2020-04-16T12:03:29
600991294
{ "authors": [ "shiva", "shivamerla", "vidhutsingh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6812", "repo": "hpe-storage/csi-driver", "url": "https://github.com/hpe-storage/csi-driver/pull/145" }
gharchive/pull-request
Implemented generic os event driven mechanism to update HPENodeInfo object if there is any change in object attributes detected . Issue: CON-882 (Continuously monitor worker node network configuration) Implementations: Implemented a generic polling mechanism that accepts anonymous function and polling interval as arguments. It has two functions 1.obj:=InitTask(...) This is basically used to initialize the task object which registers the anonymous function that needs to be polled along with poll frequency. Also, it invokes a separate thread to monitor the OS signals e.g SIGTERM, SIGHUP for the graceful exit of the polling task. 2.obj.start() This is basically to kick starts the polling in a loop unit os signal is received. It periodically polls the registered anonymous fn. This should be invoked as a separate goroutine. Created an anonymous wrapper function around getNodeInfo() that is passed to InitTask() with default 30 Minutes polling frequency. Testing: I have tested the code by manually deleting the cluster worker node's one of the data network interfaces and have validated that the network changes are dynamically picked by Node Server and also HPENodeInfo is updated. Hi, I'm not sure how to get these email to stop. Can you please remove me from your mailing list? -Shiva On Mon, Apr 27, 2020, 8:36 AM Shiva Krishna Merla<EMAIL_ADDRESS>wrote: @shivamerla approved this pull request. looks good.. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/hpe-storage/csi-driver/pull/145#pullrequestreview-401074021, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAB2ZBZRZRNVFF3UUBKPH3ROWRAFANCNFSM4MJSNJDQ . Hi, I'm not sure how to get these email to stop. Can you please remove me from your mailing list? … -Shiva I guess you need to Unsubscribe notification. That should stop. Since I haven't made any user-specific configuration it was the default. Hi, I'm not sure how to get these email to stop. Can you please remove me from your mailing list? … -Shiva I guess you need to Unsubscribe notification. That should stop. Since I haven't made any user-specific configuration it was the default. @shiva Looks like folks have been tagging you instead of me by mistake. My GitHub ID is @shivamerla. I am not sure how i can control it. Can you unsubscribe from notifications on this repo? Will try. Thanks for your responses. -Shiva On Mon, Apr 27, 2020, 9:49 PM Shiva Krishna Merla<EMAIL_ADDRESS>wrote: Hi, I'm not sure how to get these email to stop. Can you please remove me from your mailing list? … <#m_2134139001607331619_> -Shiva I guess you need to Unsubscribe notification. That should stop. Since I haven't made any user-specific configuration it was the default. @shiva https://github.com/shiva Looks like folks have been tagging you instead of me by mistake. My GitHub ID is @shivamerla https://github.com/shivamerla. I am not sure how i can control it. Can you unsubscribe from notifications on this repo? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/hpe-storage/csi-driver/pull/145#issuecomment-620378754, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAB2ZEG5Z26UNSZAROD6T3ROZN6LANCNFSM4MJSNJDQ .
2025-04-01T06:38:58.782787
2016-11-10T18:34:17
188584403
{ "authors": [ "tobiduer" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6813", "repo": "hpi-swt2/workshop-portal", "url": "https://github.com/hpi-swt2/workshop-portal/issues/35" }
gharchive/issue
US_4.3: Email to participants As organizer I want to send a bulk mail to all participants of an event in order to avoid coping the email addresses. Dependency: Event, Role Organizer, Role Pupil Additional details: The sender of the Email can be a single person (e.g<EMAIL_ADDRESS>or a mailing list (e.g. mintcamp2016@hpi.de). Please ensure, that this will not lead to access violation and discuss your solution with the HPI admins and my team.
2025-04-01T06:38:58.784345
2016-11-24T08:57:51
191464406
{ "authors": [ "bjrne", "chrisma" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6814", "repo": "hpi-swt2/workshop-portal", "url": "https://github.com/hpi-swt2/workshop-portal/issues/63" }
gharchive/issue
Missing information in seeds.rb As @radscheit already pointed out, db/seeds.rb is missing information to seed the database. Please add at least the required fields for a profile (:first_name, :last_name, :gender, :birth_date, :email, :school, :street_name, :zip_code, :city, :state, :country) /cc @hefowe @cmfcmf @Philipp-Bode Maybe there could be a test that makes sure that db/seeds.rb can be loaded?
2025-04-01T06:38:58.877323
2022-03-18T16:06:37
1173759910
{ "authors": [ "hrynko", "p3v9d5ui" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6815", "repo": "hrynko/vue-pdf-embed", "url": "https://github.com/hrynko/vue-pdf-embed/pull/17" }
gharchive/pull-request
Update readme with typedarray source Here's a small change to the README, with information on using a Uint8Array for the source, which I forgot to add in my previous PR. Thanks!
2025-04-01T06:38:58.889637
2024-10-30T01:30:56
2622642517
{ "authors": [ "vuvuong003" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6816", "repo": "hsalway1/WolfJobs", "url": "https://github.com/hsalway1/WolfJobs/issues/8" }
gharchive/issue
JobDetails unit test fails Failed test output: tests/components/Job/JobDetails.test.tsx > JobDetails > renders JobDetails TypeError: Cannot read properties of undefined (reading 'join') Added a conditional statement to check if the requiredSkills array existed before attempting to call join on it. Bug fixed!
2025-04-01T06:38:58.894061
2020-05-07T02:15:32
613724456
{ "authors": [ "williaster" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6817", "repo": "hshoff/vx", "url": "https://github.com/hshoff/vx/pull/678" }
gharchive/pull-request
new(vx-demo): convert Brush to codesandbox :memo: Documentation :house: Internal TODO [ ] fix initialBrushPosition in sandbox Part of #624, re-writes the Brush demo to link out to code-sandbox. Link (update branch to master upon merge). @kristw @hshoff I'm just gonna merge this and continue to debug the initialBrushPosition separately
2025-04-01T06:38:58.904287
2015-03-28T06:06:07
64905202
{ "authors": [ "brettwooldridge" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6818", "repo": "hstove/issue_stats", "url": "https://github.com/hstove/issue_stats/pull/25" }
gharchive/pull-request
Fix #23 decrease horizontal width of badges Fix #23 decrease horizontal width of badges Closing to squash.
2025-04-01T06:38:59.070663
2021-02-21T11:40:37
812835099
{ "authors": [ "scala-steward" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6820", "repo": "http4s/http4s-netty", "url": "https://github.com/http4s/http4s-netty/pull/116" }
gharchive/pull-request
Update sbt-release to 1.0.14 Updates com.github.gseitz:sbt-release from 1.0.13 to 1.0.14. I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "com.github.gseitz", artifactId = "sbt-release" } ] labels: sbt-plugin-update, semver-patch Superseded by #117.
2025-04-01T06:38:59.075043
2017-10-30T15:42:50
269648481
{ "authors": [ "ChristopherDavenport", "calvinbrown085", "rossabaker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6821", "repo": "http4s/http4s", "url": "https://github.com/http4s/http4s/issues/1504" }
gharchive/issue
Http4s Testing Documentation In our documentation we are lacking examples and demonstration of using http4s-testing to create unit tests for our services. We have some implementations in http4s.g8, but we should have more concrete examples that show a consensus that people can work from in the Website documentation. @ChristopherDavenport Could you point me to http4s-testing when you have a chance? I can't seem to find it. It's the testing subproject. It gets published as: "org.http4s" %% "http4s-testing" % Http4sVersion We have Http4sMatchers, which help with Specs2. If you search our tests for the methods therein, you'll find examples of their use. (Aside: it's not great that our testing module has a dependency on specs2. I think we should consider whether that's the right home.) Other interesting things in there are the ArbitraryInstances, which go nicely with Scalacheck. We have also started adding some laws that go well with discipline. I suspect these are more useful inside http4s than outside, though I can imagine them being used in more advanced testing scenarios. @rossabaker Awesome, I will play around with that! Thank you! I've been looking around at this a little bit, one thing i'm unclear on is where these docs are supposed to live. In the docs/ folder? Definitely seems good enough to close out this issue. If we want we can revisit arbitraries and law testing at some point in the future.
2025-04-01T06:38:59.078574
2016-05-03T03:21:39
152692230
{ "authors": [ "LeifW", "rossabaker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6822", "repo": "http4s/http4s", "url": "https://github.com/http4s/http4s/issues/622" }
gharchive/issue
Default bind address: <IP_ADDRESS> or <IP_ADDRESS>? The default has tripped a couple people up recently. I looked at a few other servers for prior art, and found support for both. Without looking, what do you expect? I'd say <IP_ADDRESS> On May 2, 2016 8:21 PM, "Ross A. Baker"<EMAIL_ADDRESS>wrote: The default has tripped a couple people up recently. I looked at a few other servers for prior art, and found support for both. Without looking, what do you expect? — You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub https://github.com/http4s/http4s/issues/622 Vetoed. See other ticket.
2025-04-01T06:38:59.089301
2017-07-01T09:32:53
239938946
{ "authors": [ "ixti", "janko-m", "mikegee" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6823", "repo": "httprb/http", "url": "https://github.com/httprb/http/pull/418" }
gharchive/pull-request
Further upload improvements This pull request adds two more upload improvements. First commit changes HTTP::Request::Body to use IO#readpartial if available, and second commit adds a buffer string for memory-efficient uploads; both are explained in detail in their commit descriptions. I was hoping that, once this is merged (or rejected), we could release a new version of HTTP.rb and form_data.rb with the streaming upload feature. Does this need an addition to the test suite to exercise both #read and #readpartial? @mikegee The example that uses StringIO as a "Enumerable IO" already exercises #readpartial, and the "non-Enumerable" FakeIO example exercises #read. Thanks for the detailed commit messages. I was about to ask why #readpartial was preferable. @mikegee I made what I think is an improvement in the last commit. Instead of writing out the #read/#readpartial logic and using a buffer string, we can use IO.copy_stream which will do all that for us (see the Rubinius implementation), with the help of a BlockIO object which acts as a "destination IO" and just calls the block with the data that IO.copy_stream "writes". Here we use a trick of creating a Proc objects from an implicit block by calling Proc.new without any arguments. So, this: def foo block = Proc.new end is equivalent to def foo(&block) end This way we can still use yield in other branches of the conditional. Wow! TIL about def foo; Proc.new; end @ixti Updated. I brought back returning Enumerator in HTTP::Request::Body#each because it makes tests simpler. Also, because of https://github.com/jruby/jruby/issues/4701 I changed the tests for IO objects to just verify that yielded chunks sum up to the entire content. @britishtea @ixti Updated. @ixti @britishtea Do you think this one is ready for merging? Sorry I was on a vacation and completely off the grid. Will handle this ASAP. @ixti Quick bump 😃 Sorry took me so long. Merged. Thank you! @ixti No problem at all! Sorry that I've been nagging about it, I just thought about writing a section in Shrine documentation about how to upload files with various HTTP libraries, and I wanted to wait until HTTP.rb had streaming IO uploads, so that I don't have to change the documentation later. Now that we merged these final memory optimizations, and made deflating work with IO/Enumerable uploads (https://github.com/httprb/http/pull/412), do you think we can release a new HTTP.rb version with the new streaming IO uploads feature? @janko-m I want to work-on/merge all outstanding PRs that can be merged in relatively fast, and then we'll be good to release 3.0.0.pre1 I just want to merge max breaking changes as we ae doing that anyway ;)) @ixti Sounds great, thank you!
2025-04-01T06:38:59.114070
2024-12-09T10:05:37
2726569099
{ "authors": [ "benzhangdragonplus", "huanngzh", "nicoptere", "yejr0229" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6824", "repo": "huanngzh/MV-Adapter", "url": "https://github.com/huanngzh/MV-Adapter/issues/8" }
gharchive/issue
How to achieve renderings from other perspectives? Thanksfor your efforts! The effect is very good, but the images are all from a horizontal perspective. How about the effect images from other perspectives? For example, Looking down on objects We look forward to your reply! hey, a part of the answer lies in the way the camera is initialized : https://github.com/huanngzh/MV-Adapter/blob/612211253f79d383656289ab85e8df22b9f46ea4/scripts/inference_t2mv_sdxl.py#L83-L92 specifically the values of elevation_deg=[0, 0, 0, 0, 0, 0], # <= rotation about the X axis : the "up/down" angle distance=[1.8] * num_views, #<= distance from center azimuth_deg=[x - 90 for x in [0, 45, 90, 180, 270, 315]], # <= rotation about the Y axis : the "turn table" angle these values are partly hardcoded (elevation & azimut_deg) and partly dynamic (distances is based on num_views) so first idea is to make everything dynamic like so: # initialise rotations values azimuts = [] distances = [] elevations = [] elevation = 0 step = int( 360 / num_views ) # 'turn table' angle for i in range( 0, 360, step): azimuts.append( i - 90 ) elevations.append( elevation ) distances.append( 1.8 ) print( "azimut values", len(azimuts), azimuts, 'step', step ) # with: a num_views of 9: => azimut values 9 [-90, -50, -10, 30, 70, 110, 150, 190, 230] step 40 # Prepare cameras cameras = get_orthogonal_camera( azimuth_deg=azimuts, elevation_deg=elevations, distance=distances, left=-0.55, right=0.55, bottom=-0.55, top=0.55, device=device, ) NB: the get_orthogonal_camera method doesn't have a num_views argument so you must provide the azimuth_deg (or add num_views to get_orthogonal_camera's args). this should create 9 views around the model BUT it doesn't work and I don't understand why. it does create 9 images but they always use the same 6 (default) angles... this is what the above call with 9 views produces the images 2 & 3, 4 & 5, 6 & 7 are computed individually (notice some slight differences) but have the same camera angles that correspond to the 'default' angles. any hint as to why & how to fix would be appreciated :) Hi, your understanding of the hyperparameters is correct. However, sorry we haven't released our arbitrary view generation, as shown in README. Hi, thanks for your great work, I'd like to know when will you release your inference code for generating arbitrary number of views?
2025-04-01T06:38:59.119092
2021-02-02T21:31:18
799707680
{ "authors": [ "Whisper40", "huashengdun" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6825", "repo": "huashengdun/webssh", "url": "https://github.com/huashengdun/webssh/issues/206" }
gharchive/issue
Cross Origin Problem I started your app on my server like that : wssh --address='<IP_ADDRESS>' --origin="*" --port=8888 --xsrf=False --debug And i access it from my computer to the server with my custom address. The problem is that it does not connect :( E.G : It's working without problem from localhost to localhost. But i need the access from * to my server Do you see what's wrong ? What error complained by your webssh server? Hello, this is the error that appears on the server It seems that the problem was on the nginx configuration ! Strange thing I will close it, thanks for help
2025-04-01T06:38:59.120175
2021-10-03T22:28:04
1014569126
{ "authors": [ "stefan-reich" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6826", "repo": "huashengdun/webssh", "url": "https://github.com/huashengdun/webssh/issues/244" }
gharchive/issue
Some keys appear wrong on German keyboard Hi, this one is weird. Generally the assignments are correct, but for example # appears as / in an WebSSH terminal. OK
2025-04-01T06:38:59.153962
2024-09-28T15:57:18
2554358539
{ "authors": [ "NateEag", "bpetit" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6827", "repo": "hubblo-org/scaphandre", "url": "https://github.com/hubblo-org/scaphandre/pull/393" }
gharchive/pull-request
Fix a few typos in why.md As a native US English speaker, these minor mistakes caught my eye while reading about this very cool project. So, I took a minute to clean them up. Thank you for this PR @NateEag ! Merged
2025-04-01T06:38:59.170760
2018-11-07T21:38:14
378479945
{ "authors": [ "huchenme", "tinchox5" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6828", "repo": "huchenme/github-trending-api", "url": "https://github.com/huchenme/github-trending-api/issues/11" }
gharchive/issue
Based on this awesome API Hi @huchenme! Maybe you want to know that I'm working on this project: https://github.com/zircleUI/github-trending-plus that heavily uses your API 😄 🎉 Thank you again for your work!! Hi @tinchox5, this is nice to know! Your site is fun to play with, nice animations 😄 @tinchox5 Just updated README and added your project to this section https://github.com/huchenme/github-trending-api#projects-using-github-trending-api 🙌 Thank you!! On Wed, Feb 13, 2019 at 10:40 AM Hu Chen<EMAIL_ADDRESS>wrote: @tinchox5 https://github.com/tinchox5 Just updated README and added your project to this section https://github.com/huchenme/github-trending-api#projects-using-github-trending-api 🙌 — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/huchenme/github-trending-api/issues/11#issuecomment-463247357, or mute the thread https://github.com/notifications/unsubscribe-auth/ALBcDTH1n5XIbPUQPk896crJ3TAwLrMgks5vNDIGgaJpZM4YTZyu .
2025-04-01T06:38:59.195596
2023-11-23T16:26:01
2008534893
{ "authors": [ "GaParmar", "JakobLS", "SunMarc", "billvsme", "faceslog", "khayamgondal", "klopez89", "sr5434", "xiaohaipeng" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6830", "repo": "huggingface/accelerate", "url": "https://github.com/huggingface/accelerate/issues/2182" }
gharchive/issue
ValueError: Query/Key/Value should either all have the same dtype, --enable_xformers_memory_efficient_attention fails with the following error when run with accelerate File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/xformers/ops/fmha/__init__.py", line 348, in _memory_efficient_attention_forward_requires_grad inp.validate_inputs() File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/xformers/ops/fmha/common.py", line 121, in validate_inputs raise ValueError( ValueError: Query/Key/Value should either all have the same dtype, or (in the quantized case) Key/Value should have dtype torch.int32 query.dtype: torch.float32 key.dtype : torch.float16 value.dtype: torch.float16 Steps: 0%| | 0/1000 [00:02<?, ?it/s] [2023-11-25 14:03:17,898] torch.distributed.elastic.multiprocessing.api: [ERROR] failed (exitcode: 1) local_rank: 0 (pid: 5264) of binary: /anaconda/envs/diffusers-ikin/bin/python Traceback (most recent call last): File "/anaconda/envs/diffusers-ikin/bin/accelerate", line 8, in <module> sys.exit(main()) File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/accelerate/commands/accelerate_cli.py", line 47, in main args.func(args) File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/accelerate/commands/launch.py", line 985, in launch_command multi_gpu_launcher(args) File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/accelerate/commands/launch.py", line 654, in multi_gpu_launcher distrib_run.run(args) File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/torch/distributed/run.py", line 797, in run elastic_launch( File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/torch/distributed/launcher/api.py", line 134, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) File "/anaconda/envs/diffusers-ikin/lib/python3.8/site-packages/torch/distributed/launcher/api.py", line 264, in launch_agent raise ChildFailedError( torch.distributed.elastic.multiprocessing.errors.ChildFailedError: ============================================================ diffusers/examples/dreambooth/train_dreambooth_lora_sdxl.py FAILED I am running accelerate as following accelerate launch diffusers/examples/dreambooth/train_dreambooth_lora_sdxl.py \ --pretrained_model_name_or_path="stabilityai/stable-diffusion-xl-base-1.0" \ --instance_data_dir={input_dir} \ --output_dir={output_dir} \ --instance_prompt=instance_prompt \ --mixed_precision="fp16" \ --resolution=1024 \ --train_batch_size=1 \ --gradient_accumulation_steps=4 \ --learning_rate=1e-4 \ --lr_scheduler="constant" \ --lr_warmup_steps=0 \ --checkpointing_steps=500 \ --max_train_steps=1000 \ --seed="0" \ --checkpoints_total_limit=5 \ --enable_xformers_memory_efficient_attention Accelerate config { "compute_environment": "LOCAL_MACHINE", "debug": false, "distributed_type": "MULTI_GPU", "downcast_bf16": false, "machine_rank": 0, "main_training_function": "main", "mixed_precision": "no", "num_machines": 1, "num_processes": 2, "rdzv_backend": "static", "same_network": false, "tpu_use_cluster": false, "tpu_use_sudo": false, "use_cpu": false } Versions: xformers==0.0.23.dev687 accelerate==0.24.1 torch==2.1.0 torchvision==0.16.1 Thanks for reporting @khayamgondal, is it working without accelerate ? It looks like the query don't have the same dtype as the key and the value. I would suggest trying to find out why it is the case. If I run accelerate without --enable_xformers_memory_efficient_attention flag, training works fine. Llooks like somehow xformers upscales query vector to float32. +1 @sr5434 can you try without accelerate ? it doesn't seems to be an issue with accelerate since we don't modify the dtype. Maybe it is probably something in train_dreambooth_lora_sdxl.py that modifies the dtype. LMK how it goes ! If I run accelerate without --enable_xformers_memory_efficient_attention flag, training works fine. Looks like somehow xformers upscales query vector to float32. this made all the difference. thank you! @klopez89 that seems more like a workaround though @SunMarc This is the source code: https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_lora_sdxl.py Also when I just use python the error stays the same Thanks for clarifying @sr5434 ! This issue is not related to accelerate then. You should probably open an issue on diffusers repository. My guess is that xformers attention is not working as expected in the script. i think the problem is caused by xformers and pytorch used different cuda version to compile, xformers is compiled used cuda 11.8 or higher,but the cuda version your pytorch compiled is 11.7 or lower.solution is to download xfomers source code and compile it in your environment @SunMarc I was able to fix the problem by disabling xformers +1 I get the same error when running the train_text_to_image_lora_sdxl.py script. I've made sure to align the CUDA version as suggestion above but no difference. accelerate==0.26.1 torch==2.1.2+cu118 torchvision==0.16.2+cu118 xformers==0.0.23.post1+cu118 Can the same Dreambooth fixes be applied here as well? Hi @JakobLS , can you try to build xformers from source ? I'm running this in a cloud VM and it's freezing when I do that. Tried it twice - in my original and in a new I created. If I don't align the CUDA versions before, as I showed above, I get the following error: The detected CUDA version (11.8) mismatches the version that was used to compile PyTorch (12.1). Please make sure to use the same CUDA versions. Hi @JakobLS , you need to either reinstall pytorch with the cu121. Or the you can the version of CUDA of your driver to 11.8 Hi @SunMarc, I might not be explaining myself properly. That's what I wanted to show with my first comment. The VM comes installed with CUDA 11.8. If I install a Pytorch version with CUDA 11.8 and run it with no mixed precision it works fine. If I run it with mixed_precision="fp16" I get the following error: ValueError: Query/Key/Value should either all have the same dtype, or (in the quantized case) Key/Value should have dtype torch.int32 query.dtype: torch.float32 key.dtype : torch.float16 value.dtype: torch.float16 Similarly if I install a Pytorch version with CUDA 12.1 and run it with no mixed precision it also works fine. But if I run it with mixed_precision="fp16" I get the exact same error. It doesn't like when I run it with mixed precision so it looks like the issue might be there rather than in the CUDA version. I'd like to use mixed precision though due to the speed advantage. Let me know if you want me to open a new issue for this btw. @SunMarc, when I install from source, the installation automatically terminates and the terminal window shuts down. No xformers is being installed. Using the following script: pip install ninja pip install -v -U git+https://github.com/facebookresearch/xformers.git@main#egg=xformers This is the installed Pytorch: accelerate==0.26.1 torch==2.1.2+cu118 torchvision==0.16.2+cu118 and CUDA (checking with nvcc --version): nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2022 NVIDIA Corporation Built on Wed_Sep_21_10:33:58_PDT_2022 Cuda compilation tools, release 11.8, V11.8.89 Build cuda_11.8.r11.8/compiler.31833905_0 This is indeed strange that you are not able to install from source. Let's try this then: pip3 install -U xformers --index-url https://download.pytorch.org/whl/cu118. You should get the 11.8 version this way. That successfully installs xformers: accelerate==0.26.1 torch==2.1.2+cu118 torchvision==0.16.2+cu118 xformers==0.0.23.post1+cu118 But when I now execute the script train_text_to_image_lora_sdxl.py as shown above, I get the same error as OP: ValueError: Query/Key/Value should either all have the same dtype, or (in the quantized case) Key/Value should have dtype torch.int32 query.dtype: torch.float32 key.dtype : torch.float16 value.dtype: torch.float16 If I run the script without --enable_xformers_memory_efficient_attention I instead get the following error: ValueError: Attempting to unscale FP16 gradients. ValueError: Attempting to unscale FP16 gradients. error means that the model is on fp16. See this thread for more information. As for the other issue, it seems that compiling both pytorch and xformers with the same cuda version doesn't work. I will try to debug that. Thanks for looking into it. I'll keep --enable_xformers_memory_efficient_attention enabled to reap its benefits though. 🙂 If it helps you, it's working if I remove mixed precision like this: accelerate launch $SCRIPT_PATH \ --enable_xformers_memory_efficient_attention \ # Same as before Still, I'd like to use mixed_precision="fp16" for the speed boost. Still the same issue thanks for the workaround @JakobLS Has anyone found a fix to this? I still facing the same issues. It's now some time ago and I'm not working on it anymore. But I think I "solved" it by using mixed_precision="fp16" over xformers instead since the latter seem to lead to clearer degradation in image quality which fp16 doesn't. You could also try to update to later cuda, torch and xformers versions if you really want to use xformers (I think I tried that and got it to work).
2025-04-01T06:38:59.209540
2023-09-04T10:57:21
1880028195
{ "authors": [ "GeauxEric", "LaurentMazare", "akhildevelops" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6831", "repo": "huggingface/candle", "url": "https://github.com/huggingface/candle/issues/734" }
gharchive/issue
[On Cuda] Tensor::randn fails to generate random floats for odd no: of samples I've run the sample code, with different shapes, mentioned in Readme file (https://github.com/huggingface/candle/blob/main/README.md#get-started) use candle_core::{Device, Tensor}; fn main() -> Result<(), Box<dyn std::error::Error>> { let device = Device::Cpu; let a = Tensor::randn(0f32, 1., (4, 5), &device)?; // This will generate random values let b = Tensor::randn(0f32, 1., (5, 3), &device)?; // This will error out let c = a.matmul(&b)?; println!("{c}"); Ok(()) } It will error out during runtime with the message: Cuda(Curand(CurandError(CURAND_STATUS_LENGTH_NOT_MULTIPLE))). It works if I change to CPU. After digging deeper into cudarc library, found that a pseudorandom generator is used for generating floats and cudarand supports only even number of output samples for pseudorandom generator. How can I generate odd number of samples ? One workaround is to index the tensor b and reshape: #[cfg(test)] mod tests { use candle_core::{Device, IndexOp, Tensor}; #[test] fn it_works() -> Result<(), Box<dyn std::error::Error>> { let device = Device::new_cuda(0)?; let a = Tensor::randn(0f32, 1., (4, 5), &device)?; // This will generate random values let b = Tensor::randn(0f32, 1., (5, 4), &device)?; // This is OK let b_ = b.i((.., 0..3))?.reshape((5,3))?; // Reshape is needed println!("b_ shape: {:?}", b_.shape()); let c = a.matmul(&b_)?; println!("{a}"); println!("{b}"); println!("{c}"); Ok(()) } } Result: b_ shape: [5, 3] [[-1.8936, 0.9397, 1.0709, -0.0135, -0.2310], [-0.4147, -0.1733, 0.3322, 0.1012, 0.4609], [-0.5161, 0.1109, -0.1549, 2.1302, 0.7367], [-0.3432, 1.0344, 1.1740, 0.8580, 0.8757]] Tensor[[4, 5], f32, cuda:0] [[-0.4003, 0.7206, -0.9397, -0.0915], [ 2.3167, 1.5755, -0.7861, 0.1864], [ 0.2528, -0.8081, -0.5926, -1.3103], [-1.9073, 0.8630, 0.1234, -1.7610], [ 0.4231, -0.7811, -0.2071, -0.2975]] Tensor[[5, 4], f32, cuda:0] [[ 3.1338, -0.5808, 0.4523], [-0.1496, -1.1131, 0.2461], [-3.3269, 1.1909, 0.5999], [ 1.5647, 0.4901, -1.2617]] Tensor[[4, 3], f32, cuda:0] Since the underlying cuRAND only supports even numbers, maybe candle can hide the workaround in the randn API. Right, I've actually just merged #793 that should hopefully fix this (at the expense of generating an additional value when the number of elements is odd). Closing as hopefully fixed now, but please re-open if you run into any more issues.
2025-04-01T06:38:59.212265
2022-11-06T16:35:36
1437463660
{ "authors": [ "HuggingFaceDocBuilderDev", "camartinezbu" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6832", "repo": "huggingface/course", "url": "https://github.com/huggingface/course/pull/366" }
gharchive/pull-request
Spanish translation of Chapter 5 ¡Hi! This is the spanish translation for chapter 5 according to issue #38. The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.
2025-04-01T06:38:59.237016
2023-09-14T11:36:54
1896382218
{ "authors": [ "albertvillanova", "codecov-commenter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6833", "repo": "huggingface/datasets-server", "url": "https://github.com/huggingface/datasets-server/pull/1805" }
gharchive/pull-request
Replace deprecated typing Since Python 3.9 (this project is in Python 3.9.15), some typing hints are deprecated. This PR replaces deprecated type hints. Codecov Report Patch coverage: 100.00% and project coverage change: -4.44% :warning: Comparison is base (f24b758) 90.50% compared to head (e40e7ba) 86.06%. Report is 1 commits behind head on main. Additional details and impacted files @@ Coverage Diff @@ ## main #1805 +/- ## ========================================== - Coverage 90.50% 86.06% -4.44% ========================================== Files 222 87 -135 Lines 13641 3187 -10454 ========================================== - Hits 12346 2743 -9603 + Misses 1295 444 -851 Flag Coverage Δ jobs_cache_maintenance ? jobs_mongodb_migration 85.92% <100.00%> (ø) libs_libcommon ? services_admin 87.44% <100.00%> (ø) services_api ? services_rows 82.75% <100.00%> (ø) services_search 79.16% <ø> (ø) services_sse-api 94.32% <100.00%> (ø) services_worker ? Flags with carried forward coverage won't be shown. Click here to find out more. Files Changed Coverage Δ services/search/src/search/routes/search.py 54.62% <ø> (ø) ...s/mongodb_migration/src/mongodb_migration/check.py 34.48% <100.00%> (ø) ...ngodb_migration/src/mongodb_migration/collector.py 100.00% <100.00%> (ø) ...ation/src/mongodb_migration/deletion_migrations.py 69.51% <100.00%> (ø) ..._20230511100700_queue_delete_indexes_with_force.py 76.00% <100.00%> (ø) ...30516101600_queue_delete_index_without_revision.py 73.07% <100.00%> (ø) ...bs/mongodb_migration/src/mongodb_migration/plan.py 100.00% <100.00%> (ø) ...he_add_partial_field_in_config_parquet_and_info.py 100.00% <100.00%> (ø) ..._cache_add_features_field_in_split_duckdb_index.py 100.00% <100.00%> (ø) jobs/mongodb_migration/tests/test_plan.py 95.45% <100.00%> (ø) ... and 3 more ... and 135 files with indirect coverage changes :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
2025-04-01T06:38:59.238910
2024-02-02T00:34:40
2113800655
{ "authors": [ "HuggingFaceDocBuilderDev", "PierreCounathe" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6834", "repo": "huggingface/deep-rl-class", "url": "https://github.com/huggingface/deep-rl-class/pull/481" }
gharchive/pull-request
Unit 3 proposal updates In Unit 3, in the description of Double Deep Q Networks, I thought that the descriptions of the respective roles of the DQN and of the Target Network were a bit confusing, which is why I am proposing an update that - I think - matches the pseudo-code provided in the Unit. Maybe I am mistaken and just did not understand these paragraphs. Let me know. The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2025-04-01T06:38:59.242787
2023-11-03T00:01:39
1975222291
{ "authors": [ "HuggingFaceDocBuilderDev", "mishig25" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6835", "repo": "huggingface/hub-docs", "url": "https://github.com/huggingface/hub-docs/pull/1079" }
gharchive/pull-request
[Docs Model UI] Gallery component Follow up to https://github.com/huggingface/hub-docs/pull/1076#issuecomment-1791164555 The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.
2025-04-01T06:38:59.246197
2023-02-09T16:19:46
1578200001
{ "authors": [ "HuggingFaceDocBuilderDev", "SBrandeis", "julien-c", "osanseviero" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6836", "repo": "huggingface/hub-docs", "url": "https://github.com/huggingface/hub-docs/pull/669" }
gharchive/pull-request
Add Space disk space info Related discussion https://huggingface.slack.com/archives/C048K60MPNF/p1675778705684199 The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. Would it make sens to display that information in the pricing page instead, and maybe link to there from the docs? It would avoid discrepancies between the docs and the actual prices / hardware specs, and keep moon-landing as the source of truth on that matter WDYT @julien-c? You mentioned nicer in the docs that in /settings, but what about /pricing? no strong opinion – we can also merge them here for now, and think about it later i.E. i think on /pricing for instance "disk space" might be a bit too much info, so i think it makes sense here Alright, merging ahead then!
2025-04-01T06:38:59.248060
2023-03-01T10:35:10
1604704271
{ "authors": [ "HuggingFaceDocBuilderDev", "dveni", "julien-c" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6837", "repo": "huggingface/hub-docs", "url": "https://github.com/huggingface/hub-docs/pull/701" }
gharchive/pull-request
Update models-downloading.md Update the file downloading function, I got this warning using cached_download: file_download.py:629: FutureWarning: `cached_download` is the legacy way to download files from the HF hub, please consider upgrading to `hf_hub_download` The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. thanks @dveni!
2025-04-01T06:38:59.250881
2023-11-20T14:11:54
2002295062
{ "authors": [ "Wauplin", "remyleone" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6838", "repo": "huggingface/huggingface_hub", "url": "https://github.com/huggingface/huggingface_hub/issues/1846" }
gharchive/issue
Issue detecting graphviz on macos Describe the bug When I run huggingface-cli env while having graphviz installed, I expect to see the version of graphviz showing up. Reproduction brew install graphviz huggingface-cli env Logs $ huggingface-cli env Copy-and-paste the text below in your GitHub issue. - huggingface_hub version: 0.15.1 - Platform: macOS-14.1.1-arm64-arm-64bit - Python version: 3.11.5 - Running in iPython ?: No - Running in notebook ?: No - Running in Google Colab ?: No - Token path ?: /Users/rleone/.cache/huggingface/token - Has saved token ?: True - Who am I ?: remyleone - Configured git credential helpers: 360000, osxkeychain - FastAI: 2.7.13 - Tensorflow: 2.15.0 - Torch: 2.1.0 - Jinja2: 3.1.2 - Graphviz: N/A - Pydot: 1.4.2 - Pillow: 9.4.0 - hf_transfer: 0.1.4 - gradio: 4.4.1 - numpy: 1.24.3 - ENDPOINT: https://huggingface.co - HUGGINGFACE_HUB_CACHE: /Users/rleone/.cache/huggingface/hub - HUGGINGFACE_ASSETS_CACHE: /Users/rleone/.cache/huggingface/assets - HF_TOKEN_PATH: /Users/rleone/.cache/huggingface/token - HF_HUB_OFFLINE: False - HF_HUB_DISABLE_TELEMETRY: False - HF_HUB_DISABLE_PROGRESS_BARS: None - HF_HUB_DISABLE_SYMLINKS_WARNING: False - HF_HUB_DISABLE_EXPERIMENTAL_WARNING: False - HF_HUB_DISABLE_IMPLICIT_TOKEN: False - HF_HUB_ENABLE_HF_TRANSFER: False (base) (k8s|admin@cli-k8s-beautiful-proskuriakova-7b18357b-9c0a-46ea-af9a-5427d97f17c2:kubeflow) # rleone @ rleone-macbook in ~ [15:08:58] $ which dot /opt/homebrew/bin/dot (base) (k8s|admin@cli-k8s-beautiful-proskuriakova-7b18357b-9c0a-46ea-af9a-5427d97f17c2:kubeflow) # rleone @ rleone-macbook in ~ [15:09:05] $ dot --version dot - graphviz version 9.0.0 (20230911.1827) ### System info ```shell $ huggingface-cli env Copy-and-paste the text below in your GitHub issue. - huggingface_hub version: 0.15.1 - Platform: macOS-14.1.1-arm64-arm-64bit - Python version: 3.11.5 - Running in iPython ?: No - Running in notebook ?: No - Running in Google Colab ?: No - Token path ?: /Users/rleone/.cache/huggingface/token - Has saved token ?: True - Who am I ?: remyleone - Configured git credential helpers: 360000, osxkeychain - FastAI: 2.7.13 - Tensorflow: 2.15.0 - Torch: 2.1.0 - Jinja2: 3.1.2 - Graphviz: N/A - Pydot: 1.4.2 - Pillow: 9.4.0 - hf_transfer: 0.1.4 - gradio: 4.4.1 - numpy: 1.24.3 - ENDPOINT: https://huggingface.co - HUGGINGFACE_HUB_CACHE: /Users/rleone/.cache/huggingface/hub - HUGGINGFACE_ASSETS_CACHE: /Users/rleone/.cache/huggingface/assets - HF_TOKEN_PATH: /Users/rleone/.cache/huggingface/token - HF_HUB_OFFLINE: False - HF_HUB_DISABLE_TELEMETRY: False - HF_HUB_DISABLE_PROGRESS_BARS: None - HF_HUB_DISABLE_SYMLINKS_WARNING: False - HF_HUB_DISABLE_EXPERIMENTAL_WARNING: False - HF_HUB_DISABLE_IMPLICIT_TOKEN: False - HF_HUB_ENABLE_HF_TRANSFER: False Thanks it solved it. I was thinking that dependencies were somewhat installed from a different lib but it wasn't. Glad to know your problem's fixed!
2025-04-01T06:38:59.254017
2023-11-22T15:38:50
2006606079
{ "authors": [ "HuggingFaceDocBuilderDev", "Wauplin" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6839", "repo": "huggingface/huggingface_hub", "url": "https://github.com/huggingface/huggingface_hub/pull/1854" }
gharchive/pull-request
Send user_agent in HEAD calls It looks like for the past 14 months we were not sending the user agent in the HEAD call when downloading a file thanks to a change I made in https://github.com/huggingface/huggingface_hub/pull/1058... Thanks @patrickvonplaten for reporting (private slack thread). This means we don't have user_agent (i.e. library_name/version + additional information like pipeline_class) in our stats when model files are already cached. This PR fixes it and adds a regression test for it. cc @julien-c @osanseviero weird that we haven't found out about it before. (also, deprecate http_user_agent -a completely unused method- + add get_hf_file_metadata to HfApi -it wasn't there before-) The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.
2025-04-01T06:38:59.255909
2022-02-02T10:26:57
1121748467
{ "authors": [ "jimypbr", "michaelbenayoun" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6840", "repo": "huggingface/optimum-graphcore", "url": "https://github.com/huggingface/optimum-graphcore/pull/30" }
gharchive/pull-request
Bart and T5 What does this PR do? Adds training support for: BartForConditionalGeneration T5ForConditionalGeneration Evaluation for these models will come in another PR but some building blocks are already implemented here. There are also examples for summarization and translation finetuning, but instructions on how to use them will also come in another PR. 🎉 looks good to me
2025-04-01T06:38:59.257778
2023-10-04T15:04:02
1926434876
{ "authors": [ "bocchris-aws", "michaelbenayoun" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6841", "repo": "huggingface/optimum-neuron", "url": "https://github.com/huggingface/optimum-neuron/issues/249" }
gharchive/issue
Tensor parallelism: saved model can't be loaded After running finetuning using run_summarization.py for a couple of steps, the directory to which the model shards have been saved does not contain a config.json. Hence, the model stored can't be loaded using from_pretrained():` OSError: PATH_TO_SAVED_MODEL_DIR does not appear to have a file named config.json. Should the config.json be saved together with the shards? You need to provide the parent directory instead. The parent directory contains all the config files.
2025-04-01T06:38:59.260421
2022-04-25T14:38:38
1214611466
{ "authors": [ "HuggingFaceDocBuilderDev", "JingyaHuang", "fxmarty" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6842", "repo": "huggingface/optimum", "url": "https://github.com/huggingface/optimum/pull/154" }
gharchive/pull-request
Remove unused argument What does this PR do? Remove an unused argument in compute_loss_ort() from Trainer class. @JingyaHuang (I can not add reviewers so I ping you!) Before submitting [x] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case). The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. Closing as corrected in https://github.com/huggingface/optimum/pull/189/ Sorry, I did not see this.
2025-04-01T06:38:59.262295
2024-10-25T15:49:57
2614485807
{ "authors": [ "raghavian", "rwightman" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6843", "repo": "huggingface/pytorch-image-models", "url": "https://github.com/huggingface/pytorch-image-models/issues/2313" }
gharchive/issue
[FEATURE] Additional metrics for ImageNet validation results The results files for ImageNet, for example this one reports top-1, top-5 accuracy along with other measures. Can the total training time of each model also be added? Inference times are available in other results files but without the exact number of training epochs. Without the training epochs for all the models, it is not straightforward to estimate the total training time per model. This will be useful to study the performance versus training time resource consumption, for example. @raghavian models have been trains on such a wide variety of compute resources across a long period of time, the effort to pull that information together would be too high relative to continuing library maintenance and ongoing feature development.
2025-04-01T06:38:59.267042
2023-02-08T13:41:31
1576146136
{ "authors": [ "Narsil", "mrsteyk" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6844", "repo": "huggingface/safetensors", "url": "https://github.com/huggingface/safetensors/issues/178" }
gharchive/issue
Explicit automatic alignment of header My use case might be overly specific, but when writing/using vectorised code on mmap'd safetensors file header sometimes causes everything to have an odd-numbered pointer which breaks even 16 bit vectorisation. Is there a possibility that python bindings will get an option to save with an explicit alignment? Just padding the header should be enough for most use cases. No it's not overly specific, actually there's already a PR for that. https://github.com/huggingface/safetensors/pull/148 I was waiting for more need for it before merging, but it seems this is picking up in low level frameworks where alignment could really help speed up load times. Also I love the project ! If you want pure rust ML framework I recommend https://github.com/coreylowman/dfdx (Still very early on, but there's at least a lot to inspire from IMO). For instance I implemented https://github.com/Narsil/fast_gpt2 (without dfdx, more like your approach, but still stealing the mkl bindings from dfdx to get the performance ! ) Thank you, @Narsil! Yeah, doing math low level isn't that popular apart from people who know how to code and are on "sub-par" HW by today's standards. Also thanks for the mention of dfdx and your repo. I didn't even consider trying to use any BLAS lib. isn't that popular apart from people who know how to code and are on "sub-par" HW by today's standards. It's still the future in my eyes. The ML fields is somewhat settling and not experimenting as much as it used to, performance is becoming a real concern for anything at scale. And all the python solution for performance are way to clunky to beat compiled code. This is a very personal view. Closing because #148 is merged
2025-04-01T06:38:59.281567
2023-06-01T05:21:58
1735481908
{ "authors": [ "Steven-N", "austinmw", "gsaivinay", "monuminu", "mrwadams", "philschmid", "razasaad", "songfeng", "yapweiyih" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6845", "repo": "huggingface/text-generation-inference", "url": "https://github.com/huggingface/text-generation-inference/issues/390" }
gharchive/issue
trust_remote_code is not available in SageMaker Endpoint System Info Hi Team , As i can see there is not way i can set trust_remote_code option in sagemaker_endpoint.sh #!/bin/bash if [[ -z "${HF_MODEL_ID}" ]]; then echo "HF_MODEL_ID must be set" exit 1 fi export MODEL_ID="${HF_MODEL_ID}" if [[ -n "${HF_MODEL_REVISION}" ]]; then export REVISION="${HF_MODEL_REVISION}" fi if [[ -n "${SM_NUM_GPUS}" ]]; then export NUM_SHARD="${SM_NUM_GPUS}" fi if [[ -n "${HF_MODEL_QUANTIZE}" ]]; then export QUANTIZE="${HF_MODEL_QUANTIZE}" fi text-generation-launcher --port 8080 Information [ ] Docker [ ] The CLI directly Tasks [ ] An officially supported command [ ] My own modifications Reproduction llm_model = HuggingFaceModel( role=role, image_uri=llm_image, env={ 'HF_MODEL_ID': hf_model_id, 'HF_MODEL_QUANTIZE': json.dumps(use_quantization), 'SM_NUM_GPUS': json.dumps(number_of_gpu) } ) This gives error for Falcon-40b Expected behavior The endpoint should get created @OlivierDehaene Hi, do I currently need to build this locally, or is this latest updated docker image uploaded to a public container repository that I can pull from? We already working on releasing the new Image for SageMaker. I ll keep you posted here. @austinmw @songfeng @OlivierDehaene New image is now available. See Phil's latest blog post for how to deploy Falcon 7B & 40B - https://www.philschmid.de/sagemaker-falcon-llm @Markus-Zeggel The tgi 0.8.2 works for me. Also, make sure you update the sagemaker sdk or maybe create a new env from scratch. trust_remote_code = True env={ 'HF_MODEL_ID': hf_model_id, # 'HF_MODEL_QUANTIZE': json.dumps(use_quantization), 'SM_NUM_GPUS': json.dumps(number_of_gpu), 'HF_MODEL_TRUST_REMOTE_CODE': json.dumps(trust_remote_code) } Hi, I exactly followed the https://www.philschmid.de/sagemaker-falcon-llm tutorial. (The only difference is the instance type) However, I still get the error message Loading tiiuae/falcon-7b-instruct requires you to execute the configuration file in that repo on your local machine. Make sure you have read the code there to avoid malicious use, then set the option trust_remote_code=True to remove this error. I tried this config: # TGI config config = { 'HF_MODEL_ID': "tiiuae/falcon-7b-instruct", # model_id from hf.co/models # 'SM_NUM_GPUS': json.dumps(number_of_gpu), # Number of GPU used per replica 'MAX_INPUT_LENGTH': json.dumps(1024), # Max length of input text 'MAX_TOTEL_TOKENS': json.dumps(2048), # Max length of the generation (including input text) 'HF_MODEL_TRUST_REMOTE_CODE': "true" # 'HF_MODEL_QUANTIZE': "bitsandbytes", # comment in to quantize } Am I missing something? Is this the correct image:<PHONE_NUMBER>84.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-tgi-inference:2.0.0-tgi0.8.2-gpu-py39-cu118-ubuntu20.04 ? Thanks! @Markus-Zeggel The tgi 0.8.2 works for me. Also, make sure you update the sagemaker sdk or maybe create a new env from scratch. trust_remote_code = True env={ 'HF_MODEL_ID': hf_model_id, # 'HF_MODEL_QUANTIZE': json.dumps(use_quantization), 'SM_NUM_GPUS': json.dumps(number_of_gpu), 'HF_MODEL_TRUST_REMOTE_CODE': json.dumps(trust_remote_code) } Hello, would like to know if anybody actually managed to use SSE streaming in Sagemaker using this TGI server image. Thanks. @gsaivinay sagemaker is currently not supporting SSE. Apologies as I know this isn't the best place for it, but I wasn't sure where else to ask this... Is anyone aware if we can provide the scripts that Sagemaker needs ourselves rather than pulling from huggingface.co/models which occurs when you define 'HF_MODEL_ID'? We're behind a proxy so cannot do it directly from sagemaker. We have other models that work fine in the same environment, however we saw that the falcon-7b has a few python scripts in the HF repo (https://huggingface.co/tiiuae/falcon-7b/tree/main). The models that work fine do not have these python scripts which from other research seems to be why it pulls from HF...? @Steven-N we created a blog post https://www.philschmid.de/sagemaker-llm-vpc @philschmid, i tried with tiiuae/falcon-rw-1b on sagemaker==2.170.0, still get the following error ValueError: Loading /opt/ml/model requires you to execute the configuration file in that repo on your local machine. Make sure you have read the code there to avoid malicious use, then set the option `trust_remote_code=True` to remove this error. Code to reproduce. import torch from peft import PeftModel from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer from huggingface_hub import snapshot_download import sagemaker import boto3 sess = sagemaker.Session() # sagemaker session bucket -> used for uploading data, models and logs # sagemaker will automatically create this bucket if it not exists sagemaker_session_bucket=None if sagemaker_session_bucket is None and sess is not None: # set to default bucket if a bucket name is not given sagemaker_session_bucket = sess.default_bucket() try: role = sagemaker.get_execution_role() except ValueError: iam = boto3.client('iam') role = iam.get_role(RoleName='sagemaker_execution_role')['Role']['Arn'] print(f"sagemaker role arn: {role}") print(f"sagemaker session region: {sess.boto_region_name}") MODEL_ID = "tiiuae/falcon-rw-1b" CACHED_DIR = "../cache" MERGE_MODEL_DIR = "merged_model_test" model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.float16, low_cpu_mem_usage=True, # device_map="auto", trust_remote_code=True, cache_dir=CACHED_DIR, ) tokenizer = AutoTokenizer.from_pretrained( MODEL_ID, cache_dir=CACHED_DIR, ) model.save_pretrained(MERGE_MODEL_DIR, safe_serialization=True) tokenizer.save_pretrained(MERGE_MODEL_DIR, safe_serialization=True) import os parent_dir = os.getcwd() # change to model dir os.chdir(MERGE_MODEL_DIR) # use pigz for faster and parallel compression !tar -cf model.tar.gz --use-compress-program=pigz * # change back to parent dir os.chdir(parent_dir) from sagemaker.s3 import S3Uploader # upload model.tar.gz to s3 s3_model_uri = S3Uploader.upload(local_path=str(MERGE_MODEL_DIR + "/model.tar.gz"), desired_s3_uri=f"s3://{sess.default_bucket()}/test-model") print(f"model uploaded to: {s3_model_uri}") from sagemaker.huggingface import get_huggingface_llm_image_uri, HuggingFaceModel import json image_uri = get_huggingface_llm_image_uri("huggingface", version="0.8.2") print(f"llm image uri: {image_uri}") instance_type = "ml.g4dn.2xlarge" health_check_timeout = 300 trust_remote_code = True config = { "HF_MODEL_ID": "/opt/ml/model", # path to where sagemaker stores the model "MAX_INPUT_LENGTH": json.dumps(2048), # Max length of input text "MAX_TOTAL_TOKENS": json.dumps(3000), # Max length of the generation (including input text) "HF_MODEL_TRUST_REMOTE_CODE": json.dumps(trust_remote_code) } # create HuggingFaceModel with the image uri llm_model = HuggingFaceModel( role=role, image_uri=image_uri, model_data=s3_model_uri, env=config, ) endpoint_name = sagemaker.utils.name_from_base("test") predictor = llm_model.deploy( endpoint_name=endpoint_name, initial_instance_count=1, instance_type=instance_type, model_data_download_timeout=10 * 60, container_startup_health_check_timeout=10 * 60, wait=False, ) print(predictor.endpoint_name) falcon-rw-1b is a different model than the 7B or 40B model. Hey team, I was able to reproduce the same error with using the new TGI 0.8.2 image with deployment of falcon-7b following this guide and setting the 'HF_MODEL_TRUST_REMOTE_CODE': json.dumps(True),. The only difference was the Instance Size (ml.m5.2xlarge). I can't correlate if instance size is what raises the exception of trust_remote_code=true.
2025-04-01T06:38:59.287797
2024-06-16T20:43:43
2355970516
{ "authors": [ "Narsil", "rYoussefAli" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6846", "repo": "huggingface/text-generation-inference", "url": "https://github.com/huggingface/text-generation-inference/pull/2071" }
gharchive/pull-request
Fix local installation after Rust 1.79 and transformers 4.41.2 I was trying to install text-generation-inference locally without Docker, and I encountered several problems while doing so. I collected all those problems and made the fixes here so that people do not encounter such tedious problems again. Fixes # (issue) The rust-toolchain.toml file was overriding Rust version to 1.78.0, making the installation fail as the inline const feature in rust requires 1.79.0. Upgraded the transformers version. Fixed the error: "str" has no attribute "logits" in rw.py when the model starts Before submitting [ X] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case). [ X] Did you read the contributor guideline, Pull Request section? Who can review? Anyone in the community is free to review the PR once the tests have passed. Feel free to tag members/contributors who may be interested in your PR. @OlivierDehaene @Narsil The rust-toolchain is already fixed on main (and you don't seem to fix it it seems). Rw.py should be in it's own PR (and with a reproducer, no need for a test since we don't really maintain that hard non flash models, no custom kernels models anymore). As for the transformers version you are simply modifying the benchmark calls which is an optional dependency, and the lockfile only points to an old version because some package seems to depend on an old version. Regardless we never manually touch the lock file, only pyproject.toml so the solution would need to be in it.
2025-04-01T06:38:59.480723
2023-07-09T12:25:19
1795379083
{ "authors": [ "hughexp" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6847", "repo": "hughexp/jx3-minikube_home", "url": "https://github.com/hughexp/jx3-minikube_home/pull/11" }
gharchive/pull-request
chore: promote jx-aspnet-app-01 to version 0.0.2 this commit will trigger a pipeline to generate the actual kubernetes resources to perform the promotion which will create a second commit on this Pull Request before it can merge jx-aspnet-app-01 Changes in version 0.0.2 Chores release 0.0.2 (jenkins-x-bot) add variables (jenkins-x-bot) Other Changes These commits did not use Conventional Commits formatted messages: Update release.yaml (hughexp) [APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: To complete the pull request process, please assign You can assign the PR to them by writing /assign in a comment when ready. The full list of commands accepted by this bot can be found here. Needs approval from an approver in each of these files: OWNERS Approvers can indicate their approval by writing /approve in a comment Approvers can cancel approval by writing /approve cancel in a comment @hughexp: The following test failed, say /retest to rerun them all: Test name Commit Details Rerun command verify b39f949a1d17fe74fe66617312701be5af6c690e link /test verify Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
2025-04-01T06:38:59.497650
2022-01-09T02:11:18
1097084767
{ "authors": [ "knu" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6848", "repo": "huginn/huginn", "url": "https://github.com/huginn/huginn/issues/3062" }
gharchive/issue
Replace RR with RSpec Mocks RSpec Mocks is widely used nowadays, and having to learn RR in order to write tests is becoming an obstacle to new contributors. https://github.com/kjvarga/rr-to-rspec-converter can be of help.
2025-04-01T06:38:59.514352
2022-02-01T22:06:23
1121248789
{ "authors": [ "BlobCodes" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6849", "repo": "hugopl/gtk4.cr", "url": "https://github.com/hugopl/gtk4.cr/issues/11" }
gharchive/issue
Gdk::Cursor.new_from_name must return Gdk::Cursor but it is returning (Gdk::Cursor | Nil) Trying to create the following cursor: Gdk::Cursor.new_from_name("pointer", nil) Results in the aforementioned compile-time error. The generated method: def self.new_from_name(name : ::String, fallback : Gdk::Cursor?) : self # gdk_cursor_new_from_name: (Constructor) # @fallback: (nullable) # Returns: (transfer full) # Handle parameters fallback = if fallback.nil? Pointer(Void).null else fallback.to_unsafe end # C call _retval = LibGdk.gdk_cursor_new_from_name(name, fallback) # Return value handling Gdk::Cursor.new(_retval, GICrystal::Transfer::Full) unless _retval.null? end This seems to apply to other methods as well (which previously worked): GdkPixbuf::Pixbuf.new_from_file GdkPixbuf::Pixbuf.new_from_file_at_size
2025-04-01T06:38:59.559969
2015-09-11T19:53:47
106082165
{ "authors": [ "cncoleman", "esjewett" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6850", "repo": "humanitiesplusdesign/palladio", "url": "https://github.com/humanitiesplusdesign/palladio/issues/53" }
gharchive/issue
Make it possible to define a bounding box for the component map There are good use cases for a fixed map that displays data, but is not zoomable. This would make the components a bit closer to illustrations. Related to #40. Should be able to tackle both at the same time. This particular setting would probably best be an option added to the component map, along with an option to set the center point and zoom level. This exposes an API for setting center and zoom level. We can do bounds if we want as well, but the functionality is more or less equivalent and possibly more accessible. In the component context: map.getOptions().centerCoordinates([50, 0]); map.getOptions().zoomLevel(6);
2025-04-01T06:38:59.655310
2023-02-21T16:34:00
1593779960
{ "authors": [ "nikspz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6851", "repo": "hummingbot/hummingbot", "url": "https://github.com/hummingbot/hummingbot/issues/6103" }
gharchive/issue
Client - typo in gateway generate-certs message version: development 1.13.0 There's typo in gateway generate-certs message Steps: Clone and install development branch Source or Docker Start the Client execute gateway generate-certs command pay attention to the message Actual: pass phase Expected: pass phrase Fixed by #6119
2025-04-01T06:38:59.660190
2024-04-23T08:25:35
2258264328
{ "authors": [ "Venj-ADL", "neauoire" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6852", "repo": "hundredrabbits/Orca", "url": "https://github.com/hundredrabbits/Orca/issues/297" }
gharchive/issue
Is this repo dead? Seems to be a long time since any dev work happened here. Is this project dead? Just asking so I know whether to allocate time/energy into learning the system in case it will never be updated again. I understand maintaining software is not easy and I am am certainly not expecting anything, just more curious. I saw Orca when it was first released and thought it was super interesting but didn't have the time to invest in it back then. Checking in again and it seems to be abandoned (which is fine, but perhaps mark if so it that is the case :) ). Cheers for all your work so far. It's still very much alive. The javascript implementation is complete, so I don't need to update it, the other implementations(uxn, cli, etc..) still change from time to time, but only rarely. The Orca specification is frozen, the next time this repo will be modified is when the webmidi API changes.
2025-04-01T06:38:59.667207
2016-08-28T00:05:28
173621718
{ "authors": [ "andoks", "ruslo" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6853", "repo": "hunter-packages/googletest", "url": "https://github.com/hunter-packages/googletest/pull/10" }
gharchive/pull-request
Backport fix for building unit-tests for gtest itself with Mingw backport of bugfix for https://github.com/google/googletest/issues/708 from upstream googletest. fix #9 https://ci.appveyor.com/project/andoks/googletest Release: https://github.com/hunter-packages/googletest/releases/tag/1.8.0-hunter-p3 Testing: https://ci.appveyor.com/project/ingenue/hunter/build/1.0.860 https://travis-ci.org/ingenue/hunter/builds/155816959 https://ci.appveyor.com/project/andoks/googletest Well, this test use Visual Studio: -- Building for: Visual Studio 14 2015 -- The C compiler identification is MSVC 19.0.24213.1 -- The CXX compiler identification is MSVC 19.0.24213.1 -- Check for working C compiler using: Visual Studio 14 2015 -- Check for working C compiler using: Visual Studio 14 2015 -- works -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working CXX compiler using: Visual Studio 14 2015 -- Check for working CXX compiler using: Visual Studio 14 2015 -- works MinGW still not working: https://ci.appveyor.com/project/ingenue/hunter/build/1.0.860/job/2rvaic4wy6penu6w I've saved this pull request in branch hunter.pr-10 and reverted hunter branch to the previous state. Sorry, I'm not too familiar with appveyor, and obviously did not take enough time to set up building with MinGW properly. It seems like even though the upstream PR was accepted, it still has not fixed all the issues with building with MinGW on windows (https://github.com/google/googletest/pull/721). An in addition, I'm still running with self-patched hunter (#8), and did not test the cherry-pick locally. If I submit another PR, I'll make sure to set up testing properly with Appveyour. Here is the configuration with AppVeyor testing with MinGW: https://github.com/forexample/hunter-simple/blob/master/appveyor.yml. By the way you can send me pull request with working configuration (or may be even upstream).
2025-04-01T06:38:59.668806
2017-06-29T00:07:51
239331857
{ "authors": [ "Fotospecht", "hunterjm", "marlonespindola", "syros1977" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6854", "repo": "hunterjm/fifa-autobuyer", "url": "https://github.com/hunterjm/fifa-autobuyer/issues/193" }
gharchive/issue
Plans for the future @hunterjm any plans in continuing this auto buyer for FIFA 18?? Most likely. 👍 What this programs needs is a scheduler that lets you configure at what times the program runs, for how many hours, at random times etc. All of my accounts have been banned after some time, even with RPM=4. The autobuyer has to mimic human behaviour better than with the current version. I hope you keep going with this ab for fifa 18! @hunterjm is there any chance to make this work on a mobile device? that would be great.
2025-04-01T06:38:59.719841
2024-10-03T07:34:31
2563337759
{ "authors": [ "gabriel-piles", "mllife" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6857", "repo": "huridocs/pdf-document-layout-analysis", "url": "https://github.com/huridocs/pdf-document-layout-analysis/issues/91" }
gharchive/issue
Reading Order model integration simplification Hello, I want to continue the discussion from here - https://huggingface.co/HURIDOCS/pdf-reading-order/discussions/1 Thanks for all of your work, I am exploring to integrate your code into my project but, need more simplification. Any way to do this without Poppler and do this with some other pdf parsing framework. Is it possible for you to modularise part where its extract pdf stuff and then make a separate code specifically for adding results from segment model and all the other stuff into a pipeline for easy integration with different backbones for parsing. Or just make input as simple as list of segments in a page {contained text, bbox of segment, type of segment} and output as dict{bbox: order}; <did i miss anything else required?> please share a sample jupyter notebook. regards. Thank you for your valuable suggestion. We understand the importance of interchangeability for the parser and appreciate your insight. While making this change would require a significant time investment, it's not currently a top priority due to our ongoing commitments to supporting NGOs in other areas. However, we recognize the potential benefits and encourage you to explore this feature by creating a fork of the project. The codebase is already structured to facilitate this change, particularly within the PdfFeatures class and its initialization. With some additional effort, you should be able to implement the interchangeable parser functionality in a repository fork.
2025-04-01T06:38:59.728426
2023-05-08T12:45:21
1700187783
{ "authors": [ "delihus", "macstepien" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6858", "repo": "husarion/rosbot_xl_manipulation_ros", "url": "https://github.com/husarion/rosbot_xl_manipulation_ros/pull/2" }
gharchive/pull-request
Rewrite joy servo node in cpp bump::minor It was necessary due to performance issues with Python node. Then I try to close gripper in RViz with gripper move group only left gripper joint is moving in simulation. Screencast from 19.05.2023 23:24:14.webm Done: #3 #4 #5
2025-04-01T06:38:59.739011
2023-03-24T20:27:06
1639991330
{ "authors": [ "IbraheemAlSaady", "berengamble", "keremnalbant", "patrick-ve" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6859", "repo": "hwchase17/langchain", "url": "https://github.com/hwchase17/langchain/issues/1976" }
gharchive/issue
StructuredOutputParser - Allow users to get multiple items from response. Allow users to choose the type in the schema (string | List[string]) Allow users to get multiple json objects (get JSON array) in the response. I achieved it by replacing the {{ }} -> [[ ]] as follows: prompt.replace("""{{ "ID": string // IDs which refers to the sentences. "Text": string // Sentences that contains the answer to the question. }}""", """[[ "ID": string // IDs which refers to the sentences. "Text": string // Sentences that contains the answer to the question. ]]""") And got a list of json objects with this method. Very hacky, but this is what I did to have it return an array of objects. output_parser = StructuredOutputParser.from_response_schemas( response_schemas=[ ResponseSchema( name="country_code", description="two letter country code" ), ResponseSchema(name="city", description="city name"), ResponseSchema( name="places", description="""array of of 10 places in the following format: [ {{ "name": string // name of the place', "types": [string] // types of the place' }} ] """, ), ] ) format_instructions =output_parser.get_format_instructions() format_instructions =format_instructions.replace( '"places": string', '"places": array of objects' ) Which generates: ```json { "country_code": string // two letter country code "city": string // city name "places": array of objects // array of of 10 places in the following format: [ { "name": string // name of the place', "types": [string] // types of the place' } ] } I managed to achieve this by passing type to the ResponseSchema. Something like this ResponseSchema( name="someList", description="description to model with the example: [{name: string, other: string}]", type="array(objects)" ) I managed to achieve this by passing type to the ResponseSchema. Something like this ResponseSchema( name="someList", description="description to model with the example: [{name: string, other: string}]", type="array(objects)" ) Is there a TypeScript equivalent of ResponseSchema for Langchain.js? It doesn't look like so, but it would be greatly helpful for the JS/TS community.
2025-04-01T06:38:59.762778
2013-02-08T18:59:45
10791182
{ "authors": [ "bobthecow", "patkan" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6860", "repo": "hwstartup/TheMakerMap.com", "url": "https://github.com/hwstartup/TheMakerMap.com/issues/24" }
gharchive/issue
What other data sets should be included in The Maker Map? Possibilities: 3D printers (private/commercial) Maker events Hackerspaces (from the wiki, check for activity) Things that already live on other maps? ??? Discuss. I think including the locations of all fablabs worldwide would also be beneficial. This data is already organised in the form of a map on fablabs.io
2025-04-01T06:38:59.788431
2017-01-24T03:44:32
202717541
{ "authors": [ "hycis", "shakacs" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6861", "repo": "hycis/Mozi", "url": "https://github.com/hycis/Mozi/issues/3" }
gharchive/issue
How to setup batch_size? There seems no where to set batch_size, as neither datablock nor singleblock class have an input entry for it. Singleblock and datablock is a subclass of dataset so you can just put batch_size as argument to set batch_size https://github.com/hycis/Mozi/blob/master/example/datablocks_example.py#L62 https://github.com/hycis/Mozi/blob/master/example/datablocks_example.py#L62 On 24 Jan 2017, at 11:44 AM, shakacs<EMAIL_ADDRESS>wrote: There seems no where to set batch_size, as neither datablock nor singleblock class have an input entry for it. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/hycis/Mozi/issues/3, or mute the thread https://github.com/notifications/unsubscribe-auth/ADWIiW6EKIZLKajmYZzTBaCryxoxqZkEks5rVXOggaJpZM4Lr1QH. Thanks. The reason i post this issue is that when i tried to run the code in tourist, it throws "'Mnist' object has no attribute 'batch_size'", the same exception thrown once i tried to print the batch_size of a datablock object, and i traced into the source and found that only IterMatrix class has a member called batch_size. May be i got something wrong. Yap, the dataset object itself contains the train, valid and test itermatrix, and dataset itself indeed does not have the attribute of batch_size. However you can call the batch_size by looking into the itermatrix, for example dataset = Mnist(batch_size=32) print(dataset.train.batch_size) On 24 Jan 2017, at 12:13 PM, shakacs<EMAIL_ADDRESS>wrote: Thanks. The reason i post this issue is that when i tried to run the code in tourist, it throws "'Mnist' object has no attribute 'batch_size'", the same exception thrown once i tried to print the batch_size of a datablock object, and i traced into the source and found that only IterMatrix class has a member called batch_size. May be i got something wrong. — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/hycis/Mozi/issues/3#issuecomment-274705250, or mute the thread https://github.com/notifications/unsubscribe-auth/ADWIiVLEI-3rxz8CKYfKLkckV2bPLe0Vks5rVXpTgaJpZM4Lr1QH.
2025-04-01T06:38:59.795341
2022-03-07T19:57:07
1161857978
{ "authors": [ "hydrocheck", "pkdash" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6862", "repo": "hydroshare/hydroshare", "url": "https://github.com/hydroshare/hydroshare/pull/4536" }
gharchive/pull-request
[WIP] 4535 Remove netcdf resource type NOTE: This PR replaces another closed PR #4528 Pull Request Checklist: [ ] Positive Test Case Written by Dev [ ] Automated Testing [ ] Sufficient User and Developer Documentation [ ] Passing Jenkins Build [ ] Peer Code review and approval Positive Test Case [Enter positive test case here] METRIC VALUE https://sonarqube.cuahsi-workstation.com:9000/dashboard?id=hydroshare-4536 @sblack-usu I made a change as per your suggestion. Can you re-approve?
2025-04-01T06:38:59.811652
2019-01-03T02:45:28
395442097
{ "authors": [ "flashery", "fletch3555" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6864", "repo": "hyn/multi-tenant", "url": "https://github.com/hyn/multi-tenant/issues/698" }
gharchive/issue
cURL Error #:Could not resolve host: tenant4.hubshake.localhost Description I always get this error when trying to CURL on the API from wordpress to laravel application. cURL Error #:Could not resolve host: tenant4.hubshake.localhost .. Actual behavior Right now it always showing me this error cURL Error #:Could not resolve host: tenant4.hubshake.localhost .. Expected behavior It should got to the API and call the function in that API and perform the operation inside that function. .. Information hyn/multi-tenant version: 5.* laravel version: 5.5.* database driver and version: mysql webserver software and version: xampp v3.2.2 php version: 7.1 tenancy.php config <?php /* * This file is part of the hyn/multi-tenant package. * * (c) Daniël Klabbers<EMAIL_ADDRESS> * * For the full copyright and license information, please view the LICENSE * file that was distributed with this source code. * * @see https://laravel-tenancy.com * @see https://github.com/hyn/multi-tenant */ use Hyn\Tenancy\Database\Connection; return [ 'models' => [ /** * Specify different models to be used for the global, system database * connection. These are also used in their relationships. Models * used have to implement their respective contracts and * either extend the SystemModel or use the trait * UsesSystemConnection. */ // Must implement \Hyn\Tenancy\Contracts\Customer 'customer' => \Hyn\Tenancy\Models\Customer::class, // Must implement \Hyn\Tenancy\Contracts\Hostname 'hostname' => \Hyn\Tenancy\Models\Hostname::class, // Must implement \Hyn\Tenancy\Contracts\Website 'website' => \Hyn\Tenancy\Models\Website::class ], 'website' => [ /** * Each website has a short random hash that identifies this entity * to the application. By default this id is randomized and fully * auto-generated. In case you want to force your own logic for * when you need to have a better overview of the complete * tenant folder structure, disable this and implement * your own id generation logic. */ 'disable-random-id' => false, /** * The random Id generator is responsible for creating the hash as mentioned * above. You can override what generator to use by modifying this value * in the configuration. * * @warn This won't work if disable-random-id is true. */ // 'random-id-generator' => Hyn\Tenancy\Generators\Uuid\ShaGenerator::class, 'random-id-generator' => App\Tenancy\Generators\DBNameGenerator::class, /** * Enable this flag in case you're using a driver that does not support * database username or database name with a length of more than 32 characters. * * This should be enabled for MySQL, but not for MariaDB and PostgreSQL. */ 'uuid-limit-length-to-32' => env('LIMIT_UUID_LENGTH_32', false), /** * Specify the disk you configured in the filesystems.php file where to store * the tenant specific files, including media, packages, routes and other * files for this particular website. * * @info If not set, will revert to the default filesystem. */ 'disk' => null, /** * Automatically generate a tenant directory based on the random id of the * website. Uses the above disk to store files to override system-wide * files. * * @info set to false to disable. */ 'auto-create-tenant-directory' => true, /** * Automatically rename the tenant directory when the random id of the * website changes. This should not be too common, but in case it happens * we automatically want to move files accordingly. * * @info set to false to disable. */ 'auto-rename-tenant-directory' => true, /** * Automatically deletes the tenant specific directory and all files * contained within. * * @see * @info set to true to enable. */ 'auto-delete-tenant-directory' => false, /** * Time to cache websites in minutes. Set to false to disable. */ 'cache' => 10, ], 'hostname' => [ /** * If you want the multi tenant application to fall back to a default * hostname/website in case the requested hostname was not found * in the database, complete in detail the default hostname. * * @warn this must be a FQDN, these have no protocol or path! */ 'default' => env('TENANCY_DEFAULT_HOSTNAME'), /** * The package is able to identify the requested hostname by itself, * disable to get full control (and responsibility) over hostname * identification. The hostname identification is needed to * set a specific website as currently active. * * @see src/Jobs/HostnameIdentification.php */ 'auto-identification' => env('TENANCY_AUTO_HOSTNAME_IDENTIFICATION', true), /** * In case you want to have the tenancy environment set up early, * enable this flag. This will run the tenant identification * inside a middleware. This will eager load tenancy. * * A good use case is when you have set "tenant" as the default * database connection. */ 'early-identification' => env('TENANCY_EARLY_IDENTIFICATION', false), /** * Abort application execution in case no hostname was identified. This will throw a * 404 not found in case the tenant hostname was not resolved. */ 'abort-without-identified-hostname' => true, /** * Time to cache hostnames in minutes. Set to false to disable. */ 'cache' => 10, ], 'db' => [ /** * The default connection to use; this overrules the Laravel database.default * configuration setting. In Laravel this is normally configured to 'mysql'. * You can set a environment variable to override the default database * connection to - for instance - the tenant connection 'tenant'. */ 'default' => env('TENANCY_DEFAULT_CONNECTION'), /** * Used to give names to the system and tenant database connections. By * default we configure 'system' and 'tenant'. The tenant connection * is set up automatically by this package. * * @see src/Database/Connection.php * @var system-connection-name The database connection name to use for the global/system database. * @var tenant-connection-name The database connection name to use for the tenant database. */ 'system-connection-name' => env('TENANCY_SYSTEM_CONNECTION_NAME', Connection::DEFAULT_SYSTEM_NAME), 'tenant-connection-name' => env('TENANCY_TENANT_CONNECTION_NAME', Connection::DEFAULT_TENANT_NAME), /** * The tenant division mode specifies to what database websites will be * connecting. The default setup is to use a new database per tenant. * In case you prefer to use the same database with a table prefix, * set the mode to 'prefix'. * * @see src/Database/Connection.php */ 'tenant-division-mode' => env('TENANCY_DATABASE_DIVISION_MODE', 'database'), /** * The database password generator takes care of creating a valid hashed * string used for tenants to connect to the specific database. Do * note that this will only work in 'division modes' that set up * a connection to a separate database. */ // 'password-generator' => Hyn\Tenancy\Generators\Database\DefaultPasswordGenerator::class, 'password-generator' => App\Tenancy\Generators\DefaultPasswordGenerator::class, /** * The tenant migrations to be run during creation of a tenant. Specify a directory * to run the migrations from. If specified these migrations will be executed * whenever a new tenant is created. * * @info set to false to disable auto migrating. * * @warn this has to be an absolute path, feel free to use helper methods like * base_path() or database_path() to set this up. */ 'tenant-migrations-path' => database_path('migrations/tenant'), /** * Seeds the newly created tenant database based on this Seeder. * * @info requires tenant-migrations-path to be in use. * * @warn specify a valid fully qualified class name. * @example App\Seeders\AdminSeeder::class */ 'tenant-seed-class' => false, /** * Automatically generate a tenant database based on the random id of the * website. * * @info set to false to disable. */ 'auto-create-tenant-database' => true, /** * Automatically rename the tenant database when the random id of the * website changes. This should not be too common, but in case it happens * we automatically want to move databases accordingly. * * @info set to false to disable. */ 'auto-rename-tenant-database' => true, /** * Automatically deletes the tenant specific database and all data * contained within. * * @info set to true to enable. */ 'auto-delete-tenant-database' => false, ], 'folders' => [ 'config' => [ /** * Merge configuration files from the config directory * inside the tenant directory with the global configuration files. */ 'enabled' => true, /** * List of configuration files to ignore, preventing override of crucial * application configurations. */ 'blacklist' => ['database', 'tenancy', 'webserver'], ], 'routes' => [ /** * Allows adding and overriding URL routes inside the tenant directory. */ 'enabled' => true, /** * Prefix all tenant routes. */ 'prefix' => null, ], 'trans' => [ /** * Allows reading translation files from a trans directory inside * the tenant directory. */ 'enabled' => true, /** * Will override the global translations with the tenant translations. * This is done by overriding the laravel default translator with the new path. */ 'override-global' => true, /** * In case you disabled global override, specify a namespace here to load the * tenant translation files with. */ 'namespace' => 'tenant', ], 'vendor' => [ /** * Allows using a custom vendor (composer driven) folder inside * the tenant directory. */ 'enabled' => true, ], 'media' => [ /** * Mounts the assets directory with (static) files for public use. */ 'enabled' => true, ] ] ]; webserver.php config <?php /* * This file is part of the hyn/multi-tenant package. * * (c) Daniël Klabbers<EMAIL_ADDRESS> * * For the full copyright and license information, please view the LICENSE * file that was distributed with this source code. * * @see https://laravel-tenancy.com * @see https://github.com/hyn/multi-tenant */ return [ /** * Apache2 is one of the most widely adopted webserver packages available. * * @see http://httpd.apache.org/docs/ * @see https://www.digitalocean.com/community/tutorials/how-to-install-linux-apache-mysql-php-lamp-stack-on-ubuntu */ 'apache2' => [ /** * Whether the integration with Apache2 is currently active. */ 'enabled' => false, /** * Define the ports of your Apache service. */ 'ports' => [ /** * HTTP, non-SSL port. * * @default 80 */ 'http' => 80, /** * HTTPS, SSL port. * * @default 443 */ 'https' => 443 ], /** * The generator taking care of hooking into the Apache services and files. */ 'generator' => \Hyn\Tenancy\Generators\Webserver\Vhost\ApacheGenerator::class, /** * The view that holds the vhost configuration template. */ 'view' => 'tenancy.generators::webserver.apache.vhost', /** * Specify the disk you configured in the filesystems.php file where to store * the tenant vhost configuration files. * * @info If not set, will revert to the default filesystem. */ 'disk' => null, 'paths' => [ /** * Location where vhost configuration files can be found. */ 'vhost-files' => [ '/etc/apache2/sites-enabled/' ], /** * Actions to run to work with the Apache2 service. */ 'actions' => [ /** * Action that asserts Apache2 is installed. */ 'exists' => '/etc/init.d/apache2', /** * Action to run to test the apache configuration. */ 'test-config' => 'apache2ctl -t', /** * Action to run to reload the apache service. */ 'reload' => 'apache2ctl graceful' ] ] ], /** * Nginx webserver support. * * @see http://nginx.org */ 'nginx' => [ /** * Whether the integration with nginx is currently active. */ 'enabled' => false, /** * The php sock to be used. */ 'php-sock' => 'unix:/var/run/php/php7.0-fpm.sock', /** * Define the ports of your nginx service. */ 'ports' => [ /** * HTTP, non-SSL port. * * @default 80 */ 'http' => 80, /** * HTTPS, SSL port. * * @default 443 */ 'https' => 443 ], /** * The generator taking care of hooking into the nginx services and files. */ 'generator' => \Hyn\Tenancy\Generators\Webserver\Vhost\NginxGenerator::class, /** * The view that holds the vhost configuration template. */ 'view' => 'tenancy.generators::webserver.nginx.vhost', /** * Specify the disk you configured in the filesystems.php file where to store * the tenant vhost configuration files. * * @info If not set, will revert to the default filesystem. */ 'disk' => null, 'paths' => [ /** * Location where vhost configuration files can be found. */ 'vhost-files' => [ '/etc/nginx/sites-enabled/' ], /** * Actions to run to work with the Nginx service. */ 'actions' => [ /** * Action that asserts nginx is installed. */ 'exists' => '/etc/init.d/nginx', /** * Action to run to test the nginx configuration. */ 'test-config' => '/etc/init.d/nginx configtest', /** * Action to run to reload the nginx service. */ 'reload' => 'systemctl restart nginx' ] ] ] ]; Then that would be why. Please read the article I linked. A subdomain is a full domain, but with a bit of syntactic sugar. It still needs to be explicitly defined somewhere, which you have not done yet ahhh now I know.. thank you so much man... No problem
2025-04-01T06:38:59.818284
2022-02-04T14:30:38
1124266996
{ "authors": [ "kasperisager", "mafintosh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6865", "repo": "hypercore-protocol/hypercore-next", "url": "https://github.com/hypercore-protocol/hypercore-next/pull/83" }
gharchive/pull-request
Use turbo-hash-map to avoid stringifying keys Depends on https://github.com/mafintosh/turbo-hash-map/pull/3 to maintain browser compatibility. Moved to protomux
2025-04-01T06:38:59.826621
2017-08-14T14:21:16
250045708
{ "authors": [ "bergwolf", "laijs" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6866", "repo": "hyperhq/runv", "url": "https://github.com/hyperhq/runv/pull/554" }
gharchive/pull-request
kill two todos in deleteContainer operation Also clean up glog trailing newline a bit in cli/container.go. @laijs updated. LGTM
2025-04-01T06:38:59.827129
2015-05-22T05:36:29
79297900
{ "authors": [ "nstoddard", "seanmonstar" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6867", "repo": "hyperium/hyper", "url": "https://github.com/hyperium/hyper/issues/541" }
gharchive/issue
Make OpenSSL an optional dependency It can be tricky to install OpenSSL, especially on Windows. It would be nice if it were an optional dependency. Closed via #577
2025-04-01T06:38:59.831859
2020-09-25T10:58:44
708845536
{ "authors": [ "LucioFranco", "alce", "gardk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6868", "repo": "hyperium/tonic", "url": "https://github.com/hyperium/tonic/issues/463" }
gharchive/issue
Allow custom error types in RPC methods Feature Request Motivation One of the things I love about Rust is it's stellar error-handling capabilities and how it helps write correct code. When writing RPC implementations with this library I've noticed excessive and duplicated use of map_err or unwrap and no custom error types used in those functions. For some functions with lots of fallible operations, or when there's several functions that can return many of the same errors. I've extracted those out into separate functions and defined a custom error type so I can use ? and have some centrally defined error handling strategy for the common ones. This works, but being able to do this straight in the impl would be really nice. Proposal As I see it there's two possibilities here: Bake some magic into the tonic::async_trait macro (or provide a separate one, maybe a function attribute) that wraps the implementation functions in an outer function that returns tonic::Status error like before, but coerces the actual return type based on some trait. Allow the user to define an error type for each generated function in the trait and convert it in the generated service stub code. This one I'm not that big a fan of cause it currently requires defining the type for every trait function until associated_type_defaults lands... The trait used for conversion I think could look something like this: trait ToStatus { fn code(&self) -> Code; fn status(&self) -> Status { let code = self.code(); Status::new(code, code.description()) } } impl<T: ToStatus> From<T> for Status { fn from(custom: T) -> Self { custom.status() } } // example impl impl ToStatus for Code { fn code(&self) -> Code { *self } } What do you guys think? I'd be more than happy to attempt implementing this myself, but I've never written any real macro code before. :) Thanks for reading! I've cobbled together an attribute macro that seems to get the job done: gardk/tonic_rpc_wrapper There is some upcoming work to improve internal error handling and I have been thinking about ways to make it easier to use richer error model with google.rpc.Status. It may be worth it to look into all this, including your proposal, in an unified way. It will take a little while though, so thank you for your patience. Personally, I share your observation about Status conversions. I have tried a couple of approaches, none of which I completely like. It's a minor thing and I'm not sure what can be done about it or how but I believe it's worth considering. You likely want to take advantage of Try to convert errors into Status I believe that should work. Usually, what I see is people will have custom code to convert their internal errors into Status, I think this is the correct method. Thank you for the responses! I look forward to seeing any future work you may have in store for this. As for the comment about using Try. This will only work directly (without Result::map_error) when the fallible function being called returns a first-party error if I'm not mistaken? What prompted me posting this issue was that I wanted to treat Diesel's errors the same for many of my handlers within a service. And there is no direct conversion from those into Status that I'm aware of. What I ended up doing was mapping any handled third-party errors into a custom error type, then using ? for conversion to Status, which is almost the same. However, it looks like you're aware of this little snag and may be looking into it, which was really all I wanted :) @gardk Yeah, the intermediate error type imo is great too. I think its always good to funnel errors into a "application" defined one that then can get translated into say Status or some other http style error. So I think that is the correct solution!
2025-04-01T06:38:59.833236
2024-10-17T21:47:46
2595821657
{ "authors": [ "tottoto" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6869", "repo": "hyperium/tonic", "url": "https://github.com/hyperium/tonic/pull/2014" }
gharchive/pull-request
chore: Start development of version 0.13 Motivation Starts the development of version 0.13. This is needed to tell cargo-semver-checks that the development includes breaking changes. Solution Updates versions to 0.13.0-dev. I opened this pull request as I think we probably could start development of the next version including breaking changes as @LucioFranco merged #1892 which includes breaking changes. This change is needed for the pull requests such as #2006 and #2013.
2025-04-01T06:38:59.850056
2020-10-20T17:49:01
725812542
{ "authors": [ "SaidShah", "litong01" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6870", "repo": "hyperledger-labs/minifabric", "url": "https://github.com/hyperledger-labs/minifabric/issues/104" }
gharchive/issue
Getting the history of transactions Hello, I know if we run minifab blockquery it will generate the transactions for the last block, but is there a way we can generate the history of all transactions or the history of all blocks and transactions? Use minifab blockquery -b command to get transactions of the block. Please close this issue. -b blocknumber
2025-04-01T06:38:59.852233
2021-08-19T16:35:51
974848258
{ "authors": [ "Patrik-Stas" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6871", "repo": "hyperledger/aries-vcx", "url": "https://github.com/hyperledger/aries-vcx/issues/338" }
gharchive/issue
Remove dependency on indy module from protocol module Module protocols should be a "codification" of state machines as defines by Aries RCFs. These should not directly tie to libindy or any specific type of credentials or frameworks. This was solved a while ago by hiding indy dependencies behind traits. Closing
2025-04-01T06:38:59.857321
2022-11-18T07:26:21
1454612413
{ "authors": [ "Gabriel-Trintinalia" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6872", "repo": "hyperledger/besu", "url": "https://github.com/hyperledger/besu/pull/4701" }
gharchive/pull-request
Start privacyCluster non-bootnodes in parallel (acceptance test improvement) Signed-off-by: Gabriel Trintinalia<EMAIL_ADDRESS>PR description Start non-bootnodes in parallel for privacy clusters (acceptance test only) may reduce some acceptance tests duration by 15% Documentation [x] I thought about documentation and added the doc-change-required label to this PR if updates are required. Changelog [x] I thought about the changelog and included a changelog update if required. @Gabriel-Trintinalia any reason not to merge this change? is there more to do? It improved the tests on my machine but I had no data to validate on CircleCi. I think we can merge, it definitely does not hurt.
2025-04-01T06:38:59.858933
2023-05-29T05:13:02
1730075757
{ "authors": [ "macfarla", "non-fungible-nelson", "pinges" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6873", "repo": "hyperledger/besu", "url": "https://github.com/hyperledger/besu/pull/5507" }
gharchive/pull-request
Revert the revert of the tx selection commit The original PR (#5396) caused a problem when running on Frontier. The world state update was committed too late and the receipt was created with the old world state root hash. @pinges shall we resolve and merge? To check for no side effects on frontier block syncing, started a full sync node with this PR and one on latest main commit https://app.circleci.com/pipelines/github/hyperledger/besu/22489/workflows/5495f46f-5114-4cc0-b4d3-46280dd735ef/jobs/139706/artifacts Both have got past block 200,000 so have synced frontier blocks successfully.
2025-04-01T06:38:59.862095
2024-04-04T16:22:19
2225980125
{ "authors": [ "bestbeforetoday" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6874", "repo": "hyperledger/fabric-admin-sdk", "url": "https://github.com/hyperledger/fabric-admin-sdk/pull/192" }
gharchive/pull-request
Update dependencies to address GO-2024-2687 Disable automatic update of dependencies that do not address security vulnerabilities. These updates appear to regularly mess up the Go dependencies (go.mod/go.sum). grpc.Dial() and grpc.DialContext() are deprecated in current gRPC versions, and replaced by grpc.NewClient(). LGTM, btw @bestbeforetoday , grpc changes some behaviors in the NewClient, any impact? One place in the code used grpc.DialContent(). This tries to establish the network connection immediately rather than waiting until the connection is actually used to make a request. In practice I don't think this makes any noticeable difference. It just means that an existing network issue will cause an error when a request is made instead when the connection is created. A network issue can occur at any time so errors could always happen when a request is made, even if the initial connect was successful.
2025-04-01T06:38:59.869452
2020-06-09T18:19:12
635659157
{ "authors": [ "stephyee" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:6875", "repo": "hyperledger/fabric-samples", "url": "https://github.com/hyperledger/fabric-samples/pull/202" }
gharchive/pull-request
Rename marbles_transfer to asset-transfer-secured-agreement Rename marbles_transfer to a more generic and streamlined asset Move sample to new directory Remove TODO for marbles by size query The images in the readme still uses marbles as the asset in the diagrams. In a follow-up we can generate new diagrams.