id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2749549144
Test Issue - jhoos Test Body - adpujhzffd Test Comment - xvqli
gharchive/issue
2024-12-19T08:06:55
2025-04-01T06:45:33.086866
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/56792", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2750623539
Test Issue - kofuf Test Body - mpgyavjrka Test Comment - tanou
gharchive/issue
2024-12-19T15:04:39
2025-04-01T06:45:33.087853
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/56802", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1279856467
Test Issue - tutek Test Body - lhqkpxlpyw Test Comment - haoyy
gharchive/issue
2022-06-22T09:35:42
2025-04-01T06:45:33.088592
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/6008", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1334898577
Test Issue - opxwl Test Body - vbwkiwqcth Test Comment - vikoy
gharchive/issue
2022-08-10T16:44:07
2025-04-01T06:45:33.089297
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/9942", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2001883206
Fix additionalRuleLabels default type Objective Fix additionalRuleLabels default type Why Helm report warnings when additionalRuleLabels is customized due to incorrect default variable type. coalesce.go:289: warning: destination for prometheus-rules.prometheus-postgresql-alerts-chart.global.additionalRuleLabels is a table. Ignoring non-table value ([]) coalesce.go:289: warning: destination for prometheus-rules.prometheus-rds-alerts-chart.global.additionalRuleLabels is a table. Ignoring non-table value ([]) How Use dict as default value for additionalRuleLabels Release plan [ ] Merge this PR [ ] Apply with CI #sre
gharchive/pull-request
2023-11-20T10:37:38
2025-04-01T06:45:33.145477
{ "authors": [ "vmercierfr" ], "repo": "qonto/database-monitoring-framework", "url": "https://github.com/qonto/database-monitoring-framework/pull/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1094864438
iris shader not working I use the seus hrr 2.1 shader and this happened What do you mean by hrr? no idea the shader is here https://www.patreon.com/posts/45141775 It’s Half Resolution Rendering, aka rendering at half the resolution and upscaling with tricks to make it look good PTGI shader is not supported
gharchive/issue
2022-01-06T00:50:30
2025-04-01T06:45:33.156424
{ "authors": [ "IMS212", "mmmmmmmmmmmmmmmmaly", "qouteall" ], "repo": "qouteall/ImmersivePortalsMod", "url": "https://github.com/qouteall/ImmersivePortalsMod/issues/955", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1598454113
🛑 otakusan is down In 1a0f285, otakusan (https://otakusan.net/LightNovel) was down: HTTP code: 0 Response time: 0 ms Resolved: otakusan is back up in 03a5b1c.
gharchive/issue
2023-02-24T11:27:19
2025-04-01T06:45:33.210004
{ "authors": [ "quanhieu" ], "repo": "quanhieu/alive_up", "url": "https://github.com/quanhieu/alive_up/issues/1307", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2119581648
🛑 kokocon is down In 453f501, kokocon (https://tracker.kokocon.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: kokocon is back up in c16e499 after .
gharchive/issue
2024-02-05T22:30:29
2025-04-01T06:45:33.212539
{ "authors": [ "quanhieu" ], "repo": "quanhieu/alive_up", "url": "https://github.com/quanhieu/alive_up/issues/9471", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2129220449
🛑 otakusan is down In fd67e25, otakusan (https://otakusan.net/LightNovel) was down: HTTP code: 0 Response time: 0 ms Resolved: otakusan is back up in 6be9b86 after 6 minutes.
gharchive/issue
2024-02-11T23:22:36
2025-04-01T06:45:33.214868
{ "authors": [ "quanhieu" ], "repo": "quanhieu/alive_up", "url": "https://github.com/quanhieu/alive_up/issues/9698", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2515927700
App crashes while streaming data & converting to markdown When we stream data (in my case from a chatbot) to the frontend and then parse the markdown using markdown-to-jsx if the chatbot starts outputting a code block the app would crash (only for HTML). This happens as: Since we stream the responses from chatbot the chatbot starts the code with opening the code block and then starts writing its HTML but since the code block isn't closed markdown-to-jsx detects it as plain HTML and renders it inside a <p> tag which causes DOM errors as the <html> tag cannot be a child of the <p> tag. I believe markdown-to-jsx should detect the starting of the code block and till the code block is not closed by '```' it should just detect the chars after the starting as the code. We're facing a similar issue where invalid markdown causes a CPU blowout. We essentially need a way to handle markdown which contains invalid html/markdown. Perhaps there's a way to strip this invalid markup or escape it?
gharchive/issue
2024-09-10T09:36:35
2025-04-01T06:45:33.217040
{ "authors": [ "ItzCrazyKns", "willtpwise" ], "repo": "quantizor/markdown-to-jsx", "url": "https://github.com/quantizor/markdown-to-jsx/issues/602", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
171735959
ENH: Remove dropna from beta function Sometimes returns have an np.nan value which causes beta to return np.nan. In the case that either contains np.nan, calculate covariance of only those dates that have both returns and benchmark return values. ====================================================================== FAIL: test_max_drawdown_transformation_1 (empyrical.tests.test_stats.TestStats) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/travis/miniconda/envs/testenv/lib/python3.5/site-packages/nose_parameterized/parameterized.py", line 365, in standalone_func return func(*(a + p.args), **p.kwargs) File "/home/travis/build/quantopian/empyrical/empyrical/tests/test_stats.py", line 188, in test_max_drawdown_transformation assert constant*max_dd <= transformed_dd AssertionError ====================================================================== FAIL: test_max_drawdown_transformation_4 (empyrical.tests.test_stats.TestStats) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/travis/miniconda/envs/testenv/lib/python3.5/site-packages/nose_parameterized/parameterized.py", line 365, in standalone_func return func(*(a + p.args), **p.kwargs) File "/home/travis/build/quantopian/empyrical/empyrical/tests/test_stats.py", line 190, in test_max_drawdown_transformation assert constant*max_dd >= transformed_dd AssertionError These seem unrelated, but why are we getting them? PR looks good to me. Would be good to have a test with nans here.
gharchive/pull-request
2016-08-17T19:07:39
2025-04-01T06:45:33.219041
{ "authors": [ "ahgnaw", "twiecki" ], "repo": "quantopian/empyrical", "url": "https://github.com/quantopian/empyrical/pull/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
607314583
Feature request: Ability to get changed df without the filters I want to use QGrid to present a table to a user and give him the ability to edit certain things. The tables sometimes can be very big and the user will do a lot of filtering and each time will change other rows. My problem is, that if the last filter will not be cleared, I will not be able to detect all the changes. I will be able to see only the rows that presented on screen using the get_changed_df function but maybe the user did also other changes previously. My request is to expose a new function get_unfiltered_changed_df that will give me all the changes done to the table but without the filtering (maybe even without the sorting, but this is less important). Another option is to have new flags in the get_changed_df(filtered=True, sorted=True) function which by default will behave the same as today but will give the option to get unfiltered/unsorted df. Thanks. I see that already today it exist in a private variable and can be accessed grid._unfiltered_df but it's not officially supported. Also, if no official function for this, the internal grid._unfiltered_df might not be supported in later versions I think this function would be very important. Without this the user is obliged to manually reset all the filters before saving the work. This is really awkward and error prone. I am using this for a visualization and I saved partial results several times before understanding that i forot to remove a filter I agree with @alucab Not being able to get a view of the edited, but unfiltered version of the DataFrame forces me to disallow filtering in situations where I want to post the differences between the original df and the changed_df. I like qgrid a lot already. Adding this feature would help me out a lot, though. _unfiltered_df does not reflect edits
gharchive/issue
2020-04-27T07:41:27
2025-04-01T06:45:33.223327
{ "authors": [ "HariniMadhi", "Schlendrian", "alucab", "motcke", "rachel-chocron" ], "repo": "quantopian/qgrid", "url": "https://github.com/quantopian/qgrid/issues/313", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
477460360
MAINT: Add Tests to new XWAR Calendar https://github.com/quantopian/trading_calendars/pull/61 https://github.com/quantopian/trading_calendars/pull/74
gharchive/issue
2019-08-06T15:50:32
2025-04-01T06:45:33.224781
{ "authors": [ "dmichalowicz", "gmanoim-quantopian" ], "repo": "quantopian/trading_calendars", "url": "https://github.com/quantopian/trading_calendars/issues/62", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
789536581
Error in /pitch-a-talk Error in quantsquills.com Error in /pitch-a-talk Bad Gateway View on Bugsnag Stacktrace src/components/TalkPitchForm.js:54 - v/l</< src/components/TalkPitchForm.js:48 - promise callback*v/l< src/components/useForm.js:78 - BoRJ/t.a/f< .cache/production-app.js:181 - UxWs/</</< .cache/production-app.js:179 - UxWs/</< .cache/production-app.js:153 - promise callback*UxWs/< .cache/production-app.js:39 - promise callback*UxWs View full stacktrace Created automatically via Bugsnag Sendgrid appeared to have deleted our API key for some reason. New one created, site re-built, all working fine. Sendgrid appeared to have deleted our API key for some reason. New one created, site re-built, all working fine.
gharchive/issue
2021-01-20T01:26:19
2025-04-01T06:45:33.228125
{ "authors": [ "drzax" ], "repo": "quantsquills/quantsquills.com", "url": "https://github.com/quantsquills/quantsquills.com/issues/296", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
515637680
[KED-1129] Nodes with a list of arguments that includes a repeated argument throw a ValueError when running Description If a node argument list contains the same argument more than once, then an error is thrown rather than the node's func being called with the repeated arg in the appropriate places. Steps to Reproduce Simple example illustrating the problem: def add(a, b): return a + b pipeline = Pipeline( [ node(add, ["a", "b"], outputs="a_and_b", name="add_a_b"), node(add, ["a", "a"], outputs="twice_a", name="add_a_a"), node(add, ["b", "b"], outputs="twice_b", name="add_b_b"), ] ) I would expect this to work fine, but instead produces the following error: ValueError: Node add_a_a: add([a,a]) -> [twice_a] expected 2 input(s) ['a', 'a'], but got the following 1 input(s) instead: ['a']. Possible fix In node.Node._run_with_list there is the following check that seems to be too strict: https://github.com/quantumblacklabs/kedro/blob/2bafa6f032d88f0ce7e2cb004da419a4762e2f49/kedro/pipeline/node.py#L432 If changed to just if not all_available: ... then the code above works successfully, but I have not tested whether this causes problems elsewhere. Your Environment Kedro version used (pip show kedro or kedro -V): 0.15.1 Python version used (python -V): Python 3.6.9 :: Anaconda, Inc. Operating system and version: macOS Mojave Hi @simon-b thanks for raising this issue - it's been on our backlog to investigate so I suspect we'll move it up on our priority list. Updated with the relevant internal ticket number to make it easier to track. Closing this as resolved in commit https://github.com/quantumblacklabs/kedro/commit/8f076cf9ba48edadde1227893b313d6bc6f0fa18 Thanks again for raising this!
gharchive/issue
2019-10-31T18:03:09
2025-04-01T06:45:33.236227
{ "authors": [ "lorenabalan", "simon-b" ], "repo": "quantumblacklabs/kedro", "url": "https://github.com/quantumblacklabs/kedro/issues/159", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
975559708
Allow initialization of MeasurementKeys with empty strings An empty string is still a string and is used occasionally as a placeholder in user code. We should support creation of measurement keys with empty strings. Question to cirq-maintainers. Is this sufficient to prevent problems with backward incompatibility or should we roll back instead? Can you add a reference to the PR that introduced the change so we have it on the record. GitHub will also show this PR to any future viewers of the previous one The PR introducing this change was intentionally breaking since There was inconsistency between cirq.measure(q) and cirq.MeasurementGate().on(q) for some qubit q. The former would assign the measurement key q while the latter would have an empty string. We decided(1, 2) that dynamically remapping empty keys based on what qubits are operated on was not a desirable behavior. So, either we could keep this inconsistency, change cirq.measure to stop assigning the "default" key or disallow empty keys in MeasurementGate. We went with the last option since it seemed that empty keys are not really much useful in real world circuits and simulations. Since our Results are keyed by a measurement key string, it doesn't seem too radical to make the key argument sort-of required when constructing measurements. @smitsanghavi At the very least, we have to at least temporarily roll this change back, since a significant amount of internal code uses the empty string as a placeholder for "anonymous" measurements. If we want to re-break this, we need to migrate that code first.
gharchive/pull-request
2021-08-20T12:20:07
2025-04-01T06:45:33.240991
{ "authors": [ "dstrain115", "mpharrigan", "smitsanghavi" ], "repo": "quantumlib/Cirq", "url": "https://github.com/quantumlib/Cirq/pull/4445", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
869275426
Add FQE docs to devsite Self-explanatory from title. cc @lamberta if you have time for a quick review. The goal is to include the FQE docs on https://quantumai.google/openfermion under a new "FQE" tab. I followed how this is done for ReCirq and Cirq - see https://github.com/quantumlib/OpenFermion-FQE/pull/81 for more details. To pull docs from the OpenFermion-FQE repo you'll need to set up separate docs infra for that project. It is much easier to keep the source of those docs in this repo and use the existing nightly jobs. To pull docs from the OpenFermion-FQE repo you'll need to set up separate docs infra for that project. It is much easier to keep the source of those docs in this repo and use the existing nightly jobs. Ahah, thanks for this 🙏 @ncrubin would you prefer to Migrate fqe/docs into this repo as suggested? Set up separate docs infra for FQE? Other? (1) is easy and I can do this (today). (2) I don't know how to do at the moment. Pulled in the FQE notebooks as we discussed, PTAL. Note for posterity: I did not write 1800+ lines here, just moved the notebooks from https://github.com/quantumlib/openfermion-fqe.
gharchive/pull-request
2021-04-27T21:22:50
2025-04-01T06:45:33.246131
{ "authors": [ "lamberta", "rmlarose" ], "repo": "quantumlib/OpenFermion", "url": "https://github.com/quantumlib/OpenFermion/pull/724", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1686650574
Support Milestoned and Demilestoned events for Pull Requests. As per title. Currently milestones events are supported for issues only. Although I'm not sure if there is really an event specific to PRs. How I'm currently handling my Pull Request check through GHIssue. private fun checkPullRequest(issue: GHIssue) { if (issue !is GHPullRequest && !issue.isPullRequest) { return } val pullRequest = if (issue is GHPullRequest) { issue } else { issue.repository.getPullRequest(issue.number) } val repository = pullRequest.repository val headSha = pullRequest.head.sha // Create a new Check Run and show it as pending in the UI val checkRun = repository.createCheckRun(AppBundle["check.pr.milestone.name"], headSha) .withExternalID(milestoneCheckId) .withStartedAt(Date()) .withStatus(Status.IN_PROGRESS) .create() updateCheckRun(issue, checkRun.update()) } Closing as GitHub doesn't deliver a milestoned or demilestoned event specific to PRs. These events are for issues only, from which one can retrieve the shadowed PR if necessary.
gharchive/issue
2023-04-27T11:24:52
2025-04-01T06:45:33.252915
{ "authors": [ "lppedd" ], "repo": "quarkiverse/quarkus-github-app", "url": "https://github.com/quarkiverse/quarkus-github-app/issues/472", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1072289128
App won't start with quarkus 2.5.1.Final + quarkus-logging-logback, it causes an application startup error Using quarkus version: implementation 'io.quarkiverse.logging.logback:quarkus-logging-logback:0.5.0' quarkusPluginVersion=2.5.1.Final quarkusPlatformVersion=2.5.1.Final It causes an application startup error: LoggingLogbackProcessor#removeSlf4jBinding threw an exception: java.lang.NoSuchMethodError: io.quarkus.deployment.builditem.RemovedResourceBuildItem. STACKTRACE 2021-12-06 10:30:45,561 ERROR [io.qua.dep.dev.IsolatedDevModeMain] (main) Failed to start quarkus: java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkiverse.logging.logback.deployment.LoggingLogbackProcessor#removeSlf4jBinding threw an exception: java.lang.NoSuchMethodError: io.quarkus.deployment.builditem.RemovedResourceBuildItem.<init>(Lio/quarkus/bootstrap/model/AppArtifactKey;Ljava/util/Set;)V at io.quarkiverse.logging.logback.deployment.LoggingLogbackProcessor.removeSlf4jBinding(LoggingLogbackProcessor.java:70) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$2.execute(ExtensionLoader.java:887) at io.quarkus.builder.BuildContext.run(BuildContext.java:277) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:834) at org.jboss.threads.JBossThread.run(JBossThread.java:501) at io.quarkus.runner.bootstrap.AugmentActionImpl.runAugment(AugmentActionImpl.java:330) at io.quarkus.runner.bootstrap.AugmentActionImpl.createInitialRuntimeApplication(AugmentActionImpl.java:252) at io.quarkus.runner.bootstrap.AugmentActionImpl.createInitialRuntimeApplication(AugmentActionImpl.java:60) at io.quarkus.deployment.dev.IsolatedDevModeMain.firstStart(IsolatedDevModeMain.java:93) at io.quarkus.deployment.dev.IsolatedDevModeMain.accept(IsolatedDevModeMain.java:450) at io.quarkus.deployment.dev.IsolatedDevModeMain.accept(IsolatedDevModeMain.java:67) at io.quarkus.bootstrap.app.CuratedApplication.runInCl(CuratedApplication.java:149) at io.quarkus.bootstrap.app.CuratedApplication.runInAugmentClassLoader(CuratedApplication.java:105) at io.quarkus.deployment.dev.DevModeMain.start(DevModeMain.java:145) at io.quarkus.deployment.dev.DevModeMain.main(DevModeMain.java:63) Caused by: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkiverse.logging.logback.deployment.LoggingLogbackProcessor#removeSlf4jBinding threw an exception: java.lang.NoSuchMethodError: io.quarkus.deployment.builditem.RemovedResourceBuildItem.<init>(Lio/quarkus/bootstrap/model/AppArtifactKey;Ljava/util/Set;)V at io.quarkiverse.logging.logback.deployment.LoggingLogbackProcessor.removeSlf4jBinding(LoggingLogbackProcessor.java:70) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$2.execute(ExtensionLoader.java:887) at io.quarkus.builder.BuildContext.run(BuildContext.java:277) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:834) at org.jboss.threads.JBossThread.run(JBossThread.java:501) at io.quarkus.builder.Execution.run(Execution.java:116) at io.quarkus.builder.BuildExecutionBuilder.execute(BuildExecutionBuilder.java:79) at io.quarkus.deployment.QuarkusAugmentor.run(QuarkusAugmentor.java:161) at io.quarkus.runner.bootstrap.AugmentActionImpl.runAugment(AugmentActionImpl.java:328) ... 9 more Caused by: java.lang.NoSuchMethodError: io.quarkus.deployment.builditem.RemovedResourceBuildItem.<init>(Lio/quarkus/bootstrap/model/AppArtifactKey;Ljava/util/Set;)V at io.quarkiverse.logging.logback.deployment.LoggingLogbackProcessor.removeSlf4jBinding(LoggingLogbackProcessor.java:70) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$2.execute(ExtensionLoader.java:887) at io.quarkus.builder.BuildContext.run(BuildContext.java:277) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:834) at org.jboss.threads.JBossThread.run(JBossThread.java:501) I have released 0.6.0: https://github.com/quarkiverse/quarkus-logging-logback/pull/24 Might take a while to get to central, but should be good to go in a few hours. Oops, I messed up the 0.6.0 release, so have released 0.7.0 yeap! it works pretty good! thank you @stuartwdouglas
gharchive/issue
2021-12-06T15:33:26
2025-04-01T06:45:33.257367
{ "authors": [ "eriknyk", "stuartwdouglas" ], "repo": "quarkiverse/quarkus-logging-logback", "url": "https://github.com/quarkiverse/quarkus-logging-logback/issues/22", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1903463222
Pull request 487 fix Fix for https://github.com/quarkiverse/quarkus-openapi-generator/pull/487 @hbelmiro I wasn’t sure if I should wait for @ricardozanini first, so just to avoid any misunderstanding, I created the PR. Please let me know if there's something else I can help with. Sh*t, I was wondering why I got conflicts and I just realized, I didn't create the PR from main (I thought I switched before creating the branch, but I didn't), so if that's a problem, I can delete this one and create another PR. @patr1kt0th It's not a problem if it's effectively changing what it should change. But if you want to fix it anyway, you can just rebase your branch.
gharchive/pull-request
2023-09-19T17:46:40
2025-04-01T06:45:33.259899
{ "authors": [ "hbelmiro", "patr1kt0th" ], "repo": "quarkiverse/quarkus-openapi-generator", "url": "https://github.com/quarkiverse/quarkus-openapi-generator/pull/491", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1633350555
Fix cycle decorator in Dekorate Relates to CI failures cc @metacosm Doesn't look like the tests are passing and the build is still not failing… Maybe there's an OLM install action now? Maybe there's an OLM install action now? I don't know what you mean or how aN OLM install action might help here. The problem is that the exit 1 from a custom script is not being propagated to the process and hence the step is not failing. And I don't know how to fix this issue yet. it seems that the process is providing the right exit status now. It's time to actually fix the issue. I'm investigating this. operatorhubio-catalog-cdmlc 0/1 CrashLoopBackOff 6 8m58s this seems to be the issue. @metacosm it seems the issue was caused by an incompatibility of the latest version of OLM and Kind. Using 0.23.0 works fine: https://github.com/quarkiverse/quarkus-operator-sdk/pull/537/files#diff-696f0cdcddf966043bf093f3677ce9ff73b8b9342ca1110cd4575e92153e9c9fR62 I'm changing the PR to ready for review.
gharchive/pull-request
2023-03-21T07:18:06
2025-04-01T06:45:33.263744
{ "authors": [ "Sgitario", "metacosm" ], "repo": "quarkiverse/quarkus-operator-sdk", "url": "https://github.com/quarkiverse/quarkus-operator-sdk/pull/537", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2377366838
Add a special entry to make this compatible with Quarkus 3.9+ @ia3andy I found this commit but TBH I'm not sure whether I actually tested it locally or so... It works and should be released asap to be in Quarkus 3.12 (and also maybe backported in previous)! I have no idea how to backport something in the platform? CC @gsmet @mkouba I think you need to set the new version in the platform, then backport that PR
gharchive/pull-request
2024-06-27T07:23:18
2025-04-01T06:45:33.265431
{ "authors": [ "ia3andy", "mkouba" ], "repo": "quarkiverse/quarkus-qute-web", "url": "https://github.com/quarkiverse/quarkus-qute-web/pull/94", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
39563310
Labels Hi, How do I display labels for ranges in bar units without 3rd party plugins? Thanks Thanks mate
gharchive/issue
2014-08-05T20:45:28
2025-04-01T06:45:33.324499
{ "authors": [ "KawiNinja", "praneybehl" ], "repo": "quarterto/Elessar", "url": "https://github.com/quarterto/Elessar/issues/73", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
149735546
Fix regression regex Running regression tests, I got the error: [debug] Running TaskDef(quasar.StreamingQueryRegressionSpec, specs2 Specification fingerprint, false, [SuiteSelector]) java.nio.charset.MalformedInputException: Input length = 2 at java.nio.charset.CoderResult.throwException(CoderResult.java:281) at sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:339) at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178) at java.io.InputStreamReader.read(InputStreamReader.java:184) at java.io.BufferedReader.read1(BufferedReader.java:210) at java.io.BufferedReader.read(BufferedReader.java:286) at java.io.Reader.read(Reader.java:140) at scala.io.BufferedSource.mkString(BufferedSource.scala:96) at quasar.regression.QueryRegressionTest.quasar$regression$QueryRegressionTest$$$anonfun$46(QueryRegressionTest.scala:262) at quasar.regression.QueryRegressionTest$lambda$$textContents$1.apply(QueryRegressionTest.scala:262) at quasar.regression.QueryRegressionTest$lambda$$textContents$1.apply(QueryRegressionTest.scala:262) Printing out the file path at QueryRegressionTest.scala:262 showed FileIn(DirIn(DirIn(DirIn(DirIn(DirIn(Current,DirName(it)),DirName(src)),DirName(main)),DirName(resources)),DirName(tests)),FileName(.caseWithGroup.test.swp)) which showed the problem - the regression file regex was matching on files beginning with ., and then trying to load that .swp file as the regression test file. This PR fixes that. Travis failed with [error] x using disposition to download as zipped directory [error] Gave up after only 0 passed tests. 2 tests were discarded. (DataServiceSpec.scala:59) which should be entirely unrelated. Running travis again. Hm, this time the build appears to have hung at or near the end of the regression tests. No output has been received in the last 10 minutes, this potentially indicates a stalled build or something wrong with the build itself. The build has been terminated I shall run it again. Is this something that's been seen before? 👍🏾
gharchive/pull-request
2016-04-20T10:59:35
2025-04-01T06:45:33.378469
{ "authors": [ "alissapajer", "sellout" ], "repo": "quasar-analytics/quasar", "url": "https://github.com/quasar-analytics/quasar/pull/1171", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
502353013
About preprocessing Hi, This repo is fantastic! I have questions about the preprocessing part. I only saw 'return x' in the preprocess_input in resnet.py. Is that true for all the resnet pretrained models? Also, for the input tensor, does it expect 'RGB' or 'BGR'? Thanks! Joey Hi, thanks, resnet models have been trained on RGB images with range 0-255 (uint8) Just read images in RGB uint8 and use preprcessing_function -> dont think about it )) Hi, Thanks for quick response!
gharchive/issue
2019-10-03T23:48:57
2025-04-01T06:45:33.436077
{ "authors": [ "JoeyTPChou", "qubvel" ], "repo": "qubvel/classification_models", "url": "https://github.com/qubvel/classification_models/issues/37", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
432577784
load_model issue Hi, First of all thanks for the great work. I trained a FPN model, with Resnext50 as backbone, for multiclass image segmentation. I saved the model using model.save() but when I try to retrieve in order to evaluate the performance on a new test dataset, I get a ValueError: Unknown layer: Slice. Has anyone else faced a similar problem? Ok found it. In order to load the particular model, I did the following and it worked. from segmentation_models.common.layers import ResizeImage from classification_models.common.blocks import Slice model = load_model(FILENAME_MODEL, custom_objects={'Slice': Slice, 'ResizeImage': ResizeImage}) You should be able to load model just after this line of code: import segmentation_models Layers will be registered as custom objects in global keras registry. Thanks a lot @qubvel , I was having the same issue and importing the library fixed it. Would be great if you could add this to the README so that in future people won't have to search in the issues to find this solution.
gharchive/issue
2019-04-12T13:52:45
2025-04-01T06:45:33.438674
{ "authors": [ "panakouris", "qubvel", "sarahESL" ], "repo": "qubvel/segmentation_models", "url": "https://github.com/qubvel/segmentation_models/issues/86", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1130105400
fix: social image url using gatsby image Did the minimal changes to get this to work, hope it's helpful. Hahaha its okay, thank you for the help
gharchive/pull-request
2022-02-10T13:44:01
2025-04-01T06:45:33.439576
{ "authors": [ "gtchakama", "raae" ], "repo": "queen-raae/gtchakama-myblog", "url": "https://github.com/queen-raae/gtchakama-myblog/pull/1", "license": "0BSD", "license_type": "permissive", "license_source": "github-api" }
956144014
Prefix all public facing ports in Chorus with something like 10, so 108983, 103000, 108080 etc Often I'll fire up Chorus and discover that port 3000 was already taken by my Quepid. Or a Solr is running on 8983 already... can we prefix everything with a number like 10, and then keep them on the well known port number. Ugh, why is the limit 65535? 8983 --> 108983 won't work. So, how about prefixing with a 1... So 8983 --> 18983? 3000 becomes 13000? Another option here is to not have so many public facing ports exposed or have something like Traefik or Nginx front these containers to expose with hostnames (this gets ugly since you need /etc/hosts or something to make it truely work). The first option would automatically pick an open port and then you could ask docker-compose what port that was opened on. This avoids a whole bunch of hardcoding. Its harder to use though from a "learning" standpoint. docker-compose port solr1 8983 https://docs.docker.com/compose/reference/port/ The other example is https://doc.traefik.io/traefik/user-guides/docker-compose/basic-example/, but requires domain name :( sigh, the darn port 7000 getting used by Apple is an example of why this might have been nice.
gharchive/issue
2021-07-29T19:30:07
2025-04-01T06:45:33.446050
{ "authors": [ "epugh", "risdenk" ], "repo": "querqy/chorus", "url": "https://github.com/querqy/chorus/issues/59", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2171396737
Add OVER(PARTITION BY ...) clause Why the new feature should be added Currently for example SQLExpressions.percentileCont(...) only supports .withinGroup().orderBy(...), but not .over().partitionBy(..) How the new feature should work It would be awesome to be able to write SQLExpressions.percentileCont(0.25).withinGroup().orderBy("value").over().partitionBy("series") to get PERCENTILE_CONT(0.25) WITHIN GROUP (ORDER BY value) OVER (PARTITION BY series) As there is already support for this with SQLExpressions.count() (e.g. SQLExpressions.count().over().partitionBy((Expression<?>) null)) you could try to do something like this to achieve your percentile count with OVER () new WindowFunction<>(SQLExpressions.percentileCont((Expression<Number>) null)) .partitionBy((Expression<?>) null); The nulls of course have to be replaced with your proper Expressions. :) As an exlanation: SQLExpressions.count() returns a WindowOver instance providing the .over() which then returns a WindowFunction that brings the partitionBy() method. This is not tested and only assumed to work. But the code gives me good hope, that it should. Thanks for the reply @dodgex I will give it a try and report back
gharchive/issue
2024-03-06T12:30:44
2025-04-01T06:45:33.451040
{ "authors": [ "dodgex", "enwi" ], "repo": "querydsl/querydsl", "url": "https://github.com/querydsl/querydsl/issues/3698", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
99355608
All Library Images are shown blurred When we open the QBImagePicker to select multiple images, All the images have blurred thumbnails. QBAssetsViewController.m ---> line 451 ---> When first run self.traitCollection no value @wangbiye When i add this lines i solve the problem ?.. because i to have the same problem.... Can you help me..... i am doing in swift... @sreeji44 I just change a place, which is the relevant code changes after me, but it is Objective-C. - (UICollectionViewCell *)collectionView:(UICollectionView *)collectionView cellForItemAtIndexPath:(NSIndexPath *)indexPath { QBAssetCell *cell = [collectionView dequeueReusableCellWithReuseIdentifier:@"AssetCell" forIndexPath:indexPath]; cell.tag = indexPath.item; cell.showsOverlayViewWhenSelected = self.imagePickerController.allowsMultipleSelection; // Image PHAsset *asset = self.fetchResult[indexPath.item]; CGSize itemSize = [(UICollectionViewFlowLayout *)collectionView.collectionViewLayout itemSize]; /* change here */ UIWindow *window = [UIApplication sharedApplication].keyWindow; CGFloat fixValue; if (self.traitCollection.displayScale != 0) { fixValue = self.traitCollection.displayScale; } else { fixValue = window.traitCollection.displayScale; } CGSize targetSize = CGSizeScale(itemSize, fixValue); [self.imageManager requestImageForAsset:asset targetSize:targetSize contentMode:PHImageContentModeAspectFill options:nil resultHandler:^(UIImage *result, NSDictionary *info) { if (cell.tag == indexPath.item) { cell.imageView.image = result; } }]; // Video indicator if (asset.mediaType == PHAssetMediaTypeVideo) { cell.videoIndicatorView.hidden = NO; NSInteger minutes = (NSInteger)(asset.duration / 60.0); NSInteger seconds = (NSInteger)ceil(asset.duration - 60.0 * (double)minutes); cell.videoIndicatorView.timeLabel.text = [NSString stringWithFormat:@"%02ld:%02ld", (long)minutes, (long)seconds]; if (asset.mediaSubtypes & PHAssetMediaSubtypeVideoHighFrameRate) { cell.videoIndicatorView.videoIcon.hidden = YES; cell.videoIndicatorView.slomoIcon.hidden = NO; } else { cell.videoIndicatorView.videoIcon.hidden = NO; cell.videoIndicatorView.slomoIcon.hidden = YES; } } else { cell.videoIndicatorView.hidden = YES; } // Selection state if ([self.imagePickerController.selectedAssets containsObject:asset]) { [cell setSelected:YES]; [collectionView selectItemAtIndexPath:indexPath animated:NO scrollPosition:UICollectionViewScrollPositionNone]; } return cell; } @wangbiye actually i am using the objective c version and converted the code thats all... so QBassetviewcontroller.m page i want to change it right... ok thankz a lot..... @wangbiye actually i tried those lines for the first time it seems clear.. and when i add second selection it stilll look blur images... :( @wangbiye thankz its works I'm sorry, my English is bad, I cannot quickly understand what you mean. @wangbiye its ok.. Your code work for me. :) CGSize targetSize = CGSizeScale(itemSize, fixValue); This is showing error, It worked for me when I changed this: PHImageRequestOptions* options = [[PHImageRequestOptions alloc] init]; options.synchronous = YES; options.networkAccessAllowed = YES; options.deliveryMode = PHImageRequestOptionsDeliveryModeHighQualityFormat; It was options.deliveryMode = PHImageRequestOptionsDeliveryModeFastFormat before. It used to work even without this though. This might be super late but can help people in future. Adding a multiplier in QBAssetViewController is the solution that worked for me at line 450. CGSize targetSize = CGSizeScale(itemSize, self.traitCollection.displayScale*4); //self.traitCollection.displayScale = 2. This is a default set in iOS 13.2 (or somewhere around that) multiplying it by 4 increases the resolution and you see clearer images.
gharchive/issue
2015-08-06T05:06:54
2025-04-01T06:45:33.458593
{ "authors": [ "deepak802sharma", "jankarrajpara", "mrharsh2804", "sreeji44", "wangbiye" ], "repo": "questbeat/QBImagePicker", "url": "https://github.com/questbeat/QBImagePicker/issues/86", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2568283619
Device location verification API only checks longitude For the Device Location Verification API, It can verify when the device's latitude and longitude is matching the query. { "status":200, "method":"POST", "endpoint":"location-verification", "api":"verify", "accessToken":"Bearer xxxxx", "phoneNumber":"00000000", "message":"location-verification request successful", "datetime":"2024-10-05T19:13:36.170Z", "cell":{ "phoneNumber":"00000000", "name":"team x", "message":"Hi", "accessToken":"Bearer xxxxx" }, "device":{ "phoneNumber":"00000000" }, "area":{ "type":"Circle", "location":{ "latitude":49.261177, "longitude":-123.249161 }, "accuracy":12000 }, "verificationResult":true, "lastLocationTime":"2024-10-05T19:13:36.170Z" } And when I change the longitude something incorrect, it does return false for the verification result { "status":200, "method":"POST", "endpoint":"location-verification", "api":"verify", "accessToken":"Bearer xxxxx", "phoneNumber":"00000000", "message":"location-verification request successful", "datetime":"2024-10-05T19:13:36.170Z", "cell":{ "phoneNumber":"00000000", "name":"team x", "message":"Hi", "accessToken":"Bearer xxxxx" }, "device":{ "phoneNumber":"00000000" }, "area":{ "type":"Circle", "location":{ "latitude":49.261177, "longitude":123.249161 }, "accuracy":50 }, "verificationResult":false, "lastLocationTime":"2024-10-05T19:13:36.170Z" } However, when I query it with the wrong latitude value it still returns true. { "status":200, "method":"POST", "endpoint":"location-verification", "api":"verify", "accessToken":"Bearer xxxxx", "phoneNumber":"00000000", "message":"location-verification request successful", "datetime":"2024-10-05T19:13:36.170Z", "cell":{ "phoneNumber":"00000000", "name":"team x", "message":"Hi", "accessToken":"Bearer xxxxx" }, "device":{ "phoneNumber":"00000000" }, "area":{ "type":"Circle", "location":{ "latitude":1, "longitude":-123.249161 }, "accuracy":50 }, "verificationResult":true, "lastLocationTime":"2024-10-05T19:13:36.170Z" } So i think that it is currently only checking the longitude of the coordinates of the device for if it is in the correct location as it also does error if both longitude and latitude are incorrect { "status":200, "method":"POST", "endpoint":"location-verification", "api":"verify", "accessToken":"Bearer xxxxx", "phoneNumber":"00000000", "message":"location-verification request successful", "datetime":"2024-10-05T19:24:52.281Z", "cell":{ "phoneNumber":"00000000", "name":"team x", "message":"Hi", "accessToken":"Bearer xxxxxx" }, "device":{ "phoneNumber":"0000000" }, "area":{ "type":"Circle", "location":{ "latitude":50.735851, "longitude":7.10066 }, "accuracy":50 }, "verificationResult":false, "lastLocationTime":"2024-10-05T19:24:52.281Z" } Try reducing your accuracy to the scale of 0.1 - 0.01. Your accuracy value is to large and I believe that is what is causing the problem. I've tested this with lower accuracy values and it correctly reports for different latitudes. Looks liked the accuracy is in degrees not meters
gharchive/issue
2024-10-05T19:26:31
2025-04-01T06:45:33.464527
{ "authors": [ "korih", "tolbooth" ], "repo": "questsin/UBC-Rogers-CAMARA-Hackathon", "url": "https://github.com/questsin/UBC-Rogers-CAMARA-Hackathon/issues/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1554162597
return StreamErrors for all kinds of stream cancelations Apparently @Jorropo didn't allow me to push to his branch in https://github.com/quic-go/quic-go/pull/3674, so I'm recreating this PR. @Jorropo Would you mind giving this PR a quick review? Codecov Report Merging #3681 (69c7e52) into master (b52d340) will increase coverage by 1.57%. The diff coverage is 55.56%. @@ Coverage Diff @@ ## master #3681 +/- ## ========================================== + Coverage 85.36% 86.93% +1.57% ========================================== Files 140 142 +2 Lines 10246 12085 +1839 ========================================== + Hits 8746 10505 +1759 - Misses 1111 1156 +45 - Partials 389 424 +35 Impacted Files Coverage Δ errors.go 33.33% <0.00%> (-33.33%) :arrow_down: receive_stream.go 97.67% <100.00%> (+1.65%) :arrow_up: send_stream.go 94.37% <100.00%> (+1.06%) :arrow_up: stream.go 84.00% <0.00%> (-2.49%) :arrow_down: config.go 100.00% <0.00%> (ø) logging/packet_header.go 100.00% <0.00%> (ø) internal/ackhandler/ackhandler.go 0.00% <0.00%> (ø) internal/ackhandler/frame.go 44.44% <0.00%> (ø) internal/qtls/go120.go 58.33% <0.00%> (ø) internal/ackhandler/received_packet_history.go 96.51% <0.00%> (+0.08%) :arrow_up: ... and 29 more Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
gharchive/pull-request
2023-01-24T02:00:24
2025-04-01T06:45:33.479016
{ "authors": [ "codecov-commenter", "marten-seemann" ], "repo": "quic-go/quic-go", "url": "https://github.com/quic-go/quic-go/pull/3681", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
339235238
Error when printing table for experiments An error occurs when printing an experiment for an app: $ tune ls 8a77a2d6-326f-463e-baf8-ddcef1262016 ... TypeError: Cannot set property '0' of undefined at chunkedValue.forEach (/usr/local/lib/node_modules/autotune/node_modules/table/dist/mapDataUsingRowHeightIndex.js:49:35) at Array.forEach (<anonymous>) at cells.forEach (/usr/local/lib/node_modules/autotune/node_modules/table/dist/mapDataUsingRowHeightIndex.js:48:20) at Array.forEach (<anonymous>) at unmappedRows.map (/usr/local/lib/node_modules/autotune/node_modules/table/dist/mapDataUsingRowHeightIndex.js:39:11) at Array.map (<anonymous>) at exports.default (/usr/local/lib/node_modules/autotune/node_modules/table/dist/mapDataUsingRowHeightIndex.js:30:35) at Object.exports.default [as table] (/usr/local/lib/node_modules/autotune/node_modules/table/dist/table.js:126:51) at logTable (/usr/local/lib/node_modules/autotune/dist/cli/index.js:508:25) at /usr/local/lib/node_modules/autotune/dist/cli/index.js:372:25 That's due to this bug: https://github.com/chalk/slice-ansi/pull/19 I'll do a workaround because I don't want to wait until upstream is updated. The workaround is to set truncate in the column configuration, which gives us something like this:
gharchive/issue
2018-07-08T15:37:16
2025-04-01T06:45:33.487482
{ "authors": [ "dvdsgl", "schani" ], "repo": "quicktype/autotune", "url": "https://github.com/quicktype/autotune/issues/51", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
270852159
IR enhancements for GraphQL As far as I can tell, to express the whole of GraphQL we need to enhancements to the IR: [x] Enums. Just an ordered set of names. [ ] Discriminated unions. A discriminated union would be an enum with a class type associated to each enum value. Note that GraphQL does not implicitly return an enum value that determines the class type, but it can be forced to by including __typename in the query. [ ] Interfaces. An interface is like a class, except that it's never actually instantiated as an object, and other classes can implement it. We might choose to implement interfaces by making classes subclassable. We could just have a flag on a class that says whether it's an interface or not. For an MVP we can probably do without interfaces. Note that I'm not proposing also inferring those types from input JSON (yet). For the purposes of GraphQL support we only need to have the IR support them and the renderers produce code for them. Here's an example on what we need discriminated unions for: query FragmentTyping { profiles(handles: ["zuck", "cocacola"]) { handle ...userFragment ...pageFragment } } fragment userFragment on User { friends { count } } fragment pageFragment on Page { likers { count } } In GraphQL parlance, the type of profiles could be a union of User, Page, and potentially other types, or it could be an interface which both User and Page implement (there are other possibilities, too). The expected result is that the generated type for profiles would contain friends or likers, or both, or neither, depending on what the dynamic type of the result is. The cheap way, not requiring any new support from the IR, would be to generate a type that contains both, but make them optional. I'll implement this first - it's trivial. A much nicer way is to actually generate separate types for the two, three, or four cases (see below for an explanation of the number of cases). What these types look like in generated code can be very different in different target languages. In TypeScript, for example, we could just spell out all the full types and make the profiles type a union of them. That has the disadvantage of leaving the user with no nice way of distinguishing between them - they would have to look at the __typename field, or check for which properties are non-null. Another way would be to generate actual classes, and maybe put the common fields in a shared interface. In that case the user could do instanceof to get the type, but we would have to rewrite the JSON object to instantiate those classes. In C++ we could generate separate classes with an optional common superclass, too. Here's how there might be two, three, or four different cases. Note that there might be other ways to get to those specific sets of cases, too: Two cases, either friends or likers: profiles is a union of User and Page, which are both object types. Three cases, either friends or likers or neither: profiles is a union of User, Page, and Event, all of which are object types. Three cases, either friends or likers or both: profiles is a union of User and Page, which are both interfaces, and there is a PlainUser which implements only User, a PlainPage which implements only Page, and a UserPage which implements both. Four cases: Like the latter, except that the profiles union also contains the object type Event. It's now also clear that there might be more than one dynamic type that corresponds to a static case. For example, if User is an interface then any type that implements it (and doesn't implement Page) will result in the same case. It seems to me that the minimum discriminated union type we need to nicely represent this in the IR is something like this: class DiscriminatedUnionType { discriminatingProperty: string; cases: Set<[Set<string>, Map<string, Type>]>; } discriminatingProperty is the JSON property that determines which case applies. In GraphQL it is always __typename. cases is the set of all static cases. Each one is a tuple consisting of a set of type names that map to the particular case, and a property map, like we have in ClassType. The type name is the dynamic value of discriminatingProperty. We might want to use a proper class instead of an array to represent that tuple. It's not clear to me know that we really do need interfaces in the IR, but it might still be helpful. As explained in the above comment, a good way to represent discriminated unions in many programming languages seems to be as a bunch of classes that implement a common interface which covers the properties common to all result cases. To do that, we need to have an analysis step that finds all the common properties. If we had interfaces in the IR we could implement that analysis as an IR transformation step that's only done for languages that need/want it. That would give us the ability to use interfaces for other things further down the road, like inferring common supertypes from JSON, and the backends would already support it. A discriminated union based on an interface would look different than the one described above, however.
gharchive/issue
2017-11-03T00:49:03
2025-04-01T06:45:33.498628
{ "authors": [ "schani" ], "repo": "quicktype/quicktype", "url": "https://github.com/quicktype/quicktype/issues/257", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1411446690
Improve protect_future Add a label Add a timeout Display a warning when the future takes too long to complete Call record_progress after the future completes Call record_progress after the future completes
gharchive/issue
2022-10-17T11:59:17
2025-04-01T06:45:33.500756
{ "authors": [ "fulmicoton", "guilload" ], "repo": "quickwit-oss/quickwit", "url": "https://github.com/quickwit-oss/quickwit/issues/2134", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1774168189
Merge pipeline failed when index has too many splits Describe the bug quickwit-indexer got this error 2023-06-26T07:47:49.057Z WARN spawn_merge_pipeline{index=xxx-index gen=0}: quickwit_metastore::metastore::retrying_metastore::retry: Request failed attempt_count=5 2023-06-26T07:47:49.057Z ERROR quickwit_indexing::actors::merge_pipeline: Error while spawning indexing pipeline, retrying after some time. error=Internal error: `Error, message length too large: found 50114113 bytes, the limit is: 4194304 bytes` Cause: ``. retry_count=0 retry_delay=2s Expected behavior no error Configuration: quickwit version: v0.6.1 deploy with helm, default settings (with cpu/memory request/limit increased) I found the same error(message length too large) when request /api/v1/indexes/<index-id>/splits on searcher, but successed on metastore. I found this error template in tonic's source, so I guess that error occurs on large grpc response of splits. MergePipeline query all published splits merge_pipeline.rs let query = ListSplitsQuery::for_index(self.params.pipeline_id.index_uid.clone()) .with_split_state(SplitState::Published); It seems that the mergepipline must be failed in the future because of lots of splits. Do we need a streaming splits api or just filter splits with start_timestamp (via maturation_period)? Ah interesting. Out of curiosity, how many splits do you have? Yes, our query that fetches all published splits is a bit stupid. We could fetch only immature splits, we would need to filter on split.create_timestamp. In the merge pipeline, one of the condition that filters out splits that won't undergo merges is like this: if OffsetDateTime::now_utc().unix_timestamp() >= split.create_timestamp + self.config.maturation_period.as_secs() as i64 { return true; } Ah interesting. Out of curiosity, how many splits do you have, and how did you end up in that state? 8w+ splits, this cluster is crashed last night (k8s OOMKilled), 6w+ splits is created last night (oom then restart...then oom) just to be sure, 8w is for 80 000 right? just to be sure, 8w is for 80 000 right? right I have also encountered this situation. The current temporary solution is to increase the limit condition of the query to control the number of pulls at one time. The splits that is out of the limits will never be merged at all... @fulmicoton we first fetch all splits and then filter out the mature ones. @geek-frio @dojiong we will fix this as soon as possible. Another question: to have that many splits, you are probably using partitioning? Do you have long retention too? @dojiong the fix is released in 0.6.2. It does not solve all the cases, you can still get the same error messages if you search on 10k splits for example. We will push a fix to use gRPC stream and avoid sending big messages. @geek-frio, it should work for you as well. And sorry again for the bad experience, don't hesitate to report other bugs, it's really helping us. @geek-frio, it should work for you as well. And sorry again for the bad experience, don't hesitate to report other bugs, it's really helping us. One more thing @dojiong @geek-frio : the upgrade will trigger a postgresql migration. It’s better to make a dump of the db before. Also I should have done a proper upgrade test, I will check today the upgrade process, it should work but I want to double check that. Ok, PostgreSQL migration worked fine on our cluster. It will add the column maturity_timestamp to the splits table. By default, the maturity_timestamp is set to 1970-01-01 00:00:00. This means that all your splits will be considered mature and won't undergo merge operations. This is mostly true for your splits but... it may be not for splits created recently. You may want to do an additional operation, like set the maturity_timestamp of recent splits that has a low number of documents to a timestamp in the future, like now + 48h. @fmassot I tested on my cluster, it works as expected!
gharchive/issue
2023-06-26T07:57:42
2025-04-01T06:45:33.511641
{ "authors": [ "dojiong", "fmassot", "fulmicoton", "geek-frio" ], "repo": "quickwit-oss/quickwit", "url": "https://github.com/quickwit-oss/quickwit/issues/3576", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
79887937
Proposal: Allow returning simple code results I'd love to randomize the results - no different than allowing simple code execution. In this case, it might look like the following --- - name: Test endpoint request: uri: /hello method: GET response: code: 200 body: '{"message": "hello, world", "temperatureOutside": `fmt.Sprintf("%s", rand.Intn(100)))`}' headers: content-type: text/json There is a much much much less ugly way to do it than the above, but you get the idea. Hi Thanks for the suggestion. A few things to discuss: Could you perhaps elaborate on the use case? Adding code execution certainly adds some complexity and I do want to strive to keep this as simple as possible. I've lived in worlds where fakes turn out to be quite complicated, which lead to complicated tests. That being said I dont want to throw out every idea in the name of simplicity! In this particular case, how random do you need it to be? You can achieve some randomness with the monkey config: --- - body: '{"message": "hello, world", "temperatureOutside": 30}' frequency: 0.3 - body: '{"message": "hello, world", "temperatureOutside": 10}' frequency: 0.3 - body: '{"message": "hello, world", "temperatureOutside": 50}' frequency: 0.3 ````yaml Would this suffice? Thanks again This is really close, but I guess I was hoping for true randomness - basically just returning what would be issued by the random command in Go, to make the data look a bit more real. But, you're right, it makes pure testing non-deterministic, which is, itself, an issue. On Mon, May 25, 2015 at 1:21 AM, Chris James notifications@github.com wrote: Hi Thanks for the suggestion. A few things to discuss: Could you perhaps elaborate on the use case? Adding code execution certainly adds some complexity and I do want to strive to keep this as simple as possible. I've lived in worlds where fakes turn out to be quite complicated, which lead to complicated tests. That being said I dont want to throw out every idea in the name of simplicity! In this particular case, how random do you need it to be? You can achieve some randomness with the monkey config: ---- body: '{"message": "hello, world", "temperatureOutside": 30}' frequency: 0.3- body: '{"message": "hello, world", "temperatureOutside": 10}' frequency: 0.3- body: '{"message": "hello, world", "temperatureOutside": 50}' frequency: 0.3````yaml Would this suffice? Thanks again — Reply to this email directly or view it on GitHub https://github.com/quii/mockingjay-server/issues/13#issuecomment-105163087 . Closing for now, perhaps if a stronger use case comes up we can look into it.
gharchive/issue
2015-05-23T16:18:23
2025-04-01T06:45:33.526695
{ "authors": [ "aronchick", "quii" ], "repo": "quii/mockingjay-server", "url": "https://github.com/quii/mockingjay-server/issues/13", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
127657095
Adding a number_of_units attribute to converters Background We have some methods in ET-Engine to calculate the number_of_units attribute based on their capacity and one specifically for (space/water heating technologies) demand_driven converters. There are, however, plenty of converters where a number_of_units attributes is wanted, but where these methods return 0. Especially when it comes to scalability and costs, this can even be necessary for other methods to return correct values. For instance in https://github.com/quintel/etsource/issues/800 , we need a number_of_units attribute for two car technologies to correctly calculate extra investments costs (otherwise these costs will just be multiplied by 0, the number_of_units returned as of now). First attempt I simply added a number_of_units attribute to transport_car_using_hydrogenin ETSource, added this attribute to the list of numeric attributes in note.rb in Atlas, set my Gemlock to my local Atlas in ETEngine and put V(transport_car_using_hydrogen,number_of_units) in the GQL sandbox, which worked and gave me the value I had initially put. Problem However, when I created a new scenario, I got a lot of errors of 'nil cannot be coerced into float' and my dashboard wouldn't appear. Debugging the total cost calculation led me to conversion.rb : when :converter cost * number_of_units I figured this is because ETEngine, now that number_of_units is a converter attribute, looks for it in every converter and returns nil if it is not included in the list of converter attributes. This belief was strengthened by the fact that the error arose in agriculture_burner_crude_oil, the first converter given our folder structure. Possible solution The best possible possible solution that came to my mind is the following (however, I don't know how hard it is to implement): make number_of_units an 'optional' attribute, such that ETEngine picks it up if it is included in the particular converters attributes, but 'skips' the converter (and doesn't return number_of_units = nil) if it is not there and then executes other number_of_units calculations mentioned at the start of this ticket. This solution has the advantage it doesn't require the development of new number_of_units calculation methods and that other converters which would benefit from or need a number_of_units attributes could easily get one as well. @antw , what are your thoughts, do you think this is easily implemented, do you foresee any trouble, or would you perhaps know a better solution (one that allows us to either calculate or define the number_of_units for converters without capacities)? @RobTerwel Would you mind pushing your ETSource and Atlas changes to branches on GitHub? I'll take a look and see what I can do. Would you mind pushing your ETSource and Atlas changes to branches on GitHub? Sure, ETS is h2_number_scaling and Atlas is new_attributes. Thanks! Great work, @antw ! Thanks for making this possible, also in such a small timespan! The best possible possible solution that came to my mind is the following: make number_of_units an 'optional' attribute This has been done in 13dd4ce5, and all the queries are executing correctly for me. I figured this is because ETEngine, now that number_of_units is a converter attribute, looks for it in every converter and returns nil if it is not included in the list of converter attributes. This was pretty much correct. Most converters would get a number_of_units set to nil, since they had no value defined in ETSource. I've set it so that any converter with no NoU will compute the value as before. :+1:
gharchive/issue
2016-01-20T10:55:49
2025-04-01T06:45:33.546066
{ "authors": [ "RobTerwel", "antw" ], "repo": "quintel/etengine", "url": "https://github.com/quintel/etengine/issues/813", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1713536559
🛑 Adguard Home DoT is down In 15f1faa, Adguard Home DoT ($AG_DOT) was down: HTTP code: 0 Response time: 0 ms Resolved: Adguard Home DoT is back up in 74a5ef7.
gharchive/issue
2023-05-17T09:52:04
2025-04-01T06:45:33.604548
{ "authors": [ "quyleanh" ], "repo": "quyleanh/upptime", "url": "https://github.com/quyleanh/upptime/issues/1216", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1771610380
🛑 Adguard Home DoT is down In 88614a3, Adguard Home DoT ($AG_DOT) was down: HTTP code: 0 Response time: 0 ms Resolved: Adguard Home DoT is back up in c600e1d.
gharchive/issue
2023-06-23T14:59:47
2025-04-01T06:45:33.606804
{ "authors": [ "quyleanh" ], "repo": "quyleanh/upptime", "url": "https://github.com/quyleanh/upptime/issues/2422", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1849032755
🛑 Adguard Home DoT is down In f449a96, Adguard Home DoT ($AG_DOT) was down: HTTP code: 0 Response time: 0 ms Resolved: Adguard Home DoT is back up in a8c090f.
gharchive/issue
2023-08-14T05:33:47
2025-04-01T06:45:33.608932
{ "authors": [ "quyleanh" ], "repo": "quyleanh/upptime", "url": "https://github.com/quyleanh/upptime/issues/4054", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1993415312
🛑 Adguard Home DoT is down In 0fea8f1, Adguard Home DoT ($AG_DOT) was down: HTTP code: 0 Response time: 0 ms Resolved: Adguard Home DoT is back up in 311fe2c after 31 minutes.
gharchive/issue
2023-11-14T19:23:51
2025-04-01T06:45:33.611080
{ "authors": [ "quyleanh" ], "repo": "quyleanh/upptime", "url": "https://github.com/quyleanh/upptime/issues/7208", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2022284008
🛑 Adguard Home DoT is down In 220c7c1, Adguard Home DoT ($AG_DOT) was down: HTTP code: 0 Response time: 0 ms Resolved: Adguard Home DoT is back up in f5c8511 after 55 minutes.
gharchive/issue
2023-12-03T01:40:14
2025-04-01T06:45:33.613429
{ "authors": [ "quyleanh" ], "repo": "quyleanh/upptime", "url": "https://github.com/quyleanh/upptime/issues/7803", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2092181910
🛑 Adguard Home DoT is down In da91368, Adguard Home DoT ($AG_DOT) was down: HTTP code: 0 Response time: 0 ms Resolved: Adguard Home DoT is back up in 9a796d1 after 19 minutes.
gharchive/issue
2024-01-20T17:58:21
2025-04-01T06:45:33.615512
{ "authors": [ "quyleanh" ], "repo": "quyleanh/upptime", "url": "https://github.com/quyleanh/upptime/issues/8933", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2298796250
🏃‍♂️ Sprint 15/05/24 Guidelines for Beta [ ] Getting to a consistent API between all components [ ] Dogfooding each component in production [ ] 100% coverage on all headless components (and a11y validated) [ ] Logical test coverage on styled components which are not covered by headless [ ] Visual tests for styled components [ ] Build is optimized for tree shaking (or fix namespace issues) [ ] Make docs fully optimized (shiki bug fix) Actual Tasks Must haves [ ] Fix CLI on new projects - Shai & Maieul [ ] Finish Popover tests (polyfill & critical tests) - Nate & Chris [ ] Add docker for visual testing with playwright - Shai & Nate [x] Rename ALL the styled components to dot notation - Maieul [x] #726 - Maieul [x] Complete refactor of Tooltip - Jack & Shai & Chris [x] Refactor Accordion to remove useVisibleTask$ - Jack [ ] Finish headless Checkbox - Nate [ ] Add the SSG adapter - Maieul & Shai Edit (Maïeul): [x] update package.json deps (probably "@k11r/nx-cloudflare-wrangler") for node v21.7 compatibility and above: https://github.com/cloudflare/workers-sdk/pull/5201 Nice to haves [ ] Review styled components - Maieul & Shai [x] Styled Select component (Jack) [ ] Complete sonner component (Victor) [ ] Sort out the github project [x] Change Modal API [ ] Write tests for Combobox (look at the select ones) [ ] Add the ability to add custom elements / components inside of select options [x] Fix the "preserve modules" issue [ ] Open an issue about PW component tests on PW github (Shai) [ ] Make cli E2E test working (Shai) [ ] Update the CLI to handle animation "apply" per component (Shai) [ ] Test cli in nx projects (and sub folders) (Shai) closing in favor of the GH project - https://github.com/orgs/qwikifiers/projects/1/views/1
gharchive/issue
2024-05-15T20:36:33
2025-04-01T06:45:33.642344
{ "authors": [ "shairez" ], "repo": "qwikifiers/qwik-ui", "url": "https://github.com/qwikifiers/qwik-ui/issues/740", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2074940809
1.3.2 cleanup What is it? [ ] Feature / enhancement [ ] Bug [ ] Docs / tests Description This cleans up most of the unused code. Removes @qwik-ui/tailwind Removes @qwik-ui/material Removes @qwik-ui/primitives Removes unused components Changes QwikIntrinsicElements to the new PropsOf Add Fluffy badge component and changes docs badges accordingly Removes useVisibleTask$ eslint warnings Screenshots/Demo Checklist: [x] My code follows the developer guidelines of this project [x] I have performed a self-review of my own code [x] I have made corresponding changes to the documentation [ ] Added new tests to cover the fix / functionality Looks goood!
gharchive/pull-request
2024-01-10T18:27:29
2025-04-01T06:45:33.646675
{ "authors": [ "maiieul", "thejackshelton" ], "repo": "qwikifiers/qwik-ui", "url": "https://github.com/qwikifiers/qwik-ui/pull/590", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1980440291
Data from the Pile [ ] Pile of Law [ ] HackerNews? Should compare pile of law's overlap to the other sources. Closing as duplicative of other issues.
gharchive/issue
2023-11-07T02:53:54
2025-04-01T06:45:33.749502
{ "authors": [ "StellaAthena", "blester125", "craffel" ], "repo": "r-three/licensed-pile", "url": "https://github.com/r-three/licensed-pile/issues/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1300646921
Can't run 11 billion model on A100 with 80GB Hi @craffel @muqeeth @HaokunLiu, We're trying to reproduce T-Few results for a paper, but we're getting 'CUDA out of memory' using an A100 with 80GB (your recommended setup). This is what we're running: python -m src.pl_train -c t011b.json+ia3.json+rte.json -k load_weight="pretrained_checkpoints/t011b_ia3_finish.pt" exp_name=t011b_rte_seed42_ia3_pretrained few_shot_random_seed=42 seed=42 We installed according to the README instructions and are using the default settings in the config files. We are able to run the 3 billion model using the command above, just not the 11 billion. Is there anything we are doing wrong? This is the exception: Thank you Thanks for your interest in our work! It's hard to tell from the surface. Could you share with me the full log? And if you are familiar with pytorch lightning, mind if add something like print("Memory usage at line [add something here]", torch.cuda.memory_allocated(device=None)) in functions in the start and end of training_step of EncoderDecoder.py? Hi @HaokunLiu We added the prints and attached the logs here. Looks like it runs out of memory before starting the training. (tfew3.7) unso@hf-paris-dgx-station-1:~/t-few$ python -m src.pl_train -c t011b.json+ia3.json+rte.json -k load_weight="pretrained_checkpoints/t011b_ia3_finish.pt" exp_name=t011b_rte_seed42_ia3_pretrained few_shot_random_seed=42 seed=42 > logfile Using bfloat16 Automatic Mixed Precision (AMP) GPU available: True, used: True TPU available: False, using: 0 TPU cores IPU available: False, using: 0 IPUs Reusing dataset super_glue (/home/unso/.cache/huggingface/datasets/super_glue/rte/1.0.2/d040c658e2ddef6934fdd97deb45c777b6ff50c524781ea434e7219b56a428a7) LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0] Missing logger folder: /home/unso/t-few/exp_out/t011b_rte_seed42_ia3_pretrained/log | Name | Type | Params ----------------------------------------------------- 0 | model | T5ForConditionalGeneration | 11.1 B ----------------------------------------------------- 1.1 M Trainable params 11.1 B Non-trainable params 11.1 B Total params 44,548.801Total estimated model params size (MB) Traceback (most recent call last): File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/runpy.py", line 85, in _run_code exec(code, run_globals) File "/home/unso/t-few/src/pl_train.py", line 86, in <module> main(config) File "/home/unso/t-few/src/pl_train.py", line 57, in main trainer.fit(model, datamodule) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 741, in fit self._fit_impl, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 685, in _call_and_handle_interrupt return trainer_fn(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 777, in _fit_impl self._run(model, ckpt_path=ckpt_path) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 1199, in _run self._dispatch() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 1279, in _dispatch self.training_type_plugin.start_training(self) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/plugins/training_type/training_type_plugin.py", line 202, in start_training self._results = trainer.run_stage() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 1289, in run_stage return self._run_train() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 1319, in _run_train self.fit_loop.run() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/base.py", line 145, in run self.advance(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/fit_loop.py", line 234, in advance self.epoch_loop.run(data_fetcher) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/base.py", line 145, in run self.advance(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/epoch/training_epoch_loop.py", line 193, in advance batch_output = self.batch_loop.run(batch, batch_idx) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/base.py", line 145, in run self.advance(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/batch/training_batch_loop.py", line 88, in advance outputs = self.optimizer_loop.run(split_batch, optimizers, batch_idx) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/base.py", line 145, in run self.advance(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/optimization/optimizer_loop.py", line 219, in advance self.optimizer_idx, File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/optimization/optimizer_loop.py", line 266, in _run_optimization self._optimizer_step(optimizer, opt_idx, batch_idx, closure) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/optimization/optimizer_loop.py", line 386, in _optimizer_step using_lbfgs=is_lbfgs, File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/core/lightning.py", line 1652, in optimizer_step optimizer.step(closure=optimizer_closure) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/core/optimizer.py", line 164, in step trainer.accelerator.optimizer_step(self._optimizer, self._optimizer_idx, closure, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/accelerators/accelerator.py", line 336, in optimizer_step self.precision_plugin.optimizer_step(model, optimizer, opt_idx, closure, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/plugins/precision/native_amp.py", line 80, in optimizer_step return super().optimizer_step(model, optimizer, optimizer_idx, closure, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/plugins/precision/precision_plugin.py", line 163, in optimizer_step optimizer.step(closure=closure, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/optim/lr_scheduler.py", line 65, in wrapper return wrapped(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/optim/optimizer.py", line 109, in wrapper return func(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/transformers/optimization.py", line 528, in step loss = closure() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/plugins/precision/precision_plugin.py", line 148, in _wrap_closure closure_result = closure() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/optimization/optimizer_loop.py", line 160, in __call__ self._result = self.closure(*args, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/optimization/optimizer_loop.py", line 142, in closure step_output = self._step_fn() File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/loops/optimization/optimizer_loop.py", line 435, in _training_step training_step_output = self.trainer.accelerator.training_step(step_kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/accelerators/accelerator.py", line 216, in training_step return self.training_type_plugin.training_step(*step_kwargs.values()) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/pytorch_lightning/plugins/training_type/training_type_plugin.py", line 213, in training_step return self.model.training_step(*args, **kwargs) File "/home/unso/t-few/src/models/EncoderDecoder.py", line 62, in training_step decoder_attention_mask=decoder_attention_mask, File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/transformers/models/t5/modeling_t5.py", line 1623, in forward return_dict=return_dict, File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/transformers/models/t5/modeling_t5.py", line 1020, in forward output_attentions=output_attentions, File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/transformers/models/t5/modeling_t5.py", line 696, in forward hidden_states = self.layer[-1](hidden_states) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/transformers/models/t5/modeling_t5.py", line 306, in forward forwarded_states = self.DenseReluDense(forwarded_states) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/transformers/models/t5/modeling_t5.py", line 285, in forward hidden_states = self.wo(hidden_states) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/unso/miniconda3/envs/tfew3.7/lib/python3.7/site-packages/torch/nn/modules/linear.py", line 114, in forward return F.linear(input, self.weight, self.bias) RuntimeError: CUDA out of memory. Tried to allocate 80.00 MiB (GPU 0; 79.35 GiB total capacity; 78.13 GiB already allocated; 3.62 MiB free; 78.20 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF I remember the code will print out all the args in the beginning. Could you share that with me? Sorry I think the config might be slightly off as it was meant for the 3B and not 11B versions. For the 11B variants, to fit into memory, we used a smaller batch size but still had an effect batch size of 8. Our hyperparameters werebatch_size=1 grad_accum_factor=8 eval_batch_size=2. Let us know if it still runs out of memory. thanks!
gharchive/issue
2022-07-11T12:33:04
2025-04-01T06:45:33.757619
{ "authors": [ "HaokunLiu", "danielkorat", "dptam", "eunseojo" ], "repo": "r-three/t-few", "url": "https://github.com/r-three/t-few/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
556847366
No Status effects Anyone else having no Status effects displayed, or is it just me? How do you make this work? Can you please share the memory pattern file?
gharchive/issue
2020-01-29T13:01:28
2025-04-01T06:45:33.795743
{ "authors": [ "Megagalatkikarea2", "YS-RF" ], "repo": "r00telement/SmartHunter", "url": "https://github.com/r00telement/SmartHunter/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
152190627
Previous activity doesn't show Java code: protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); setContentView(R.layout.activity_client_detail); Toolbar toolbar = (Toolbar) findViewById(R.id.toolbar); setSupportActionBar(toolbar); getSupportActionBar().setHomeButtonEnabled(true); getSupportActionBar().setDisplayHomeAsUpEnabled(true); SlidrConfig config = new SlidrConfig.Builder() .scrimEndAlpha(0f) .scrimStartAlpha(0.5f) .build(); Slidr.attach(this, config); } Theme: <style name="AppTheme.Slidr" parent="@style/AppTheme"> <item name="android:windowIsTranslucent">true</item> <item name="android:windowBackground">@android:color/transparent</item> </style> I got the same issue, but with transparent background, making the android home screen visible. Anyone? Have you add parent activity in manifest? yes I have @muhrifqii I have the same problem.But I just deal with it. The reason you could see android home screen is that the APP's background is transparent.so,you can set a app:theme on the HomeActivity in mainfest,and the style's attributes ,you can't set true @android:color/transparent.you must set your app's normally attributes,such as , false @android:color/red...and this could help you . OR:you can disallowed Slidr on your HomeActivity. its' not working in android 4.4.2 @haydarai you dont need to add parent activity in xml for this library to work :) @RomanTsekhmeistruk thanks for statusbar fix trick :)
gharchive/issue
2016-05-01T09:29:10
2025-04-01T06:45:33.801991
{ "authors": [ "PragneshNumberTank", "SSSSHIELD", "bhargavms", "haydarai", "lance2611", "muhrifqii" ], "repo": "r0adkll/Slidr", "url": "https://github.com/r0adkll/Slidr/issues/37", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
708864614
Web Interface Error Describe the bug When I try to run from the web interface it says 'NoneType' object is not subscriptable. To Reproduce Steps to reproduce the behavior: Access the Web interface Select all Containers, All Networks and All Tests See error Expected behavior Tests to run I guess and some sort of feedback provided Screenshots Desktop (please complete the following information): OS: Ubuntu Server 20.04 Browser: Brave and Firefox @The-Inamati This issue has been fixed in the latest REL, can you specify the version of DockerENT you are using. I cloned the repo the same day I posted. The current version is v0.1.15 Just run pip install DockerENT --upgrader once to check if everything is okay. I am running it, and it is running fine. On Mon, Sep 28, 2020 at 2:34 PM Inamati notifications@github.com wrote: I cloned the repo the same day I posted. The current version is v0.1.15 — You are receiving this because you were assigned. Reply to this email directly, view it on GitHub https://github.com/r0hi7/DockerENT/issues/35#issuecomment-699879743, or unsubscribe https://github.com/notifications/unsubscribe-auth/AFBMCRX5WC3J5G4OGGJVRPDSIBGQFANCNFSM4RZN2F3Q . Closing this issue, please open if you feel this issue still exit.
gharchive/issue
2020-09-25T11:30:14
2025-04-01T06:45:33.808893
{ "authors": [ "The-Inamati", "r0hi7" ], "repo": "r0hi7/DockerENT", "url": "https://github.com/r0hi7/DockerENT/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
961836108
🛑 Orchestra is down In 5bb378e, Orchestra (http://issue.chelsea.kt.co.kr) was down: HTTP code: 0 Response time: 0 ms Resolved: Orchestra is back up in 77e216b.
gharchive/issue
2021-08-05T13:36:28
2025-04-01T06:45:33.832240
{ "authors": [ "r2fresh" ], "repo": "r2fresh/chelsea", "url": "https://github.com/r2fresh/chelsea/issues/209", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126270444
Code I'm still waiting for the code! Soon. Enjoy the code. As you know, this is just a qnd one. Feel free to fork or rewrite ;) Closing.
gharchive/issue
2016-01-12T20:39:10
2025-04-01T06:45:33.833395
{ "authors": [ "f-breidenstein", "r3ek0" ], "repo": "r3ek0/infowall", "url": "https://github.com/r3ek0/infowall/issues/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1074154275
🛑 Home Assistant is down In 8a7c0d8, Home Assistant ($SECRET_SITE_HA) was down: HTTP code: 0 Response time: 0 ms Resolved: Home Assistant is back up in 6a0d95d.
gharchive/issue
2021-12-08T08:45:34
2025-04-01T06:45:33.835515
{ "authors": [ "r3f" ], "repo": "r3f/upptime", "url": "https://github.com/r3f/upptime/issues/1185", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
109534568
After a couple of playing SCPlayer does not play I played by fetching video from Documents folder. After a couple of playing, It does not play the video. What might be the problem? @erdemgc Could you please share your code here? For us, we can use the SCPlayer without problem. i am having exactly this problem http://raywenderlich.com/forums/viewtopic.php?f=2&t=6312
gharchive/issue
2015-10-02T17:23:28
2025-04-01T06:45:33.839839
{ "authors": [ "erdemgc", "ustbenchan" ], "repo": "rFlex/SCRecorder", "url": "https://github.com/rFlex/SCRecorder/issues/222", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2190173496
trying with 4 sensors i have my code done, from your file butt using platformio and 4 sensors but i can read the date from them, with android app its ok.. can you check the code? https://drive.google.com/drive/folders/1QMslgD5UtD_kUmY3Ydt3RCgHhAzKHknz?usp=sharing i recive data like this using the serial monitor code Address: 38:ab:00:00:f9:6b, manufacturer data: 281e17012c01e8, serviceUUID: 000027a5-0000-1000-8000-00805f9b34fb, rssi: -82
gharchive/issue
2024-03-16T18:22:27
2025-04-01T06:45:33.851665
{ "authors": [ "andrefcarvalho" ], "repo": "ra6070/BLE-TPMS", "url": "https://github.com/ra6070/BLE-TPMS/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
696664073
Add required OpenShift manual installation steps After https://github.com/rabbitmq/cluster-operator/pull/303 was merged, the steps in this PR are the only manual steps required for running the cluster operator on OpenShift. Thank you!
gharchive/pull-request
2020-09-09T09:46:25
2025-04-01T06:45:33.858765
{ "authors": [ "ansd", "michaelklishin" ], "repo": "rabbitmq/rabbitmq-website", "url": "https://github.com/rabbitmq/rabbitmq-website/pull/1050", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
426544406
Update access-control.md Fixed a typo. @AnthonyMastrean Please sign the Contributor License Agreement! Click here to manually synchronize the status of this Pull Request. See the FAQ for frequently asked questions. @AnthonyMastrean Thank you for signing the Contributor License Agreement! Thank you!
gharchive/pull-request
2019-03-28T15:03:57
2025-04-01T06:45:33.861581
{ "authors": [ "AnthonyMastrean", "michaelklishin", "pivotal-issuemaster" ], "repo": "rabbitmq/rabbitmq-website", "url": "https://github.com/rabbitmq/rabbitmq-website/pull/732", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
743086578
Another file manager named hunter ? Who was the first ? https://github.com/thindil/hunter https://github.com/rabite0/hunter/ Who's gonna be the first to change its name ? To be more meaningful? ;-) I can answer the first question: https://github.com/rabite0/hunter/commit/67c973c0af1d35f0f832d4eb594c12868ef78008 https://github.com/thindil/hunter/commit/9768a07787d863614a10a3d38a63f5c85560dad0 So this one was a few months earlier. The other one don't plan to change its name (see https://github.com/thindil/hunter/issues/5#issuecomment-727273187). OK we will have to live with two file managers with the same name.
gharchive/issue
2020-11-14T21:56:00
2025-04-01T06:45:33.875159
{ "authors": [ "jpggithub", "magnetophon" ], "repo": "rabite0/hunter", "url": "https://github.com/rabite0/hunter/issues/112", "license": "WTFPL", "license_type": "permissive", "license_source": "github-api" }
406354591
[OLGA] Added registred students fix bugs, fix test Description Added count of registered students to the main page, Fixed test, and bug with student countries percentage. Youtrack https://youtrack.raccoongang.com/issue/OLGA-58 Codecov Report Merging #219 into develop will increase coverage by 0.21%. The diff coverage is 100%. @@ Coverage Diff @@ ## develop #219 +/- ## =========================================== + Coverage 96.21% 96.42% +0.21% =========================================== Files 23 23 Lines 475 476 +1 =========================================== + Hits 457 459 +2 + Misses 18 17 -1 Impacted Files Coverage Δ web/olga/analytics/models.py 97.76% <100%> (+0.01%) :arrow_up: web/olga/analytics/views.py 94.15% <100%> (+0.64%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 164ec55...99359a3. Read the comment docs.
gharchive/pull-request
2019-02-04T14:29:10
2025-04-01T06:45:33.883635
{ "authors": [ "arsentur", "codecov-io" ], "repo": "raccoongang/OLGA", "url": "https://github.com/raccoongang/OLGA/pull/219", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
691934113
new Covid 19 page still needs lots of work. will be useful to see this in the review app as it's developing though. @PanosParalakis would you be able to add a route to this page on Friday please?
gharchive/pull-request
2020-09-03T12:38:16
2025-04-01T06:45:33.884850
{ "authors": [ "spencerball" ], "repo": "racedisparityaudit/ethnicity-facts-and-figures-publisher", "url": "https://github.com/racedisparityaudit/ethnicity-facts-and-figures-publisher/pull/1329", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
975209515
🛑 Pet Project is down In 2620814, Pet Project ($RINFO_SITE) was down: HTTP code: 0 Response time: 0 ms Resolved: Pet Project is back up in 3e28d62.
gharchive/issue
2021-08-20T02:49:24
2025-04-01T06:45:33.887050
{ "authors": [ "rachitkataria13" ], "repo": "rachitkataria13/upptime", "url": "https://github.com/rachitkataria13/upptime/issues/847", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
131416727
pyrax via pypi breaks duplicity Hello pyrax maintainers! It looks like recent versions of pyrax in pypi break integration with duplicity 0.6.24-1. jenkins@fff8561e79b2:/$ /usr/bin/duplicity --no-encryption restore cf+http://my_awesome_files_bucket tmp Traceback (most recent call last): File "/usr/bin/duplicity", line 1509, in <module> with_tempdir(main) File "/usr/bin/duplicity", line 1503, in with_tempdir fn() File "/usr/bin/duplicity", line 1336, in main action = commandline.ProcessCommandLine(sys.argv[1:]) File "/usr/lib/python2.7/dist-packages/duplicity/commandline.py", line 1059, in ProcessCommandLine backup, local_pathname = set_backend(args[0], args[1]) File "/usr/lib/python2.7/dist-packages/duplicity/commandline.py", line 952, in set_backend globals.backend = backend.get_backend(bend) File "/usr/lib/python2.7/dist-packages/duplicity/backend.py", line 163, in get_backend return _backends[pu.scheme](pu) File "/usr/lib/python2.7/dist-packages/duplicity/backends/_cf_pyrax.py", line 67, in __init__ % (e.__class__.__name__, str(e)), TypeError: __str__ returned non-string (type Error) jenkins@fff8561e79b2:/$ pip freeze | sort Babel==2.2.0 bzr==2.7.0.dev1 cffi==0.8.6 chardet==2.3.0 cliff==1.15.0 cmd2==0.6.8 colorama==0.3.2 configobj==5.0.6 cryptography==0.6.1 debtcollector==1.2.0 dh-virtualenv==0.11 docutils==0.12 duplicity==0.6.24 ecdsa==0.11 funcsigs==0.4 html5lib==0.999 ip-associations-python-novaclient-ext==0.1 iso8601==0.1.11 jenkins-job-builder==1.4.0 jenkins-jobs-slack==0.3.2 jenkins-view-builder==0.5.0 Jinja2==2.7.3 keyring==8.2 keystoneauth1==2.2.0 lockfile==0.8 lxml==3.4.0 MarkupSafe==0.23 mercurial==3.1.2 mock==1.0.1 monotonic==0.6 msgpack-python==0.4.7 multi-key-dict==2.0.3 ndg-httpsclient==0.3.2 netaddr==0.7.18 netifaces==0.10.4 oauthlib==0.6.3 ordereddict==1.1 os-diskconfig-python-novaclient-ext==0.1.3 oslo.config==3.4.0 oslo.i18n==3.2.0 oslo.serialization==2.2.0 oslo.utils==3.4.0 os-networksv2-python-novaclient-ext==0.25 os-virtual-interfacesv2-python-novaclient-ext==0.19 paramiko==1.15.1 pbr==1.8.1 Pillow==2.6.1 ply==3.4 prettytable==0.7.2 pyasn1==0.1.7 pycparser==2.10 pycrypto==2.6.1 Pygments==2.0.1 PyJWT==0.2.1 pyOpenSSL==0.14 pyparsing==2.0.7 pyrax==1.9.7 python-jenkins==0.4.12 python-keystoneclient==2.1.1 python-novaclient==2.27.0 pytz==2015.7 PyYAML==3.11 rackspace-auth-openstack==1.3 rackspace-novaclient==1.5 rax-default-network-flags-python-novaclient-ext==0.3.2 rax-scheduled-images-python-novaclient-ext==0.3.1 requests==2.4.3 roman==2.0.0 simplejson==3.8.1 six==1.8.0 Sphinx==1.2.3 stevedore==1.10.0 unicodecsv==0.14.1 urllib3==1.9.1 virtualenv==1.11.6 wheel==0.24.0 wrapt==1.10.6 This manifests whether pyrax versions 1.9.6 and 1.9.7 are installed via pypi. However if I pip install off HEAD in master (currently 8b55db5), things are a-ok. jenkins@c4df005fcfcd:/$ pip freeze | sort Babel==2.2.0 bzr==2.7.0.dev1 cffi==0.8.6 chardet==2.3.0 cliff==1.15.0 cmd2==0.6.8 colorama==0.3.2 configobj==5.0.6 cryptography==0.6.1 debtcollector==1.2.0 dh-virtualenv==0.11 docutils==0.12 duplicity==0.6.24 ecdsa==0.11 funcsigs==0.4 html5lib==0.999 ip-associations-python-novaclient-ext==0.1 iso8601==0.1.11 jenkins-job-builder==1.4.0 jenkins-jobs-slack==0.3.2 jenkins-view-builder==0.5.0 Jinja2==2.7.3 keyring==8.2 keystoneauth1==2.2.0 lockfile==0.8 lxml==3.4.0 MarkupSafe==0.23 mercurial==3.1.2 mock==1.3.0 monotonic==0.6 msgpack-python==0.4.7 multi-key-dict==2.0.3 ndg-httpsclient==0.3.2 netaddr==0.7.18 netifaces==0.10.4 oauthlib==0.6.3 ordereddict==1.1 os-diskconfig-python-novaclient-ext==0.1.3 oslo.config==3.4.0 oslo.i18n==3.2.0 oslo.serialization==2.2.0 oslo.utils==3.4.0 os-networksv2-python-novaclient-ext==0.25 os-virtual-interfacesv2-python-novaclient-ext==0.19 paramiko==1.15.1 pbr==1.8.1 Pillow==2.6.1 ply==3.4 prettytable==0.7.2 pyasn1==0.1.7 pycparser==2.10 pycrypto==2.6.1 Pygments==2.0.1 PyJWT==0.2.1 pyOpenSSL==0.14 pyparsing==2.0.7 pyrax==1.9.6 python-jenkins==0.4.12 python-keystoneclient==2.1.2 python-novaclient==2.27.0 pytz==2015.7 PyYAML==3.11 rackspace-auth-openstack==1.3 rackspace-novaclient==1.5 rax-default-network-flags-python-novaclient-ext==0.3.2 rax-scheduled-images-python-novaclient-ext==0.3.1 requests==2.9.1 roman==2.0.0 simplejson==3.8.1 six==1.10.0 Sphinx==1.2.3 stevedore==1.10.0 unicodecsv==0.14.1 urllib3==1.9.1 virtualenv==1.11.6 wheel==0.24.0 wrapt==1.10.6 Package diffs between a pypi install vs git+https install: $ diff /tmp/{1.9.7,master} 29c29 < mock==1.0.1 --- > mock==1.3.0 57c57 < pyrax==1.9.7 --- > pyrax==1.9.6 59c59 < python-keystoneclient==2.1.1 --- > python-keystoneclient==2.1.2 67c67 < requests==2.4.3 --- > requests==2.9.1 70c70 < six==1.8.0 --- > six==1.10.0 I tried comparing master to the v1.9.6 tag in GitHub but somethings fishy showing no differences. Also there doesnt appear to be a v1.9.7 tag in GitHub. master, v1.9.6 and v1.9.7 are all the same at the moment. There was a packaging issue causing us to bump to 1.9.7 and re-release. Based on all of those being the same, I am not sure why there would be any difference. Their setup.py files all indicate the exact same dependencies. Those dependencies are listed as: "python-novaclient==2.27.0", "rackspace-novaclient", "keyring", "requests>=2.2.1,<3", "six>=1.9.0,<2", The only thing I can see there is that your install of 1.9.7 is somehow installing six==1.8.0 which does not satisfy the pyrax dependency. We run into this in a Jenkins Docker container. The relevant pip snippets from the Dockerfile are as follows: # Upgrade Python tools RUN pip install pip==7.1.2 RUN pip install setuptools==18.2 RUN cp /usr/local/bin/pip /usr/bin/pip RUN cp /usr/local/bin/pip /usr/bin/pip2 ... RUN pip install jenkins-job-builder jenkins-view-builder jenkins-jobs-slack RUN pip install --upgrade git+https://github.com/rackspace/pyrax.git That allows us to get duplicity working. We have run into issues in the past with pbr (which is a dep of the upstream openstack modules) causing deps to be installed that do not satisfy the pyrax specifications. This may be due to pyrax not being reliant on pbr. In most cases, a second pip install --upgrade pyrax will usually resolve any dependency inconsistencies. Ugh. pbr gonna pbr ... Thanks, @sivel.
gharchive/issue
2016-02-04T17:22:39
2025-04-01T06:45:33.952780
{ "authors": [ "filler", "sivel" ], "repo": "rackspace/pyrax", "url": "https://github.com/rackspace/pyrax/issues/603", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
117638665
Having problem trying to implement redux auth middleware. The following code is how my auth middleware for redux looks like: import {replaceState} from 'redux-router'; export default store => next => action => { if (!action.requireAuth || store.getState().auth.user) { return next(action); } return next(replaceState(null, '/login')); // also tried store.dispatch(replaceState(null, '/login')); }; However, this fails to change url. It basically has no effect. Also, store.history is always null regardless whether I put it before or after reduxRouter middleware. Can someone explain what I am doing wrong? Is there an elegant way to use redux-router history actions in middleware? Also, how to access to store.history inside custom redux middleware? I've noticed that both store and store.dispatch Fighting with similar issue for a week... Looks like it's impossible. So I'm thinking how to reimplement redux-router with "auth-middleware" functionality. Yep, middleware seems to be problematic. If you want to do authentication look at this example via a higher order component: react-redux-jwt-auth Yes, saw that approach. But I think that this is wrong architecturally. Any action in my system could lead to auth checks and redirects. Maybe some actions (even not "routing actions") must be stopped. Maybe some other tactics must be applied to my system in response to some actions. Middleware is right place for that (for me). Is it wrong way to provide additional option for reduxReactRouter - "additionalMiddlewares" that will be applied inside reactReduxRouter after historyMiddleware? PS: sorry for my English @ALF-er No, middleware is definitely the right solutions for these problems. Currently there is not really active developing going on here. So I point the people to the higher order component solution, as we can´t predict if the middleware problem will be fixed any time soon...or ever. Do you accept PRs? @ALF-er Of course, though they should be described and include necessary test(s). Also take a look at this this: https://github.com/rackt/redux-router/issues/172 Thank you! Will investigate what is better: PR for redux-router or redux-simple-router =)
gharchive/issue
2015-11-18T17:32:33
2025-04-01T06:45:33.958488
{ "authors": [ "ALF-er", "Scarysize", "joonhocho" ], "repo": "rackt/redux-router", "url": "https://github.com/rackt/redux-router/issues/181", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
134090391
Event expressions - RFC Not ready to merge, no tests yet, and (almost definitely) 0.9 This is my attempt to resolve #1172, #1396, #2311, and partially #2176. It is also somewhat tied to #2369. What this does This is basically a mix of options 1 and 3 from #1172 with a twist. This takes event handling and unifies it as an expression instead of having special parsing for handling method events. This opens event handling up to everything possible with an expression in Ractive. Here's a list of benefits I've seen thus far: can call data methods e.g. on-click=".loadStuff()" can call more complex structures e.g. on-click="@this.reset.partial('foo')" or on-click=".foo().and.bar(., 'yep')" can call parent methods for tightly integrated (container) components e.g. on-click="@this.parent.doAThing(.)" chain multiple actions in an event handler e.g. on-click="@ths.toggle('foo') && @this.fire('fooed')" the code is a lot smaller and generally easier to follow as a result I have also taken the spread args and made it available anywhere, which is what ES2015 allows e.g. on-foo="fn('hi', ...arguments, 42)" is perfectly legal. I'm thinking this should be extended to any invocation expression that wants an array spread into its args, because ES2016. I implemented this by half-shimming with an IIFE based on Babel's output. There may be a better way, but this didn't require an extensive refactor. Additionally, dollar args ($1) and arguments refs are allowed to have children as per #2311. These are handled in the same way as event refs with children. Comments specifically requested Method events are handled in their current state my detecting basically ^someName(.*)\s*$, prepending @this., and handling as an expression. This seems to work pretty well in practice, but is it confusing that toggle('foo'), .toggle('foo'), and ~/toggle('foo') mean potentially very different things? With the confusion surrounding where method events actually call, should they just be deprecated and eventually removed in favor of @this.method(args)? Deprecating and removing plain method call events would eliminate the remaining inconsistency in template reference resolution. I haven't addressed proxy events in this PR yet, because I'd like some feedback first. I would prefer to support only no-args events (option 3) for auto-proxying, which would be super simple to support here. Is that what most people would like to do? The only other thing I consider tenable from a parsing (code and mental) viewpoint is to support on-click="nameExpression:argsList". That would make on-click="{{ foo ? 'bar' : 'baz'}}: {{bop}},bip,{{ { oh: 'my' } }}" become on-click="( foo ? 'bar' : 'baz' ): bop, 'bip', { oh: 'my' }". I also haven't addressed decorator and transition args yet, because I'd like some feedback first. I would prefer to drop mustaches in the directives entirely and just move to comma separated expression lists for decorators and transitions. That combined with as-decoratorName and transitionName-in/transitionName-out/transitionName-in-out (or something similar) and the changes in #2369 would allow that. This would be more consistent with event directives. Another proposal has been to treat them as method calls instead e.g. decorator="decoratorName(arg1, arg2)" which seems reasonable, but I think it's less consistent and could get weird with transitions with their in/out/in-out facets. How should deprecation be handled? My preferences Even though it's longer, @this.toggle('foo') is much more clear to me when looking at a template. I'm cool with just on-click="clicked" as a shortcut for the common re-proxy case. Mustache arguments on existing proxy events are way too confusing. I'm also not a fan of the : or the other proposed [arg1, arg2] syntax, and from the comments in #1172 it's clear that using a sigil for regular method-like syntax was a bad idea. I like as-decoratorName="arg1, etc" and fade-in-out="arg1, etc". That would entirely eliminate special syntaxes for non-bound directives (lazy and twoway can be bound in #2369). This just looks right to me <input as-validated="@this.validators.date" value="{{.when}}" /> {{#if shown}}<div fade-in-out>...</div>{{/if}}. I'd like to create a new branch for 0.9 to merge the big nasty changes ahead of time, while removing support for deprecated features. Then we could handle the deprecation notices in 0.8 without having a ton of convoluted duplicate code. That would make 0.8 and 0.9 incompatible, but it would still let people blocked on bugs in 0.7.3 have a less immediate headache. Status [x] Handle method events as expressions [x] Account for spread, arguments, and dollar args in expressions [x] Allow children of dollar args and arguments [x] Allow spread args in any invocation (it's there, so may as well use it, no?) [x] Tests , tests, and more tests - the deprecation path leaves the old proxy handling code in place, so only expression-related stuff is tested beyond the existing event tests [x] Finalize deprecation path Clean up code mostly deferred until after deprecation cycle - there's lots of code that can go away and/or be nicely refactored With the confusion surrounding where method events actually call, should they just be deprecated and eventually removed in favor of @ractive.method(args)? I'm in favor of this one. But I'd suggest @instance. @ractive looks like a reference to the Ractive global. Thinking something like @ractive.version and stuff like that. I haven't addressed proxy events in this PR yet, because I'd like some feedback first. I would prefer to support only no-args events (option 3) for auto-proxying, which would be super simple to support here. Is that what most people would like to do? Method calls all the way! Burn the old event syntax! :fireworks: :fire: :fire_engine: Another proposal has been to treat them as method calls instead e.g. decorator="decoratorName(arg1, arg2)" which seems reasonable, but I think it's less consistent and could get weird with transitions with their in/out/in-out facets. I method calls, possibly just args I guess? <div transition="transitionName({ in: { duration: 5 }, out: { duration: 10 }})></div> ...or maybe multiple setup calls <div transition="transitionName('in', 5); transitionName('out', 10);"></div> Not sure how transitions are written, just throwing in ideas. How should deprecation be handled? Partial support in 0.8 with warning log, remove from code by 0.9. Isn't that the usual mo? Even though it's longer, @ractive.toggle('foo') is much more clear to me when looking at a template. Explicit is always better than implicit. :smile: What @fskreuz said. @ractive doesn't seem to imply the current instance. @this was discarded? @current maybe? I really @instance to me, it seems much more intuitive than @ractive -- which I would assume would be the same as Ractive - as in Ractive.extend() I originally went with @ractive because just about every piece of code I've seen in the wild using Ractive has var ractive = new Ractive(...) somewhere towards the top. I hadn't considered that @ractive would point to the constructor because it's all lower case. I would think that if that were the case, it would be @Ractive based on convention (jshint, eslint, etc). Also, I think of @instance as a bit generic, so it doesn't stand out as strongly as a Ractive instance to me. At any rate, if that is to be changed it needs to be soon, because @ractive has already been out in edge for about two months but would only have been used by people who watch PRs pretty closely. This is probably as good a place as any since it is proposed to make the special ref mandatory for event methods... anyone have any strong preference? After some consideration, I still prefer @ractive. For me <on-click="foo()"> is nicer than <on-click="@ractive.foo()">, but only because it has been entrenched that foo() sits on the Ractive instance. I understand the need to align events to resolve similar to references, so I think this is a good change overall. Once folk gets used to calling @ractive.foo() maybe we can revisit shortening it to just @foo(). I appreciate the arguments against adding symbols to the template, but once @ractive.xxx is entrenched, maybe we would appreciate a shortcut more ;-) actually, I kind of like @sabob 's idea that all @ would reference the instance which is similar to coffeescript's convention to use @foo -> this.foo. but since the @ is reserved for special references (like @global), maybe instead what if it were @@ so, @@foo() would be the same as (what is now) @ractive.foo() @evs-chris I'm convinced :smile: Noted down in the docs repo to document the different @ things. @sabob @heavyk The use of @ isn't native to JS so it might be weird to newcomers. Not everyone writes CoffeeScript either. I think it should be left as @something instead of standalone @ to give it more meaning. yeah I agree with @something convention now that I think about it.. I guess my my order of preference would probably be: ~/foo() @self.foo() or @instance.foo() @ractive.foo() Using @something to refer to component instance a. I agree with @evs-chris that var ractive = new Ractive(...) provides a cognitive connection to @ractive. b. Current method events, while now familiar to us, don't fit with the rest of data refs in ractive. Looking at: on-click="clicked( foo )" why would I know that foo referred to data context but clicked referred to component context? When @ractive means the current component instance across the board then it should be easier for new people coming to ractive (though a bit of a migration road bump for us old-timers :) ): on-click="@ractive.clicked(foo)". c. We will get tired of typing @ractive.foo and @ractive.root.user and it will become an eyesore in our templates to have @ractive all over the place, which is why I introduced #2345 to introduce @foo as a shortcut for @ractive.foo and @@foo as a shortcut for @ractive.root.foo. Yes, it makes the built-in @keypath, @index a bit odd - but reserved keywords are nothing new. Proxy-event shortcut: Agree that having on-click="handler" as a shortcut for on-click="@ractive.fire( 'handler', event )" (assuming I'm understanding @evs-chris correctly) makes a lot of sense. Events bubble and fit a set of use cases different than method handlers. With the use of this.event, a very large percentage of events can be dispatched this way. Having it be a shortcut for @ractive.handler() doesn't really save you much, especially if you can write @handler() Plugin syntax a. Agree that if all element plugins use the attribute name to communicate the plugin type and name it makes things more consistent: on-click, as-tooltip, fade-in. b. Event handlers are fundamentally different than other plugin attribute values because they are arbitrary expression to respond to an event, whereas decorators and transitions are intialization values for the plugin. It will make decorators and transitions more terse if the attribute value is interpreted as an argument list: as-thing='expression' where '[' + expression ']' is what gets "interpreted" in the parser and passed to fn.apply( ractive, args ) when actually evaluated. It enables both single or double value passing via ="foo" or =" qux, 'BAR' " as well as optionspattern js:="{ foo, bar: 'BAR' }". Begin the deprecation warnings! @martypdx hmmm, I actually never cognitively connected that inconsistency with the data and instance methods (and I've written a lot of ractive now!!). I wonder how I missed that. so what you're proposing then is for it to instead be @clicked( foo )? TBH, I could dig that... I do like that syntax introduced in #2345, a lot :+1: @martypdx regarding 2, after the deprecation, yes, I would have the parser rewrite the argument of on-click="([^(]+)" as @ractive.fire($1, ...arguments). In other words as long as it's obviously not a call, it would reproxy the named event with any arguments passed (including event, if present). @fskreuz considering the syntax proposed in: #2345. I think @this would make a lot of sense instead of @ractive. The same way that this.foo would ref the foo in the context and @foo would ref the foo in the ractive instance. this would ref the context and @this would ref the instance. Consistency! @guilhermeaiolfi I'm starting to like to the idea of @this too let's just alias @this to @ractive and call it good. everyone wins and no deprecations for us old-timers :) I've got to third the @this. Although, If we do go with @this, do we even need the @? call data methods: on-click="loadStuff()" or on-click="../loadStuff()" or on-click="~/loadStuff()" call proto methods: on-click="this.loadStuff()" call complex structures on-click="this.reset.partial('foo')" or on-click="this.foo().and.bar({{this}}, 'yep')" call parent methods on-click="this.parent.doAThing({{this}})" chain multiple actions on-click="this.toggle('foo') && this.fire('fooed')" It kinda works, and avoids the @ completely, but uses {{ }} mixed in to pass expressions... which may be inconsistent with how expressions are used in other places (but this inconsistency already exists < 0.8). Although, If we do go with @this, do we even need the @? Mixing mustaches inside an expressions has been a PITA, both from an implementation perspective and IMO as a developer because you end up in "mixed-mode" javascript and html template land. The goal is to use mustaches for html and text interpolation, allow javascript expressions inside mustaches and inside plugin attribute directive values. The other benefit to use a marker instead of assuming the first token of an event handler applies a method name is that the expression itself can more more dynamic as multiple actions can be invoked: on-click="@this.set( 'x', foo ), @this.set( 'y', bar )" on-click="@this.push( 'items', newItem ).then( _ => @this.set( 'newItem' ) )" [note: AFAIK, neither of those examples is currently supported even with this PR, @evs-chris?] While arguably there is an initial learning in having both this and @this, @ would be recognizable as a special keyword. Ultimate syntax test would be something like: {{#each items}} <li on-click="@this.select( this )">{{this.name}}</li> {{/each}} I was going for succinctness in shortening that to @select but I could live with @this.select and it avoids the javascript weirdness of concatting obj.prop down to objprop. @martypdx no, but on-click="@ractive.set('x', foo) && @ractive.set('y', bar)" is. I don't know what @RichHarris' stance on arrow functions in the template is, but plain functions where explicitly excluded. I've found myself wanting simple, single statement arrows for mapping and the like before, but it's usually better to pull those out of the template. In-template functions are an entirely different 50 gallon drum of worms, though. no, but on-click="@ractive.set('x', foo) && @ractive.set('y', bar)" is That's what I thought, but that's a bit of a hack and feels error prone when people use them with functions that have a falsey return. Maybe allow the comma operator so this expression can be properly, um, expressed: on-click="@ractive.set('x', foo), @ractive.set('y', bar)" In-template functions are an entirely different 50 gallon drum of worms, though. yep. and future drum of worms for sure. It's not exactly the comma operator per se, but I added support for a list of expressions. The last value in the list is the one used for cancelling. After looking at it for a bit, I decided @this is probably not too confusing and reads and types a little better than @ractive. I personally like #2345, but I can't really gauge it's overall reception at this point. I could see how @this makes the possibility of it not landing a bit more palatable. I also added deprecation messages for old-style method and proxy-with-args events, as the consensus seems to be to move towards consistency and the existing method events are not consistent. Also, dropping method events and making proxy events simple enough to easily parse them into an expression would let a fair amount of code be put out to pasture after the deprecation cycle. I strongly agree with @fskreuz's comment: this means the current context in templates (synonymous with .). Using @this to mean the current instance can make it confusing. Making this and @this with totally different meanings is just confusing! @this and this doesn't sit well with me either. If my template ends up as: <li on-click="@this.select( this )">, I would go back and remove both this if possible to <li on-click="@ractive.select( . )"> Certainly @this is shorter to type than @ractive, but it means I have to stop and think about what the different this things mean in my template (or someone else's) Also it seems once we switch to @this and .this, this becomes meaningless and the useful bits are really the @ and ., this can be removed. I'm still in favour of the longer @ractive and see where things go from there. There are a couple of releases before v1 to iron things out. I like @this because it gels with the JS meaning. It is a bit close to this, but then any special can have a same-named reference in scope at any given point in time. I've found that the @ is a darn good visual cue on meaning because it really stands out in templates. The only other suggestion that looked ok to me was @self. I sometimes escape anonymous function binding issues when I'm ES5-ing it with var me = this;, so @me looks fairly natural to me too. So what would the vote be for the Ractive instance special ref: @this, @ractive, @self, @me, or something else? I like @this, then @me, with a tie for third with @ractive and @self. For me the answer lies in the question: So what would the vote be for the Ractive instance special ref @instance as suggested by @fskreuz It's barely longer than @ractive and it's totally clear what it means (as opposed to any other suggestion, imho) My order of preference is @ractive.something then @this.something, and if all goes well, at a later stage just @something As the joke goes "if two developers agree during a meeting, that is a majority". At the end of the day once a decision is made and documented I don't think it will matter much which keyword we use. Summary Problematic names which may cause confusion: @this: too close to this used to denote the current context in expressions (same as .) @ractive: one may think it refers to the Ractive global instead of the instance Suggested so far (in alphabetical order to not make anyone jealous ;) ): only @ or @@ @current @instance @me @self My vote goes to @self and @instance (tie). Side note of existing symbols (as far as I know, independent of the context): @index @keypath @global @key What the problem of @this being too close to this? Both are pretty close in meaning too. For people voting, have you read #2345? I know that it's a little too radical (and maybe late) and have a hard chance of acceptance but I would like to share two ideas: this (without @) to always ref the ractive instance and let just .foo to ref the context. It's more aligned with OOP. @this/@ractive/@me/whatever seems a little off considering others symbols like @index and @key. Since v0.8 it seems the recommendation seems to use {{.foo}} to avoid unnecessary lookups to get better performance. But using the context data seems the most used case in templates. So why not make it the default option? So you'd have {{foo}} (without any special char) to always be a reference to the context and something like {{^foo}} an expression that goes up to the tree looking for references. That way you know when you can get data from outside you component. That would be easier for people coming from others mustache based frameworks too and IMO easier to read. What the problem of @this being too close to this? Both are pretty close in meaning too. They are not close at all. @this would denote the current ractive instance, while this denotes a nested object belonging to data. this (without @) to always ref the ractive instance and let just .foo to ref the context. The only issue is that this is part of ractive since years. Changing it would be a major (read catastrophic) compatibility break. FWIW this.foo and .foo parse to exactly the same thing. They are not close at all. @this would denote the current ractive instance, while this is an already existing keyword denoting a nested object belonging to data. So, same idea, different scope. The only issue is that this is part of ractive since years. Changing it would be a major (read catastrophic) compatibility break. I know. That why I think it's hard to be accepted. But heyy, 0.8, now is the time to change things before 1.0. Let's make it right. FWIW this.foo and .foo parse to exactly the same thing. Sure. I was purposing to change that. I can leave with any of the options. But I think we can do better if we are willing to break compatibility. Thinking in the long run, now is the time. Let's decide for the future and not for the present. I think it would be best to stay consistent with references, where @ means "this is a meta reference that has something to do with where I am in a template construct and nothing to do with data". I think swapping this to mean @this would be considerably more confusing to newcomers based on issues that I've seen over the years. Having two distinct classes of refs is a bit easier to track: things that are in my data things that are semi-magical meta-references injected by the template . - current context @index - the current index of the nearest iteration this - more visible alias for . @key - the current key of the nearest object iteration foo - foo somewhere in data - or - an ambiguous reference that may resolve with an alias, a named index reference, or at any point up the context's hierarchy @keypath - the current context keypath relative to the component .foo - foo property of current context @rootpath - the current context keypath relative to root ../foo - foo property of context parent @global - the global object for the environment ~/foo - foo at root of data @this or @ractive or @something - the Ractive instance for the nearest component or just the root instance if none ??? Profit! At any rate, would it be better to make the @ractive special ref configurable so that each person/team can swap it for whatever is more meaningful/visible/obvious to them? Personally, I'm opposed to anything longer than @ractive for purposes of concision in templates. I frequently have elements that have multiple event handlers and I'd like to avoid getting past column 200 whenever possible. At any rate, would it be better to make the @ractive special ref configurable so that each person/team can swap it for whatever is more meaningful/visible/obvious to them? I prefer it to be something that I don't like now but are consistent between projects than have to keep this one more thing in my mind when going from one project to another because the previous team decided that @mypuppy thought it would be cool. Problematic names which may cause confusion: @this: too close to this used to denote the current context in expressions (same as .) @ractive: one may think it refers to the Ractive global instead of the instance @ractive seems so weird. If it's derived from var ractive = new Ractive(...), that ractive var is something used outside of the prototype/template, elsewhere on the page—possibly with other instances of Ractive. By that logic we could also call it @main or @app since those are frequently used (var main = new Ractive()). And that would be really weird. I still think @this is most succinct, but @instance isn't bad either. instance is also used all over the docs and in example decorator code, var instance = this. Makes sense. When you're trying to call an "instance" method, you'd use @instance. @this also makes sense because something on the "instance" is really on the prototype, and if you're familiar with prototypal calls, you use this to reference something. I guess my vote would be for both @this & @instance but not @ractive? ¯_(ツ)_/¯ @this is too confusion-prone. What about @context? @aharpervc Then this question comes up: "context to what?". The word "context" also has special meaning in Ractive, i.e. "the context stack" in templates. this is going nowhere fast :) -- what if it were @specimen instead? while examining my code, I like to think of myself as a detective examining specimens of my work :D oh, come on folks! Let's keep it concise! .@evs-chris : for the sake of sanity, please just pick one you think adequate. @fskreuz yeah, I was intentionally paralleling the use of "context" as in the right side column in @evs-chris's table Coming back to it after a bit of time has elapsed, @this still looks good, so I'm going to leave it at that. I've also updated the intro comment with the result of the other bits of discussion here. The general consensus seemed to be that this is a good way forward, so any thoughts on getting this in for 0.8 or 0.9 (or not at all)? So much goodness to wait until 0.9. Since it's ready, ship it. 0.8 from me. 0.8, ho! 0.8 is good from this end :+1: I've been coming back to something @guilhermeaiolfi said: this (without @) to always ref the ractive instance and let just .foo to ref the context. It's more aligned with OOP. @this/@ractive/@me/whatever seems a little off considering others symbols like @index and @key. If we put aside the BREAKING CHANGE aspect, I prefer an API where {{this}} always referred to the ractive instance, both in mustache references and in event handlers. You could still use {{.}} to mean the model context. In fact you could argue it's more consistent {{@this}} means current context, like current @index, @key, @keypath. Plus with aliases in 0.8 you have other options for naming model contexts beyond "this". Thoughts? @martypdx No complaints about that. I concede to whatever Chris and you like there. If you follow that train of thought, shouldn't all data/context refs start with @ and everything else be reserved keywords? So... @global -> global @root -> root @this -> this this -> .' To remain consistent @martypdx well I'm not really for it, but I'm not really against it either. I like that @ stuff stands out as 'not in my data' in the template. I typically use the . form of scoped references e.g. .foo rather than this.foo, so it doesn't really affect me on that front. I don't know how often people use this in the wild, but it's not in either mustache or handlebars as far as I can tell. From what I've seen from issues, more people tend not to use scoping at all. It looks like this was added in b74fb4df0d8189be46b7472dd675ea4ae937161f so it's been around for a while with its current meaning. All that said, I think it would be pretty easy change to make. Anyone else (especially @Rich-Harris)? @evs-chris, @JonDum for the record, I'm not actually advocating @this needs to get added under that scheme, just pointing out the option exists if people think {{.}} is inadequate. btw - self is the apparently the preferred way to get to global scope across browser, web and service workers, and node. I wonder if we might keep things simple via these mapped changes: this -> . @global -> self @this -> this @root -> this.root Only first item would be breaking change as others have not been introduced. good point (especially the service workers), I think that mapping to is quite reasonable. is there some way that could be combined in some way with your state idea in #2345 too? maybe we can define our own @ vars? or, heaven forbid a @state variable? (prolly not) I've had multiple occasions now, where having access to state would have been a more desirable approach (even though it's simple enough to add state from an external object into the template) Thread becoming long, adding just one more to hopefully seal the deal. Let's all agree to this proposal): Anything **without **@ means it's a valid JS expression. Anything **with **@ means it's something Ractive-specific. It becomes clear when you put it like this: I like that @ stuff stands out as 'not in my data' in the template. https://github.com/ractivejs/ractive/pull/2386#issuecomment-215620950 this is a meta reference that has something to do with where I am in a template construct and nothing to do with data https://github.com/ractivejs/ractive/pull/2386#issuecomment-187807506 That should be enough of a guideline to avoid confusion. We should steer away from keyword-specific to a broader view (had to re-read the thread to realize). A doc page for @-things should also be sufficient, sort of like a glossary. Really looking forward to 0.8 (as if I weren't using edge all this time). @evs-chris @fskreuz @this it is! I think we should merge this now, because it changes or supports a number of other features that are already in edge (like @ractive) Is the '@this' prefix required from now on? Does writing on-click="fire('eventNameToBeFiredOnThis', event)" still work? @kouts yes, it works without @this: http://jsfiddle.net/p9hbej0k/ @kouts, @guilhermeaiolfi Though it is deprecated if you check the console: Ractive.js: Unqualified method events are deprecated. Prefix methods with '@this.' to call methods on the current Ractive instance. whereas default firing by name w/o args is still supported: on-click="eventNameToBeFiredOnThis" So in order to be future proof should I add the '@this' to all my event calls inside my templates? Wouldn't it be better if the '@this' part was optional? Yeah, I think I said that somewhere. Not just for events, for properties in general it would be nice to avoid using '@this'. I would like to avoid those features that differ ractive from pure javascript. The exception being when there is analog way of doing that in javascript. I've been going through all my components preparing them for this change. Now having used it extensively... I'm really not a fan of all the clutter. 90% of the time I'm just doing a set(). It's far more uncommon for me to have methods in my data. Wouldn't it be better if the '@this' part was optional? +1 to that. If you did want to call a method on your data, can't you use .foo(), ../foo(), ~/foo() anyways? If we don't deprecate the old syntax and just let it remain everything works pretty dandy: http://jsfiddle.net/ddxgxupr/ (except for the {{#with foo}} {{ .test() }} {{/with}} that doesn't appear to work, may just be a bug) I'm personally opposed to anything that makes the same reference resolve to different things in different contexts. My designer buddies have a hard enough time with one set of rules :laughing: I think we should land something like #2345 that, among other things, makes @set('foo', bar) equivalent to @this.set('foo', bar). Would that get close enough? I've already switched to @this, but an alternative is always nice to have. well, I'm not fan of @this because I still think people will confuse it with this, which have totally different meanings. Now you can do this: <button on-click="@this.do(this)">foo</button> <button on-click="this.do(@this)">bar</button> If I'd be a noob, I'd just be confused :astonished: I'm a huge fan of the #2345 PR (but more for the state part about it). the @ operator feels natural to me cause I've spent a lot of time writing livescript. it's also the same in coffee-script: @set 'lala', 2 is shortcut for this.set('lala', 2) that being said, I believe @JonDum brings up a good point. in my code, I almost never use @ variables in the template (cept when I need to access global state) and I've always used ~/foo() syntax instead. so, perhaps the @ variable isn't really needed so much. I dunno... personally, I think ~/foo() is the most 'natural' for me, cause ~/ makes me think of referring to $HOME in bash.
gharchive/pull-request
2016-02-16T20:37:13
2025-04-01T06:45:34.052439
{ "authors": [ "JonDum", "aharpervc", "dagnelies", "evs-chris", "fskreuz", "guilhermeaiolfi", "heavyk", "kouts", "martypdx", "sabob", "skeptic35" ], "repo": "ractivejs/ractive", "url": "https://github.com/ractivejs/ractive/pull/2386", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2185767470
running too slow When I execute cmd_wrapper.py, I found that the execution speed during the while loop is too slow. Is there any way to improve the execution speed? from __future__ import print_function import sys import os import re import time here = os.path.dirname(os.path.abspath(__file__)) wexpectPath = os.path.dirname(here) import wexpect # Path of cmd executable: cmd_exe = 'cmd' # The prompt should be more sophisticated than just a '>'. cmdPrompt = re.compile('[A-Z]\:.+>') # Start the child process p = wexpect.spawn(cmd_exe) # Wait for prompt p.expect(cmdPrompt, timeout=5) # print the texts print(p.before, end='') print(p.match.group(0), end='') p.sendline("D:") p.expect('>', timeout=5) print('开始执行任务了') while True: # Wait and run a command. command = input("请输入:") p.sendline(command) print('开始执行...') t=time.time() print(t) index = p.expect(['>>>', 'bad', wexpect.EOF, wexpect.TIMEOUT], timeout=10) if index == 0: print('正常执行') # print the texts print(p.before) print(p.after) elif index == 1: # print the texts print(p.before) print(p.after) elif index == 2: print('出现了错误') # print the texts print(p.before) print(p.after) elif index == 3: # print the texts print('超时了') print(p.before) print(p.after) print('执行结束了...') print(time.time()) print(time.time()-t) for example when run python and import datetime,execution time more than five seconds. Actually, the speed should be very fast under normal execution. ```python Microsoft Windows [版本 10.0.19042.1706](c) Microsoft Corporation。保留所有权利。 C:\Users\caofei\Desktop\Test_All\pexpect命令的使用>开始执行任务了 请输入:python 开始执行... 1710405615.8823195 正常执行 D: D:\>python Python 3.7.9 (tags/v3.7.9:13c94747c7, Aug 17 2020, 18:58:18) [MSC v.1900 64 bit (AMD64)] on win32 Type "help", "copyright", "credits" or "license" for more information. >>> 执行结束了... 1710405622.4716284 6.589308977127075 请输入:import datetime 开始执行... 1710405633.0252686 正常执行 import datetime >>> 执行结束了... 1710405641.3068957 8.281627178192139 请输入: I tried this workaround https://github.com/raczben/wexpect/issues/53, and noticed a speedup.
gharchive/issue
2024-03-14T08:45:04
2025-04-01T06:45:34.058235
{ "authors": [ "caofei996", "dhanoosu" ], "repo": "raczben/wexpect", "url": "https://github.com/raczben/wexpect/issues/63", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
842047507
patch over sdds dependency Temporary work around hopefully until sdds import in Sirepo's opal template is dealt with. OPAL sdds dependency was resolved in Sirepo. Stephen was using the temporary patch but has been switched back to using develop.
gharchive/issue
2021-03-26T15:07:14
2025-04-01T06:45:34.133888
{ "authors": [ "cchall" ], "repo": "radiasoft/rsopt", "url": "https://github.com/radiasoft/rsopt/issues/53", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2368100956
Solve Issue 48 This PR is made for solving Issue #48. What I've discovered in my research of this problem was that a malformed JSON was made when the number of chunks was generally bigger. For example, I've found this for the actual details that @mikhailocampo provided: I've also tested them with the following files and found the same problem: https://hagan.okstate.edu/NNDesign.pdf with the topic "Neural Networks" https://diposit.ub.edu/dspace/bitstream/2445/180441/2/tfm_lichtner_bajjaoui_aisha.pdf So, I've discovered that the malformed JSON had a "properties" key in its composition, making it unappropriate for the actual Pydantic Schema. So, I developed a function to test if the response has that. If so, then it will re-try to generate another question, providing appropriate results with bigger PDF files. Although accurate, I think this is separate from the #48 Issue given that the issue encountered here is related to ValidationErrors rather than the mapping error from #48 Closed due we solved that issue in the new Develop branch
gharchive/pull-request
2024-06-23T00:55:28
2025-04-01T06:45:34.138738
{ "authors": [ "AaronSosaRamos", "mikhailocampo" ], "repo": "radicalxdev/kai-ai-backend", "url": "https://github.com/radicalxdev/kai-ai-backend/pull/49", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2147766220
UCP gets corrupted after trying to upgrade cli and radius installation Steps to reproduce Begin with an older version of rad cli installed and Radius installation on k8s cluster (v0.29 in my case) Upgrade to the latest (v0.30) rad cli using curl -fsSL "https://raw.githubusercontent.com/radius-project/radius/main/deploy/install.sh" | /bin/bash Reinstall Radius on the k8s cluster to upgrade it from v0.29 to v0.30: rad install kubernetes --reinstall Try to run rad init Observed behavior rad init results in this error: rad init Error: An unknown error was returned while testing Radius API status: Status Code: 503 Response Headers: Audit-Id: b5cbed14-b375-43d9-8907-bf36a62013b3 Cache-Control: no-cache, private Content-Length: 196 Content-Type: text/plain; charset=utf-8 Date: Wed, 21 Feb 2024 19:54:20 GMT X-Content-Type-Options: nosniff X-Kubernetes-Pf-Flowschema-Uid: c3c8ea3d-368a-4b5f-9344-590a614dba30 X-Kubernetes-Pf-Prioritylevel-Uid: 5f901053-46b1-4d28-9c29-fa1faedff1e6 Response Body: error trying to reach service: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "ucp-ca") TraceId: ac7aa107f1150e350188a8a7c8df9453 rad install kubernetes --reinstall results in a repeating error message: Reinstalling Radius version v0.30.0 to namespace: radius-system... E0221 13:19:17.841185 11297 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:17.884104 11297 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:18.702624 11297 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:18.732621 11297 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:18.910738 11297 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:18.930460 11297 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:19.188068 11297 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:19.199038 11297 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:19.518492 11297 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:19.590051 11297 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:19.821716 11297 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:19:19.886413 11297 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request Desired behavior both rad cli and Radius on the k8s cluster should be successfully upgraded to the latest version (v0.30) and rad init should complete successfully using the latest installed version of Radius. Workaround Other folks have reported this workaround to help, but it's not working for me: kubectl delete crd queuemessages.ucp.dev kubectl delete crd recipes.radapp.io resources.ucp.dev rad reinstall rad init rad Version RELEASE VERSION BICEP COMMIT 0.30.0 v0.30.0 0.30.0 f5a4a551b95168fbc2f33358bc8551c2cedece54 Operating system MacOS Monterrey, M1 chip Additional context UCP logs: 2024/02/21 21:41:30 http: TLS handshake error from 10.42.0.1:57296: remote error: tls: bad certificate │ │ 2024/02/21 21:41:31 http: TLS handshake error from 10.42.0.1:62287: remote error: tls: bad certificate │ │ 2024/02/21 21:41:32 http: TLS handshake error from 10.42.0.1:10531: remote error: tls: bad certificate │ │ 2024/02/21 21:41:32 http: TLS handshake error from 10.42.0.1:40676: remote error: tls: bad certificate │ │ 2024/02/21 21:41:33 http: TLS handshake error from 10.42.0.1:39087: remote error: tls: bad certificate │ │ 2024/02/21 21:41:47 http: TLS handshake error from 10.42.0.1:23684: remote error: tls: bad certificate │ │ 2024/02/21 21:41:50 http: TLS handshake error from 10.42.0.1:24304: remote error: tls: bad certificate │ │ 2024/02/21 21:41:51 http: TLS handshake error from 10.42.0.1:35613: remote error: tls: bad certificate │ │ 2024/02/21 21:41:51 http: TLS handshake error from 10.42.0.1:64287: remote error: tls: bad certificate │ │ 2024/02/21 21:42:02 http: TLS handshake error from 10.42.0.1:29383: remote error: tls: bad certificate │ │ 2024/02/21 21:42:17 http: TLS handshake error from 10.42.0.1:65440: remote error: tls: bad certificate Would you like to support us? [ ] Yes, I would like to support you AB#11244 even uninstallation is giving me issues: rad uninstall kubernetes Uninstalling Radius... Uninstalling Contour from namespace: radius-system E0221 13:52:04.447147 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:04.679729 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.229982 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.258676 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.611919 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.752075 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:06.253108 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:06.567221 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:07.464444 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request @ytimocin This might be related to your change in this PR - https://github.com/radius-project/radius/pull/7022/files even uninstallation is giving me issues: rad uninstall kubernetes Uninstalling Radius... Uninstalling Contour from namespace: radius-system E0221 13:52:04.447147 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:04.679729 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.229982 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.258676 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.611919 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:05.752075 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:06.253108 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:06.567221 35329 memcache.go:121] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request E0221 13:52:07.464444 35329 memcache.go:287] couldn't get resource list for api.ucp.dev/v1alpha3: the server is currently unable to handle the request @willtsai try to delete ucp-cert and reinstall radius. kubectl delete secret -n radius-system ucp-cert Turns out this behavior was expected as a part of v0.30 release -- when upgrading to v0.30 you must first uninstall and then install Radius on your cluster: rad uninstall kubernetes then rad install kubernetes
gharchive/issue
2024-02-21T21:44:45
2025-04-01T06:45:34.197664
{ "authors": [ "willtsai", "youngbupark" ], "repo": "radius-project/radius", "url": "https://github.com/radius-project/radius/issues/7187", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1821225874
🛑 GPS Receiver API is down In 961dc8b, GPS Receiver API (http://tracker.ace-energy.co.th/receiver/api/v1/health) was down: HTTP code: 504 Response time: 15037 ms Resolved: GPS Receiver API is back up in 142fdca.
gharchive/issue
2023-07-25T21:56:12
2025-04-01T06:45:34.200856
{ "authors": [ "chindanai" ], "repo": "radiuszon/upptime", "url": "https://github.com/radiuszon/upptime/issues/1789", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1882131863
feat(breakpoints): added css variables for breakpoints Added css variables for breakpoints to have flexibility on styling level. Example Scenarios As a developer, I can use radix UI breakpoints for styling components or pages on CSS level As a developer, I can change CSS variables to change breakpoint to fit on design requests. Hey @Sly777, did you test this? I don't think using custom properties in @media declarations is valid CSS, is it? Yes, very sorry @benoitgrelard - I thought first that works properly after build and dev didn't give me any error.
gharchive/pull-request
2023-09-05T14:45:54
2025-04-01T06:45:34.202943
{ "authors": [ "Sly777", "benoitgrelard" ], "repo": "radix-ui/themes", "url": "https://github.com/radix-ui/themes/pull/62", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
99205679
Add more functional code Do not merge yet Hey, this PR reimplements idle.js with a more functional code. I didn't get rid of prototype yet, but I was able to minimize impure and stateful functions to bare minimum. When I have more time I can go through the code and explain to you the design decisions. To test it, I created a examples/simple.js with this content: var Idle = require('../index'); var idle = new Idle({ preset: 'bullets', moonwalk: true }); idle.start(); @hugobessaa Aw man, thanks for this! As you know, I just started learning FP and to have a comparison like this to guide me through is really helpful. Not sure if I'm ever going to merge it though since this library is very simple and already does pretty much what I want, but I will keep this for sure! I just want to show my functional love. Don't bother to merge :smile: Some highlights: Prototype methods simply manage our pure and stateless helpers Any function there can be exported and tested without creating an instance of idle. Future improvements: Provide custom write function to render, making it easy to test it Oh, and look: not a single line were added. @hugobessaa You might wanna check the latest release which was implemented using a more functional approach :+1: Looks great!
gharchive/pull-request
2015-08-05T13:33:38
2025-04-01T06:45:34.220569
{ "authors": [ "hugobessaa", "rafaelrinaldi" ], "repo": "rafaelrinaldi/idle.js", "url": "https://github.com/rafaelrinaldi/idle.js/pull/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1708753198
Incremental compilation warnings in 1.9 When I load Rasters in julia 1.9 I get the following wall of Warnings: It seems that these warnings are all from the plotrecipes.jl file. I don't know the Plots recipes well enough, but from a user perspective this huge list of warnings looks bad. should we put the Plots recipes into a weak dependency? 18 dependencies successfully precompiled in 28 seconds. 100 already precompiled. 2 dependencies had warnings during precompilation: ┌ RasterDataSources [3cb90ccd-e1b6-4867-9617-4276c8b2ca36] │ ┌ Warning: Replacing docs for `RasterDataSources.getraster :: Union{}` in module `RasterDataSources` │ └ @ Base.Docs docs/Docs.jl:243 └ ┌ Rasters [a3a2b9e3-a471-40c9-b274-f788e487c689] │ WARNING: Method definition _prepare(DimensionalData.Dimensions.Dimension{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:193 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:304. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _prepare(Rasters.AbstractRaster{T, N, D, A} where A where D where N where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:195 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:306. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _subsample(Any, Any) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:200 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:311. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybename(Any) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:216 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:327. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybename(DimensionalData.Name{N}) where {N} in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:217 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:328. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybename(DimensionalData.NoName) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:218 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:329. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybename(Symbol) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:219 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:330. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybename(AbstractString) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:220 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:331. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_replace_missing(AbstractArray{var"#s167", N} where N where var"#s167"<:AbstractFloat) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:222 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:333. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_replace_missing(Any) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:223 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:334. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_shift(Any) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:225 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:336. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_shift(DimensionalData.Dimensions.LookupArrays.Intervals{L} where L, Any) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:226 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:337. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_shift(Any, Any) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:227 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:338. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_mapped(Tuple) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:229 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:340. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_mapped(DimensionalData.Dimensions.Dimension{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:230 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:341. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_mapped(DimensionalData.Dimensions.LookupArrays.LookupArray{T, N} where N where T, DimensionalData.Dimensions.Dimension{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:231 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:342. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_mapped(Rasters.Projected{T, A, O, Sp, Sa, MD, PC, MC, D} where D where MC where PC where MD where Sa<:DimensionalData.Dimensions.LookupArrays.Sampling where Sp<:DimensionalData.Dimensions.LookupArrays.Span where O<:DimensionalData.Dimensions.LookupArrays.Order where A<:AbstractArray{T, 1} where T, DimensionalData.Dimensions.Dimension{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:232 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:343. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_mapped(Nothing, DimensionalData.Dimensions.Dimension{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:233 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:344. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition _maybe_mapped(GeoFormatTypes.GeoFormat, DimensionalData.Dimensions.Dimension{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:234 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:345. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition refdims_title(Rasters.Band{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:238 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:349. │ ** incremental compilation may be fatally broken for this module ** │ │ WARNING: Method definition kwcall(Any, typeof(DimensionalData.refdims_title), Rasters.Band{T} where T) in module Rasters at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:238 overwritten at /home/fcremer/.julia/packages/Rasters/ymsa7/src/plotrecipes.jl:349. │ ** incremental compilation may be fatally broken for this module ** Yeah I think anshul copied some meyjods for Makie plots that I missed. Theres a branch fixing it but somehow tests failed
gharchive/issue
2023-05-13T22:11:40
2025-04-01T06:45:34.226377
{ "authors": [ "felixcremer", "rafaqz" ], "repo": "rafaqz/Rasters.jl", "url": "https://github.com/rafaqz/Rasters.jl/issues/445", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
629691285
Issue 17 Aiding to address issue-17 as well formatting. https://github.com/rahulg/treexml-rs/issues/17
gharchive/pull-request
2020-06-03T04:55:46
2025-04-01T06:45:34.242393
{ "authors": [ "mlevkov" ], "repo": "rahulg/treexml-rs", "url": "https://github.com/rahulg/treexml-rs/pull/18", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1187462234
Add support for generating a PDF invoice A client or the provider should be able to generate a pdf once the invoice is completed (for tax purposes) We may need input such as name, address, etc to generate this pdf. We can mark the pdf as "automatically generated". Not sure about the legal side of this, any support and/or insight would be appreciated! May need to include inputs for address, phone, etc. Estimated at $1000 (70% confidence)
gharchive/issue
2022-03-31T04:00:36
2025-04-01T06:45:34.247589
{ "authors": [ "dan13ram", "plor" ], "repo": "raid-guild/smart-invoice", "url": "https://github.com/raid-guild/smart-invoice/issues/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
662699351
Formatting of channel settlement notification and message Description The channel notification and message text need to get formatted nicely. We should also check, if the text is understandable. There are a lot of addresses within the notification. Acceptance criteria Addresses are truncated Icon is shown Tasks [ ] Meeting with @sashseurat @MariaNono and @taleldayekh Design of the notification panel a) Text for Sash b) It is for slow blockchain interactions c) We currently have Settle Channel Notifications d) Next one is going to be open channel e) Icons: In progress, Done, Failed f) Take the bin instead of the x to get rid of notifications g) The blocks are incrementing (just numbers) Design of the notification overlay a) Taking the same design as the notifications in the panel mockup Visual progress bar a) Making a separate issue to split this @christianbrb + @taleldayekh Here is the click-dummy for the notifications panel ready for your review: https://www.figma.com/proto/nCvDyCfu5ZXnNETzepPLwg/LC-Notifications-Panel?node-id=179%3A11&viewport=923%2C270%2C0.884325385093689&scaling=contain I've added the open channel icon, channel settled icon, and failed icon and would like your feedback on them as well. @sashseurat Crazy stuff, looks super nice 👍 :) @taleldayekh Please also take a look at https://github.com/raiden-network/light-client/issues/1902#issuecomment-677557195 when implementing this issue. I would be in favor using the clickable snack bar at footer instead of the notification overlay on the top. I would be in favor using the clickable snack bar at footer instead of the notification overlay on the top. Understood and I agree. 👍
gharchive/issue
2020-07-21T07:37:27
2025-04-01T06:45:34.253959
{ "authors": [ "christianbrb", "sashseurat", "taleldayekh" ], "repo": "raiden-network/light-client", "url": "https://github.com/raiden-network/light-client/issues/1929", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
276634604
dynamic price for a resource there's a bug where the price header isn't generated correctly if price is a callable object https://github.com/raiden-network/microraiden/blob/master/microraiden/microraiden/proxy/resources/expensive.py#L176 Fixed by PR 202
gharchive/issue
2017-11-24T14:38:22
2025-04-01T06:45:34.255357
{ "authors": [ "pcppcp" ], "repo": "raiden-network/microraiden", "url": "https://github.com/raiden-network/microraiden/issues/174", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1299599606
Tailwind requires CSS rebuild after every HTML modification. (bin/dev, hotreload gems, etc., not effective) Hi all, This one has stumped both myself and (more importantly) those with far more experience and knowledge of the Rails framework and CSS engines alike. I've attached a brief 2.5 minute video below demonstrating the issue. In effect, I can change each class on an element one time, but if I try to change two or more times, I must manually rebuild the CSS. For example, I can change a class to text-4xl from its current state of text-2xl, and it will work. But if I go to then change it to text-sm or anything other than text-2xl or text-4xl, it will effectively remove the entire class and will default to black. Same with mt classes. A perfect example is changing the default mt-28 property in application.html.erb. If I set it to, say, mt-10, that and mt-28 are the only two classes that will apply. If I tried to set any other mt level, it would fail and effectively be mt-0. I'm currently running WSL2 Ubuntu 20.04, Rails v7.0.3, Ruby v3.1.2 -- all stable releases. I use importmaps, so no package.json to speak of. I'm now a novice+ programmer, but know my way around a troubleshoot as I was tuned-in about a decade ago. I'm open to trying any and everything in order to solve this and document the solution -- it's the first thing I've seen stump everyone in the GoRails community's CSS channel. Thanks in advance, everyone! RM Link: https://streamable.com/kx0zae Your server might be serving your manually built css instead of the rebuilt bundle created by tailwindcss cli. Try running rake assets:clobber and see if that solves the issue. @robertemaxwell There's a PR submitted to make the "watch" mechanism support WSL. https://github.com/rails/tailwindcss-rails/pull/168 Does that sound like it might address this issue? See #199 for a PR which I think will resolve your issue.
gharchive/issue
2022-07-09T07:45:12
2025-04-01T06:45:34.987345
{ "authors": [ "flavorjones", "hyphenized", "robertemaxwell" ], "repo": "rails/tailwindcss-rails", "url": "https://github.com/rails/tailwindcss-rails/issues/183", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1354346372
add zlib nix libraries if pandas is detected Previously we were just adding the zlib library if numpy was detected. This PR changes it so we also add them if only pandas is detected. Fixes https://github.com/railwayapp/nixpacks/issues/462 This won't actually solve the linked issue though correct? Cause we need libstdc++ which requires LD_LIBRARY updates LD_LIBRARY is already updated with Nix libraries that the plan specifies. zlib is the Nix package needed.
gharchive/pull-request
2022-08-29T13:54:16
2025-04-01T06:45:35.008984
{ "authors": [ "coffee-cup" ], "repo": "railwayapp/nixpacks", "url": "https://github.com/railwayapp/nixpacks/pull/463", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1258254282
Task: Good defaults for artifact storage We shouldn't require much ceremony around specifying a directory to store tarballs and so on; it should Just Work. Ideally it would be shared across projects so multiple projects building an image on the same base wouldn't redownload it a bajillion times. But that conflicts a bit with MSBuild's model around intermediate files, so we'll have to design a solution. The solution is probably just "take the defaults from #6". We'll need this caching ability before our initial previews, so adding this to the preview7 milestone. This is done!
gharchive/issue
2022-06-02T14:20:13
2025-04-01T06:45:35.013394
{ "authors": [ "baronfel", "rainersigwald" ], "repo": "rainersigwald/containers", "url": "https://github.com/rainersigwald/containers/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
179020895
Installation failed (composer) & scraping from a list Hi, I keep getting this failed message when trying to install. MacBook-Pro-3:Insta mac1$ git clone https://github.com/raiym/instagram-php-scraper.git InstaPHP Cloning into 'InstaPHP'... remote: Counting objects: 561, done. remote: Total 561 (delta 0), reused 0 (delta 0), pack-reused 561 Receiving objects: 100% (561/561), 71.78 KiB | 56.00 KiB/s, done. Resolving deltas: 100% (335/335), done. Checking connectivity... done. MacBook-Pro-3:Insta mac1$ cd InstaPHP/ MacBook-Pro-3:InstaPHP mac1$ ls CHANGELOG README.md composer.json index.php src tests MacBook-Pro-3:InstaPHP mac1$ composer require raiym/instagram-php-scraper Using version ^0.3.4 for raiym/instagram-php-scraper ./composer.json has been updated Loading composer repositories with package information Updating dependencies (including require-dev) Your requirements could not be resolved to an installable set of packages. Problem 1 - Installation request for raiym/instagram-php-scraper dev-master -> satisfiable by raiym/instagram-php-scraper[dev-master]. - Can only install one of: raiym/instagram-php-scraper[v0.3.4, dev-master]. - Installation request for raiym/instagram-php-scraper ^0.3.4 -> satisfiable by raiym/instagram-php-scraper[v0.3.4]. Installation failed, reverting ./composer.json to its original content. Have I missed a crucial step? I followed your guide. As a separate comment, I would like to give this library a long list of Media Code's (or ID's) and get it to scrape comments. Can this library do it? Thanks Hello, you don't need to clone it if you use composer
gharchive/issue
2016-09-24T09:20:43
2025-04-01T06:45:35.046984
{ "authors": [ "raiym", "yarnball" ], "repo": "raiym/instagram-php-scraper", "url": "https://github.com/raiym/instagram-php-scraper/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1402089155
Create Todolist.html Simple Todo list using HTML, JS, CSS good !!!
gharchive/pull-request
2022-10-08T22:54:49
2025-04-01T06:45:35.047872
{ "authors": [ "RajMaity", "raj-ravan" ], "repo": "raj-ravan/Hacktoberfest2022", "url": "https://github.com/raj-ravan/Hacktoberfest2022/pull/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
908102988
Swipe Gesture doesn't work Idk if it's just my phone but swipe gesture doesn't work when I'm reading, I need to press > button in order for me to read the next chapter so hopefully this will be fix soon https://streamable.com/9xlnzj also it would be nice if you add horizontal/vertical swipe gesture option in reader bottom sheet similar to this one : Vertical swipe gesture to read the next chapter is not implemented. Vertical/Horizontal swipe gesture options to switch between vertical and horizontal reading mode? Oh okay. I thought my phone is having a problem xd Yep that's right I have added swipe gestures to navigate between chapters. I will close this issue as there is another one about horizontal reading mode.
gharchive/issue
2021-06-01T09:36:01
2025-04-01T06:45:35.052978
{ "authors": [ "rajarsheechatterjee", "seizue" ], "repo": "rajarsheechatterjee/lnreader", "url": "https://github.com/rajarsheechatterjee/lnreader/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
195064783
Feature: Add functionality of changing the theme Or is any other way to change the theme color of datepicker? You can use the classes mentioned in the documentation to change the color theme.
gharchive/issue
2016-12-12T19:35:39
2025-04-01T06:45:35.053841
{ "authors": [ "hemkaran", "rajeshwarpatlolla" ], "repo": "rajeshwarpatlolla/ionic-datepicker", "url": "https://github.com/rajeshwarpatlolla/ionic-datepicker/issues/287", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
102550403
ios is selected not work ios is selected not work Could you please explain briefly. android is working,but ios is bug。when me selected date is not work,top blue date nothing。click set again click start is disable。 Still i didn't get your point. Please test this in your device rajeshwarpatlolla.github.io/DatePickerForIonicFramework/demo/#/tab/date-picker Please use the same syntax as i mentioned in the description.
gharchive/issue
2015-08-22T16:49:32
2025-04-01T06:45:35.055806
{ "authors": [ "match08", "rajeshwarpatlolla" ], "repo": "rajeshwarpatlolla/ionic-datepicker", "url": "https://github.com/rajeshwarpatlolla/ionic-datepicker/issues/61", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1187611723
stats div の内部ではなくその直前に diff-stars span を挿入するように修正 Fixed #1 Summary 統計情報表示枠の溢れによる表示の崩れを修正 Changes starSpan のスタイルを修正 starSpan の挿入箇所を stats の内部から stats の直前に変更 修正後のスクリーンショット 大幅表示 中幅表示 小幅表示 マージさせていただきます。 ありがとうございます!
gharchive/pull-request
2022-03-31T07:04:25
2025-04-01T06:45:35.061016
{ "authors": [ "rakkyo150", "yatakabs" ], "repo": "rakkyo150/PredictStarNumberExtension", "url": "https://github.com/rakkyo150/PredictStarNumberExtension/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
205921442
Same background on menu for suggestions Hi, it is hard in vim-one to see the menu for the autocompletion suggestions. Dark Light Other Colorscheme Maybe we change the background of the menu. Here is how it looks like now dark light @rakr thx, looks good
gharchive/issue
2017-02-07T15:27:41
2025-04-01T06:45:35.064794
{ "authors": [ "DenniJensen", "rakr" ], "repo": "rakr/vim-one", "url": "https://github.com/rakr/vim-one/issues/32", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
475966280
Added new feature to filter by file extentions This new feature can filter files. $ statik -src=./public -only-include=*.html,*.js,*.css This issue is related to https://github.com/rakyll/statik/issues/42 I would suggest that we first discuss the feature and then we jump into the implementation. Nice job. I left a couple of comments. I am happy with this. What do you think @rakyll? What about something sorter like -filter? What about something sorter like -filter? @rakyll would you describe the API for it? Sorry maybe I was nitpicking on the flag name. What about something shorter like: -include=*.html,*.js,*.css @schivei I left one comment. We are almost there, thanks for the great work! I added the execution of the “include” option directly to the part where the files are searched, so if “include” is not passed, it must have the default expression for any “.” file. Otherwise, if it is empty, it will not include any files and will compromise the default functionality of statik. I think we are good to go. Let's wait until tomorrow morning (in case any though comes to our minds or someone else raise a concern) and then I will merge. Thanks @schivei!
gharchive/pull-request
2019-08-02T02:25:19
2025-04-01T06:45:35.089945
{ "authors": [ "jcchavezs", "rakyll", "schivei" ], "repo": "rakyll/statik", "url": "https://github.com/rakyll/statik/pull/78", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1718133279
Requirements Status link is broken? Hey, when clicking this link: It seems to be broken: fixing now. Looks like travis-ci now requires a credit card. I'll remove it.
gharchive/issue
2023-05-20T10:47:28
2025-04-01T06:45:35.097264
{ "authors": [ "Tamir198", "ralyodio" ], "repo": "ralyodio/humanparser", "url": "https://github.com/ralyodio/humanparser/issues/55", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
141858618
Remove adding module name in module_group_path if it set from cli When I generate module using this command: generamba gen MyModule module --module-group-path projectName/Scenarios Generamba adds MyModule folder in projectName/Scenarios by default. But in my opinion if I change destination with module-group-path option it should add files in same path, that I set in that option So, instead projectName/Scenarios/MyModule it should be projectName/Scenarios @IvanUshakov Thanks for your suggestion, seems reasonable. After discussion with @Beniamiiin we decided that current behavior is more predictable. If some more people will request for this feature, we'll revisit to it.
gharchive/issue
2016-03-18T12:56:59
2025-04-01T06:45:35.100308
{ "authors": [ "IvanUshakov", "etolstoy" ], "repo": "rambler-ios/Generamba", "url": "https://github.com/rambler-ios/Generamba/issues/98", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
114585895
R.pluck behaves strangely for empty lists According to the docs: Returns a new list by plucking the same named property off all objects in the list supplied. I am inferring from this statement that any pluck on an empty list would also return an empty list, but that is not the case. > var R = require('ramda') undefined > R.pluck([], 'foo') [ undefined, undefined, undefined ] Either the documentation should be clarified (not sure of the best way to do this) or the behaviour should be changed. Arguments the wrong way around. Ignore me...
gharchive/issue
2015-11-02T12:37:23
2025-04-01T06:45:35.102000
{ "authors": [ "Willyham" ], "repo": "ramda/ramda", "url": "https://github.com/ramda/ramda/issues/1491", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
548222430
why interval couldn't be '1m' Hi Ran, Thank you very much for the APT, while when I try to get the data, for example: msft = yf.Ticker("MSFT") get historical market data, here max is 5 years. msft.history(start = '2015-01-01', end = '2017-12-31', interval = '1m') I couldn't get it and I got the message: MSFT: 1m data not available for startTime=1420099200 and endTime=1514707200. Only 7 days worth of 1m granularity data is allowed to be fetched per request. I would like to get the monthly price from 2015-01-01 to 2018-12-31, could you be so kind as to give me some advice? The interval you used is for 1 minute. Use ‘1mo’ for one month. Thanks Doug! On Mon, Jan 13, 2020 at 9:38 PM Doug notifications@github.com wrote: The interval you used is for 1 minute. Use ‘1mo’ for one month. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/ranaroussi/yfinance/issues/165?email_source=notifications&email_token=AJB2444KUZGTGGEI5GHRB4DQ5VFUVA5CNFSM4KFLQNM2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEI3LJ6I#issuecomment-574010617, or unsubscribe https://github.com/notifications/unsubscribe-auth/AJB24462PBJQXXBAQNL36Q3Q5VFUVANCNFSM4KFLQNMQ . Closed as problem is solved
gharchive/issue
2020-01-10T18:18:12
2025-04-01T06:45:35.115587
{ "authors": [ "JaneLiu0617", "dpguthrie", "silvavn" ], "repo": "ranaroussi/yfinance", "url": "https://github.com/ranaroussi/yfinance/issues/165", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1025502679
When downloading historical data between date range, gives incorrect dates I try running the following code: def day_data(): data = yf.download(tickers='NVDA',interval='1d',start='2021-06-08',end='2021-06-11') return data print(day_data()) However, the output I get is as follows: 2021-06-07 175.712494 178.125000 ... 176.118576 57575600 2021-06-08 175.272507 176.214996 ... 174.499237 32384800 2021-06-09 175.157501 175.750000 ... 173.551895 38165600 I double-checked, and the markets were open on those days, and I entered June 8th as a start, but it's giving me data for June 7th. You probably have a problem with the timezone you are in (sometimes yahoo servers have diff timezones) or your computer time. I just ran the following (changed end date to saturday the 12th): import yfinance as yf def day_data(): data = yf.download(tickers='NVDA', interval='1d', start='2021-06-08', end='2021-06-12') return data print(day_data()) and the result was: [*********************100%***********************] 1 of 1 completed Open High ... Adj Close Volume Date ... 2021-06-08 175.272507 176.214996 ... 174.499237 32384800 2021-06-09 175.157501 175.750000 ... 173.551895 38165600 2021-06-10 173.500000 174.919998 ... 174.219284 28777200 2021-06-11 174.794998 179.392502 ... 178.221069 41630800 [4 rows x 6 columns] [Finished in 1.7s]
gharchive/issue
2021-10-13T17:13:36
2025-04-01T06:45:35.119032
{ "authors": [ "mistersenpai", "silvavn" ], "repo": "ranaroussi/yfinance", "url": "https://github.com/ranaroussi/yfinance/issues/853", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
216887235
WIP several fixes around stdandalone container upgrade Previous revision to instance TBD - @joshwget this will be the fix for the issue you saw when revision id wasn't returned to you on the instance.upgrade. Marking WIP till we figure the issue out. Addressed as a part of https://github.com/rancher/cattle/pull/2560, closing
gharchive/pull-request
2017-03-24T19:45:06
2025-04-01T06:45:35.134980
{ "authors": [ "alena1108" ], "repo": "rancher/cattle", "url": "https://github.com/rancher/cattle/pull/2577", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
96968021
Support x-forwarded-host Do not merge until pom maven dep is updated to non-snapshot release. LGTM
gharchive/pull-request
2015-07-24T05:18:16
2025-04-01T06:45:35.135791
{ "authors": [ "alena1108", "cjellick" ], "repo": "rancher/cattle", "url": "https://github.com/rancher/cattle/pull/717", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1886770364
Error: Cannot read properties of undefined (reading 'find')seen after Scaling down nodes in a nodepool Setup Rancher version: 2.7-head commit id: 787c056 Browser type & version: Chrome Describe the bug Error: Cannot read properties of undefined (reading 'find')seen after Scaling down nodes in a nodepool To Reproduce Deploy a node driver k3s cluster - 1 node pool all roles, 1 nodepool - 2 CP nodes, 1 nodepool with 2 worker nodes Once the cluster is Active, select one CP only node and click on Scale Down Select the other CP only node and click on Scale Down The page then loads and shows Cannot read properties of undefined (reading 'find') Expected Result No error should be seen. Console output: Cannot read properties of undefined (reading 'find') e.forEach.console.<computed> | @ | console.js:31 -- | -- | --   | (anonymous) | @ | client.js:111   | p | @ | regeneratorRuntime.js:44   | (anonymous) | @ | regeneratorRuntime.js:125   | (anonymous) | @ | regeneratorRuntime.js:69   | r | @ | asyncToGenerator.js:3   | s | @ | asyncToGenerator.js:22   | (anonymous) | @ | asyncToGenerator.js:27   | (anonymous) | @ | asyncToGenerator.js:19   | (anonymous) | @ | client.js:113   | Hn | @ | vue.runtime.esm.js:3031   | Bn | @ | vue.runtime.esm.js:3008   | Tn.e._render | @ | vue.runtime.esm.js:2687   | r | @ | vue.runtime.esm.js:3875   | e.get | @ | vue.runtime.esm.js:3446   | e.run | @ | vue.runtime.esm.js:3522   | Jr | @ | vue.runtime.esm.js:4121   | (anonymous) | @ | vue.runtime.esm.js:3143   | Kn | @ | vue.runtime.esm.js:3065   | Promise.then (async) |   |     | qn | @ | vue.runtime.esm.js:3090   | Jn | @ | vue.runtime.esm.js:3155   | ni | @ | vue.runtime.esm.js:4207   | e.update | @ | vue.runtime.esm.js:3513   | e.notify | @ | vue.runtime.esm.js:720   | set | @ | vue.runtime.esm.js:963   | Mi.Fi.set | @ | vue.runtime.esm.js:5356   | setBulkActionOfInterest | @ | index.vue:799   | mouseover | @ | index.vue:1   | Un | @ | vue.runtime.esm.js:3017   | n | @ | vue.runtime.esm.js:1815   | La.o._wrapper | @ | vue.runtime.esm.js:7473 Ticket #9665 - Test Results - ✅ PASS Verified with a Docker install of Rancher v2.8-head: Scenario 1: Validate that no error page is displayed Deploy a node driver k3s cluster - 1 node pool all roles, 1 nodepool - 2 CP nodes, 1 nodepool with 2 worker nodes Once the cluster is Active, select one CP only node and click on Scale Down Select the other CP only node and click on Scale Down Results: Scale down works properly and no undefined page shows up. /backport v2.7.next1
gharchive/issue
2023-09-08T00:40:37
2025-04-01T06:45:35.141922
{ "authors": [ "gaktive", "sowmyav27", "susesgartner" ], "repo": "rancher/dashboard", "url": "https://github.com/rancher/dashboard/issues/9665", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
557823103
K3s needs to supply Rancher UI with provider field For integrating into Rancher, the UI expects the cluster to provide the "provider" field to indicate the type of cluster. This is so the UI can show the cluster as an imported K3s cluster rather than a generic import. This appears to already be working today. The Rancher UI will show the cluster is a K3s cluster.
gharchive/issue
2020-01-30T23:40:24
2025-04-01T06:45:35.149674
{ "authors": [ "davidnuzik" ], "repo": "rancher/k3s", "url": "https://github.com/rancher/k3s/issues/1368", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
278873449
RancherUI containers tab: displayed container's IP is not refreshed after its network has been changed I moved user containers to a new network (connected to the new then disconnected from the old in CLI), however RancherUI still shows the old IP address (from the range of the old network that has already been deleted) even after server reboot. When checking the IP address of that container in CLI (docker inspect) I can see only the new network's IP address, range, gw. Cannot find any reference to the old IP, network anywhere in the inspect result. Wondering where the RancherUI gathers this info from? Would be great if it were refreshing these info when opening the containers tab on the UI. Useful Info Versions Rancher v1.6.12 Cattle: v0.183.30 UI: v1.6.30 Access Disabled Orchestration Cattle Route containers.index What does moving user containers to a new network mean? Are we talking about our own overlay network or you using Docker libnetwork network(s)? Closing as stale.
gharchive/issue
2017-12-04T05:18:39
2025-04-01T06:45:35.177821
{ "authors": [ "superseb", "vasmancs" ], "repo": "rancher/rancher", "url": "https://github.com/rancher/rancher/issues/10467", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
301030456
server redirect to wrong port for https Rancher versions: Rancher versions: 2.0 master 2/28 Steps to Reproduce: run command docker run -d --restart=unless-stopped -p 9090:80 -p 8443:443 rancher/server:master visit http://127.0.0.1:9090/ in the browser. Results: It redirects to https://127.0.0.1:9443, which is the wrong port. The correct rediect url is https://127.0.0.1:8443 There is no way for the server to know what port you mapped 443 to and redirect to it. Realistically the redirect is only useful if using the standard ports, nobody is going to seek out a non-standard http port just to be redirected to the https one.
gharchive/issue
2018-02-28T13:49:54
2025-04-01T06:45:35.181241
{ "authors": [ "vincent99", "zionwu" ], "repo": "rancher/rancher", "url": "https://github.com/rancher/rancher/issues/11738", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
331000856
Rancher catalog install not setting built-in helm variables like .Release.IsInstall Rancher versions: rancher/server or rancher/rancher: v2.0.2 rancher/agent or rancher/rancher-agent: Operating system and kernel: (cat /etc/os-release, uname -r preferred) NAME="Ubuntu" VERSION="16.04.4 LTS (Xenial Xerus)" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 16.04.4 LTS" VERSION_ID="16.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" VERSION_CODENAME=xenial UBUNTU_CODENAME=xenial Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO) Custom OpenStack VM Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB) single node rancher Environment Template: (Cattle/Kubernetes/Swarm/Mesos) Kubernetes Steps to Reproduce: Turn on helm stable catalog under global settings Install keycloak (create an answer with key=keycloak.password and value=test) Create ingress that maps / to keycloak on port 8080 Results: The keycloak helm tests are run. These should not be run when installing. The password is not set. This is because .Release.IsInstall does not appear to be set. The Rancher catalogue does not appear to be doing a standard helm install. Installing manually with helm install works fine. hi @nuwang Can you try it with Rancher v2.1.0, please? In 2.1.0, we are using Tiller.
gharchive/issue
2018-06-10T20:41:56
2025-04-01T06:45:35.186196
{ "authors": [ "loganhz", "nuwang" ], "repo": "rancher/rancher", "url": "https://github.com/rancher/rancher/issues/13934", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }