added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:35:20.571851
2020-10-24T16:27:20
728824233
{ "authors": [ "FloVanGH", "sandmor" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10268", "repo": "redox-os/orbtk", "url": "https://github.com/redox-os/orbtk/pull/383" }
gharchive/pull-request
Support for directions in gradients Context: This adds support for directions in gradients, so you can specific direction of the next manner: linear-gradient(to {top|top right|right|bottom right|bottom|bottom left|left|top left}, ...) I prefer to not attempt to document it directly because there is a similar PR in progress That's great exactly what I need 🙂. Thanks
2025-04-01T04:35:20.612548
2016-03-23T14:50:22
142980949
{ "authors": [ "mfburnett", "rata" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10269", "repo": "redspread/spread", "url": "https://github.com/redspread/spread/issues/111" }
gharchive/issue
How to handle host volumes in localkube Hi, The idea of this issue is discuss how to handle the host volumes used only for local development that, also, the absolute path changes across developers and all need to do local development. I haven't really thought about it, but maybe using a file with all the containers for production, and the one for local dev just adds the volume mounts and a "cat prod dev-volumes > dev" is done and dev is always kept up to date. Probably this is not a good idea and something better can be done, but just to start the discussion :) hey @rata, so sorry we missed this issue! since we moved work on localkube over to minikube, I believe dealing with host volumes is on their roadmap, and this issue (https://github.com/kubernetes/minikube/issues/422) has discussion on it! On Fri, Jul 29, 2016 at 05:14:29PM -0700, Mackenzie Burnett wrote: hey @rata, so sorry we missed this issue! since we moved work on localkube over to minikube, I believe dealing with host volumes is on their roadmap, and this issue (https://github.com/kubernetes/minikube/issues/422) has discussion on it! Np! I thought it was on minikube now, no that this issue was missed! :-)
2025-04-01T04:35:20.637634
2022-04-09T18:20:59
1198735097
{ "authors": [ "redwarp", "russellbanks" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10270", "repo": "redwarp/gif-wallpaper", "url": "https://github.com/redwarp/gif-wallpaper/issues/213" }
gharchive/issue
[Bug] Rare glitching of gif I downloaded this app a few days ago and it was perfectly fine. After some time, I had severe glitching of the gif and I don't know why. At the time, I was testing the reading mode on my phone (which either dampens the screen's colours or makes them grayscale) but I don't see how this could have caused it. To fix it, I went into the gif wallpaper app, removed the gif, and then re-added it, and I haven't had a problem since nor have I been able to reproduce it. I thought I would document this nontheless to see if there was anything that could be fixed/changed as a result of it. I took a screen recording of it when I noticed it was happening, as below. https://user-images.githubusercontent.com/74878137/162586476-8c2b75cf-a93b-4344-83e0-f2d35d2ffa06.mp4 You are not the first one to witness this! A friend of mine had that as well. It produces fascinating results but for sure it sucks :-D But I have not been able to consistently reproduce it, so I have no clue how to fix it for now. I'm still not sure how to reproduce, closing this
2025-04-01T04:35:20.685582
2022-06-13T16:16:27
1269669127
{ "authors": [ "jbargu", "turinglabsorg" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10271", "repo": "reef-defi/evm-provider.js", "url": "https://github.com/reef-defi/evm-provider.js/pull/20" }
gharchive/pull-request
fix main and types pointer This fix solves this annoying warning that appears anytime the script runs: Of course this fix should be also sent to npm to update public package.. Hope it helps! Hey, thanks for the heads up. It used to be like this but it was causing issue with linkage. I assume the warning is happening due to Node v16. For the moment we will leave it as is.
2025-04-01T04:35:20.703186
2022-09-02T20:15:42
1360577086
{ "authors": [ "fregante", "janpio" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10272", "repo": "refined-github/refined-github", "url": "https://github.com/refined-github/refined-github/issues/5940" }
gharchive/issue
set-default-repositories-type-to-sources: Failed to execute 'observe' on 'MutationObserver': parameter 1 is not of type 'Node'. Description New GH Projects pages trigger this error: TypeError: Failed to execute 'observe' on 'MutationObserver': parameter 1 is not of type 'Node'. at chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:7536:13 at new Promise (<anonymous>) at oneMutation (chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:7532:14) at chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:7587:51 at async chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:7585:9 at async runFeature (chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:3851:20) at async setupPageLoad (chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:3857:34) at async chrome-extension://hlepfoohegkhhmjieoechaddaejaokhf/refined-github.js:3882:29 How to replicate the issue Just open https://github.com/users/janpio/projects/3/views/1 and observe Console Extension version 22.8.23 Browser(s) used Version 104.0.5112.81 (Official Build) (64-bit) Likely because there's an extra dropdown on the page
2025-04-01T04:35:20.707293
2024-10-30T14:04:33
2624222779
{ "authors": [ "fregante", "kovsu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10273", "repo": "refined-github/shorten-repo-url", "url": "https://github.com/refined-github/shorten-repo-url/pull/64" }
gharchive/pull-request
Fix pathname is empty Test URL: https://github.com/download-directory/download-directory.github.io/issues/143 As you can see if url is https://github.com/, shortenRepoUrl function runs to end. Because these properties is empty, so it returns an empty string and makes the link is swallowed, https://github.com/refined-github/shorten-repo-url/blob/ce4fcc4b5632cf24523ed11e73da5ff876466b36/index.js#L274-L276 If we get the pathname is /, it goes to here. https://github.com/refined-github/shorten-repo-url/blob/ce4fcc4b5632cf24523ed11e73da5ff876466b36/index.js#L159-L178 so we will get a cleanHref that is github.com @fregante Please review ~ @fregante Should I merge this? I think you forget this. I thought I commented 😅 yes you can. Also a new patch version can be released
2025-04-01T04:35:20.709887
2023-07-27T18:12:09
1824925716
{ "authors": [ "alicanerdurmaz", "siddiquiaffan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10274", "repo": "refinedev/refine", "url": "https://github.com/refinedev/refine/issues/4751" }
gharchive/issue
Add support for caching with Redis Is your feature request related to a problem? Please describe. Fetching data on each request increases the load on server as well as database. Using caching can help in reducing load. Describe alternatives you've considered No response Additional context No response Describe the thing to improve Add caching with Redis to reduce load on database and reduce final cost. Hi @siddiquiaffan, just to clarify, refine is a frontend framework, and caching on Redis is not something managed directly on the frontend. The responsibility for caching, including Redis caching, typically lies with the backend/API. Moreover, it's worth mentioning that refine already has built-in caching capabilities on the client side using TanStack Query.
2025-04-01T04:35:20.747624
2022-09-06T23:02:37
1363898077
{ "authors": [ "egorio", "haochuan" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10275", "repo": "rehearsal-js/rehearsal-js", "url": "https://github.com/rehearsal-js/rehearsal-js/pull/352" }
gharchive/pull-request
feat: new flag -r to generate report Just figured out I need to call Reporter.save to have the actual report, adding a new flag -r for generating report Renamed the previous option root to basePath The reporter technically can produce several different reports (json, serif) together. It's probably can be better to have something like -r=json, -r=json,serif... but can be upgraded later. I would also recommend to add -o to specify reports output directory (planning to add it upgrade command) @egorio Thanks for the detailed info and yeah all those makes sense. I'm not aware of the different report format before, will update this PR for that.
2025-04-01T04:35:20.750761
2018-12-12T22:17:02
390432926
{ "authors": [ "StarpTech", "wooorm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10276", "repo": "rehypejs/rehype-minify", "url": "https://github.com/rehypejs/rehype-minify/issues/26" }
gharchive/issue
minify-whitespace add template tag to callapsable list Input <h1> 404 <template v-if="resource"> the {{ resource }} was </template> not found </h1> Input <h1> 404 <template v-if="resource">the {{ resource }} was</template>not found </h1> This will lead to error 404 the resource wasnotfound because the template tag wasn't identified as a collapsable tag. Due to the fact that we didn't know the context where it is used we should handle it as a collapsable element. I close it. This is the case isn't valid according to HTML specs but for web frameworks. The template tag is a container of content which isn't parsed at runtime so it depends how the user will use it. Hmm, weird! This is Vue right? Hmm, they’re doing something funky and non-HTML with template tags. Doesn’t Vue also have fragments? Would a fragment (or a span) in this case make more sense? Yes, it's vue and in angular the equivalent is ng-container. A span wouldn't be the same since the user doesn't want to create a real an extra node. In my fork I added the elements to the lists because its no huge change and provide better compatibility for web frameworks.
2025-04-01T04:35:20.754199
2022-03-10T17:56:28
1165542930
{ "authors": [ "scala-steward" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10277", "repo": "reibitto/podpodge", "url": "https://github.com/reibitto/podpodge/pull/293" }
gharchive/pull-request
Update zio-config to 2.0.3 Updates dev.zio:zio-config from 1.0.10 to 2.0.3. GitHub Release Notes - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "dev.zio", artifactId = "zio-config" } ] labels: library-update, early-semver-major, semver-spec-major, commit-count:1 Superseded by #296.
2025-04-01T04:35:20.822508
2024-04-11T08:16:06
2237168819
{ "authors": [ "Diegocampy", "ThomDietrich" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10280", "repo": "rejuvenate/lovelace-horizon-card", "url": "https://github.com/rejuvenate/lovelace-horizon-card/issues/125" }
gharchive/issue
no translation into Italian of the moon phases but the English name remains followed by (!) I inserted the card and the English translation only works for the sun, not the moon. The moon integration however works perfectly if I look at it from the settings outside of this card Hello and thanks for reporting this issue! This is a bug introduced with HA 2024.04, already reported in https://github.com/rejuvenate/lovelace-horizon-card/issues/122 We are looking for willing developers to provide a bug fix PR.
2025-04-01T04:35:20.823902
2023-06-08T13:26:58
1747895912
{ "authors": [ "Vocaloid2048", "ikechan8370" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10281", "repo": "reko-beep/hsr-data", "url": "https://github.com/reko-beep/hsr-data/issues/10" }
gharchive/issue
O_VALUE Hi, this project is awesome and useful. Recently I have used master branch to download raw_data, but the data seems to become inaccurate these days (hash of some images became 404). After changing O_VALUE to "e2f24f9ed0117cf0db2e1ecf375f7e6a", the problems solved. Maybe the O_VALUE should be changed regularly? Sorry for asking a dummy question, is that there have any possible way to get O_VALUE by user/myself? It seems O_VALUE has changed again ...
2025-04-01T04:35:20.826011
2023-07-14T15:41:33
1805086278
{ "authors": [ "dhimmel" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10282", "repo": "related-sciences/nxontology-ml", "url": "https://github.com/related-sciences/nxontology-ml/pull/5" }
gharchive/pull-request
Legacy RS Classification for terms in EFO OTAR Slim v3.43.0 Add the RS Classification for terms in EFO OTAR Slim v3.43.0. This is a legacy classification but can be useful for comparison or training in the context of this repository. @yonromai I put this in a data directory, but feel free to suggest an alternative directory structure. Noting explanation of the label generating process at https://github.com/related-sciences/nxontology-ml/issues/13#issuecomment-1731267413
2025-04-01T04:35:20.829631
2020-02-21T03:01:39
568705961
{ "authors": [ "nilanjansiromani", "vio" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10283", "repo": "relative-ci/bundle-stats", "url": "https://github.com/relative-ci/bundle-stats/issues/626" }
gharchive/issue
TypeError: Cannot read property 'split' of undefined We get an error when trying to compare two jsons ➜ product git:(bundle-stats-comparison) ✗ bundbundle-stats --html ./bundleStats/baseline.json ./bundleStats/bundle-stats.json ✔ Read Webpack stats files ↓ Read baseline data [skipped] → Multiple stat files already set. ↓ Write baseline data [skipped] → Not a baseline job (see --baseline). ✖ Process data → Cannot read property 'split' of undefined Generate reports Save reports TypeError: Cannot read property 'split' of undefined at getAssetName (/Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/assets/get-asset-name.js:17:35) at /Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/transforms/assets-webpack-transform.js:49:45 at Array.reduce (<anonymous>) at assetsWebpackTransform (/Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/transforms/assets-webpack-transform.js:44:32) at createStats (/Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/stats/create.js:97:82) at createJob (/Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/jobs/create-job.js:53:41) at /Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/jobs/create-jobs.js:28:126 at Array.reduce (<anonymous>) at createJobs (/Users/n0s00jx/codebase/work/product/node_modules/@bundle-stats/utils/lib/jobs/create-jobs.js:28:53) at Task.task (/Users/n0s00jx/codebase/work/product/node_modules/bundle-stats/bin/run.js:84:27) { context: [Object: null prototype] { artifacts: [ [Object], [Object] ] } } Hi @nilanjansiromani, thanks for raising this! Will try to investigate, but can you please share more details about your setup ? version of bundle-stats / webpack / node any errors when running bundle-stats --demo webpack stats files @nilanjansiromani I was able to replicate the error when passing invalid webpack stats. The latest version (2.2.0) is going to validate the entire structure and give you an error message. Feel free to re-open this issue if you still have problems.
2025-04-01T04:35:20.837277
2015-08-29T20:16:40
103877520
{ "authors": [ "fson", "leebyron", "steveluscher" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10284", "repo": "relayjs/relay-starter-kit", "url": "https://github.com/relayjs/relay-starter-kit/pull/22" }
gharchive/pull-request
Bundle GraphiQL with the starter kit Make the experience of starting with Relay and GraphQL smoother by including GraphiQL in the development server out of the box. GraphiQL is installed as a dev dependency and can be accessed on http://localhost:3000/graphiql Note: npm currently complains about the React peer dependency. This is going to be fixed by https://github.com/graphql/graphiql/pull/14 @leebyron, are you thinking of baking this into express-graphql? Yeah, it would be nice to integrate this into express-graphql. @dschafer and I discussed this earlier today in fact. There are some challenges, but I'll investigate it later this week. I'd love an update on this @leebyron. What's the current thinking? https://github.com/graphql/express-graphql/pull/16
2025-04-01T04:35:20.854330
2024-04-11T21:26:18
2238648434
{ "authors": [ "remal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10285", "repo": "remal-github-actions/create-tag", "url": "https://github.com/remal-github-actions/create-tag/pull/956" }
gharchive/pull-request
Merge template repository changes: remal-github-actions/template-typescript - autoclosed Template repository changes. If you close this PR, it will be recreated automatically. Autoclosing the PR, as no files will be changed after merging the changes from template-sync branch into main branch.
2025-04-01T04:35:20.861868
2017-02-13T22:02:00
207352598
{ "authors": [ "remcoder" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10286", "repo": "remcoder/gcode-previewer", "url": "https://github.com/remcoder/gcode-previewer/issues/1" }
gharchive/issue
Colors for 'zones' Skirts, perimeters, solid layers, supports etc are annotated in some flavors of gcode. Currently only Cura's syntax is picked up. Slic3r doesn't seem to produce this data but S3D does. So let's add colors for S3D flavored gcode. fixed in eed7c80ec12339224c51c040714805e17a7e4063
2025-04-01T04:35:20.863842
2023-06-14T10:29:54
1756567824
{ "authors": [ "giamir", "remcovaes" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10287", "repo": "remcovaes/web-test-runner-vite-plugin", "url": "https://github.com/remcovaes/web-test-runner-vite-plugin/pull/5" }
gharchive/pull-request
feat(config): add the ability to pass vite config as plugin args Hello @remcovaes, First of all thanks for maintaining this package: it is very useful. 🙏 In my current project we are using it in a complex workspace setup where we abstract away test configurations from individual packages into a shared one. We need more granular control on the config picked up by the vite dev-server and therefore we are using a monkey-patched version of your package. I believe what we have patched could be useful also to somebody else therefore this PR. Additionally it would be nice to simply remove a dedicated patched version from our project and simply use your package as a dependency. I have merged it in (after some small changes) as it can also be useful if you want to add React or Vue support without the need for a Vite config. Thanks for you contribution!
2025-04-01T04:35:20.878450
2019-06-30T23:38:09
462459086
{ "authors": [ "LeffelMania", "rex-remind101" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10288", "repo": "remind101/AutoGraph", "url": "https://github.com/remind101/AutoGraph/pull/105" }
gharchive/pull-request
Cleaner json api Call autograph.send(includingJSONResponse: request) and it will return a result (data, error) where data.value is the serialized object value and data.json is the json value. This keeps the API cleaner. No additional tuples inside of the success result and user can choose to request the wrapped struct with json only if they actually want it. We should probably change the name of data.value though, maybe data.serializedObject? I'm not actually sure I think this interface is better or cleaner. But I don't feel strongly against it, and I don't really have a compelling argument other than my feelings, so I'm happy to give it a whirl. 👍
2025-04-01T04:35:20.880838
2016-08-22T04:45:47
172368919
{ "authors": [ "ejholmes", "jer", "marksteve" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10289", "repo": "remind101/conveyor", "url": "https://github.com/remind101/conveyor/issues/64" }
gharchive/issue
AMI no longer exists Parameter validation failed: parameter value ami-2d097847 for parameter name AmiId does not exist. Rollback requested by user. 👍 same here. Is there a new AMI? Is this the right one? ami-f4435895 Hey @marksteve / @jer. Sorry for the late response on this. I just verified that this AMI exists and is marked as Public, so not sure why you'd be running into this issue. You can always use packer to build the AMI from scratch within your AWS account with make ami (however, it's been a while since an AMI was built from scratch, so it's possible you may run into issues with that).
2025-04-01T04:35:20.885335
2019-03-02T03:50:15
416353717
{ "authors": [ "fubar", "russellballestrini" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10290", "repo": "remind101/ssm-env", "url": "https://github.com/remind101/ssm-env/issues/25" }
gharchive/issue
A silent-fail option would be useful I am using SSM parameters to populate environment variables. It's a common pattern for applications to check if an environment variable is present and non-empty before using it, ie. expecting that it may not be set. In that sense it would be useful if ssm-env had a silent-fail option that would resolve a non-existent SSM parameter to an empty value instead of throwing an error. @russellballestrini what do you think? Could you show an example? I feel like dealing with the error of a non-existent parameter would be done outside the tool but I might not fully understand your usecase. As it is, I don't have the ability to deal with non-existent parameters outside of this tool as this tool throws an error when it encounters one, which is exactly the issue I'm looking to find a more robust solution for. Presume I have an application that uses a default DB port but allows that to be overwritten via configuration based on environment variables (as is common practice): const config = { dbPort: env.DB_PORT || 1234, foo: env.FOO || '', }; Assume I have set up SSM parameter foo with value "bar", but parameter /db/port has not been created. FOO=ssm://foo DB_PORT=ssm:///db/port ssm-env node foo.js This application will fail to start because ssm-env throws an error. This creates undesired coupling between the command and the state of configuration (ie. existence of SSM parameters); or in other words, if I configure an optional config value by creating an SSM parameter, I now also have to modify the command to include that env var/SSM param pair. What I'm looking for is something like a --silent-fail option that would result in: FOO=bar DB_PORT= Now my application can properly handle configuration with optional values. @russellballestrini any thoughts on this? I'll review a PR, since it is opt-in it shouldn't break existing workflows. https://github.com/remind101/ssm-env/pull/26 I just merged this in, going to close this issue!
2025-04-01T04:35:20.905274
2023-08-23T11:33:53
1863149591
{ "authors": [ "omelhus" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10292", "repo": "remix-run/remix", "url": "https://github.com/remix-run/remix/pull/7237" }
gharchive/pull-request
bug: multipart with boundary wrapped with " fails When a multipart/form-data upload is done from .NET it wraps the value of the boundary property in the content-type header with ", like "multipart/form-data; boundary=\"--------------------------890934293568639326555573\"", and this breaks the upload in unstable_parseMultipartFormData. Closes: Did not bother with creating an issue. Fix included in PR. [ ] Docs [x] Tests Testing Strategy: This test covers this code: integration/form-data-multipart-test.ts I guess you can just cherry pick the three lines to fix this :)
2025-04-01T04:35:20.913813
2018-10-25T12:54:53
373932914
{ "authors": [ "alfonz19", "remkop" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10293", "repo": "remkop/picocli", "url": "https://github.com/remkop/picocli/issues/529" }
gharchive/issue
require subcommand I have defined command and several subcommands, but it seems, that subcommand is just optional. Is there/could there be a way, how to flag, that selection of some subcommand is required? One way would be to show an error if the top-level command is invoked without subcommand: @Command(name = "top", subcommands = {Sub1.class, Sub2.class}) class TopCommand implements Runnable { public void run() { System.err.println("Please invoke a subcommand"); new CommandLine(new TopCommand()).usage(); } public static void main(String[] args) { CommandLine.run(new TopCommand(), args); } } @Command(name = "sub1) class Sub1 implements Runnable { public void run() { System.out.println("All good, executing Sub1"); } } @Command(name = "sub2) class Sub2 implements Runnable { public void run() { System.out.println("All good, executing Sub2"); } } Yes, thanks! (sorry for delay, I had to verify it). But to proceed with this topic: • if someone specified subcommands say via annotation, doesn't it mean, that one of them should be used? • also there is some state, which I don't undestand: if I do what you adviced, I'd expect, that when subcommand is not specified, it will produce failure. But say when I pass "-V" and no command, it succeed. That's great, that's what it should do, but it's not what I expected. However if you pass another option of top command, it will fail. This brings me to idea, that there should be flag: fail on missing command. Which would be saving lots of boilerplace code when there is a hierarchy of commands. • also it feels little bit weird, that I'm in middle of CommandLine processing and creating new instance, and I'm in TopCommand runnable and creating another one to print it's usages... if someone specified subcommands say via annotation, doesn't it mean, that one of them should be used? Not necessarily, there may be command hierarchies where it is valid to invoke the top-level command. when I pass "-V" and no command, it succeed. That's great, that's what it should do, but it's not what I expected. The --version and --help options are special: if these options are specified, the normal validation for missing required options etc is bypassed. Also, the run method is not invoked. See https://picocli.info/#_help_options for details. there should be flag: fail on missing command. Which would be saving lots of boilerplace code when there is a hierarchy of commands. I'm not sure I understand. Can you elaborate? it feels little bit weird, that I'm in middle of CommandLine processing and creating new instance, and I'm in TopCommand runnable and creating another one to print it's usages... You are right, sorry. You can use this instead of creating a new TopCommand instance: class TopCommand implements Runnable { public void run() { System.err.println("Please invoke a subcommand"); new CommandLine(this).usage(); } Got it. To elaborate on that "failOnMissing": you said, that there hierarchies, where it might be valid to skip command. There might be hierarchies, where it's invalid. But I doubt, that there will be hierarchy, where some of defined command will be optional and some not. If that's true, then there can be annotation option say "oneOfSubCommandRequired" with boolean value, and generated behavior as we're adding here with .usage() call. btw. I know it's not related to this topic, but if we're already speaking I don't want to create separate thread for that — is it possible to have dependency between options? If this is specified, that has to be also? Or this is subject of manual coding? Thanks. Soon: dependencies between options in on the TODO list: https://github.com/remkop/picocli/issues/295 I hope to be able to get to it in picocli 3.8 (but no guarantees). Until then applications will need to manually take care of this. I've created this StackOverflow question to help build a knowledge base for other users: https://stackoverflow.com/q/53043937/1446916 @alfonz19 Let me know if you have any follow-up questions. If not, can this ticket be closed? thanks for everything!
2025-04-01T04:35:20.926731
2024-06-27T10:00:28
2377715757
{ "authors": [ "remyschwab" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10294", "repo": "remyschwab/TetRex", "url": "https://github.com/remyschwab/TetRex/issues/53" }
gharchive/issue
Amino Acid Reduced Alphabet Search Indexing with the reduced alphabet is pretty straightforward but verifying is not! Querying: [ ] Preprocessing the query by replacing letters with one (need mapping of chars) [ ] Alter graph construction default case to place character nodes in a buffer instead of instantly adding them to NFA [ ] What's the best way to read the reference files? The graph simplification isn't strictly necessary... It might be a better idea to have a separate method for graph construction when there is a reduction since it's slightly more complex. It's also probably smart to replace the current I/O approach with a view
2025-04-01T04:35:20.936302
2017-02-26T15:35:54
210326135
{ "authors": [ "mfazekas", "renchap", "sevos" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10295", "repo": "renchap/webpacker-react", "url": "https://github.com/renchap/webpacker-react/issues/17" }
gharchive/issue
Idea: ERB helper to easy enabling HMR I have been thinking on how to improve our HMR support and have a way to toggle it globally while reducing the boilerplate. The best idea I have is to provide a Ruby helper and have people use packs/application.js.erb. The interface would look like: # packs/application.js.erb <%= register_react_components { Hello: 'components/Hello', App: 'components/App', } %> Which will generate the following JS code: import Hello from 'components/Hello' WebpackerReact.register(Hello) if (module.hot) { module.hot.accept( 'components/Hello', () => WebpackerReact.renderOnHMR(Hello) ) } import App from 'components/App' WebpackerReact.register(App) if (module.hot) { module.hot.accept( 'components/App', () => WebpackerReact.renderOnHMR(App) ) } We can then have a Rails config flag to enable / disable HMR, as well as a register_react_components param to override it. This will also allow to only enable HMR (and lot relevant code) depending on Rails environment. Webpacker enables ERB processing in Webpack config by default, so this should work out of the box. I dont fully like the need to have the users move their packs to ERB, but it looks like much more flexible than any other solutions I thought about. @sevos @mfazekas @daninfpj any feedback? Again, why this javascript can't be part of the JS library? Is this to solve the following warning? Yes, you can find more context here: https://github.com/renchap/webpacker-react-example/pull/2#issuecomment-278063213 We havent been able to find documentation on the internals for HMR in Webpack 2. I would have preferred a 100% JS solution where you only have to register your component, but it does not seem possible due to the way HMR works. @renchap i like the idea, alternative would be a babel plugin, but the .erb solution is just much simpler. 👍 @sevos do you think we can have a simpler pure-JS solution? I tried again to look at how to make this work with Webpack & hot-reloading, without any results. I will work on implementing the ERB otherwise. First, what do you think if we define the requirements explicitly? Like a real ticket at a software product dev. What info needs to be passed from-to, etc. Sure. I have several goals: avoid the need for boilerplate to be added for HMR (in the common cases at least) be able to easily enable / disable HMR depending on a Rails config option. Ideally I would like for new users to try it out just by flipping this option and running ./bin/webpack-dev-server --hot in place of webpack-watcher completely disable all HMR-related code in production (smaller pack)
2025-04-01T04:35:20.994745
2024-12-16T13:16:37
2742371790
{ "authors": [ "Sam-Newman", "arein", "vanyasav" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10296", "repo": "reown-com/appkit", "url": "https://github.com/reown-com/appkit/issues/3470" }
gharchive/issue
[docs] Inconsistency in appkit version between npm and github Summary Link to the related docs page https://github.com/reown-com/appkit Thanks for raising - we will look into it! We're aware of this - will fix soon
2025-04-01T04:35:21.007488
2022-09-17T12:08:34
1376755338
{ "authors": [ "Keep-up-sharma", "andreasjansson", "ishowta" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10297", "repo": "replicate/cog-stable-diffusion", "url": "https://github.com/replicate/cog-stable-diffusion/issues/21" }
gharchive/issue
Investigate: img2img prompt_strength=1.0 doesn't seem to destroy the init image completely When prompt_strength=1, SD should behave as though there were no init image. However, that doesn't seem to be the case as has been reported on Discord: https://discord.com/channels/775512803439280149/881521351381385216/1019910914692108309 We should investigate why this happens and fix it! https://github.com/replicate/cog-stable-diffusion/pull/22 Created a pull request Maybe strength is the right key. https://github.com/huggingface/diffusers/blob/a4d5b59f132126c06c1a6b1f266ee44c70440cce/src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_img2img.py#L55
2025-04-01T04:35:21.010566
2022-05-12T13:33:10
1233994618
{ "authors": [ "bfirsh", "evilstreak" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10298", "repo": "replicate/cog", "url": "https://github.com/replicate/cog/issues/607" }
gharchive/issue
Specifying an unsupported CUDA version generates a cryptic error If you specify a version of CUDA in cog.yaml that doesn't exist in pkg/config/cuda_base_image_tags.json then cog build fails with a pretty cryptic error message: $ cog build panic: runtime error: index out of range [0] with length 0 goroutine 1 [running]: github.com/replicate/cog/pkg/config.latestCuDNNForCUDA({0x14000213958, 0x6}) /Users/dominicbaggott/replicate/cog/pkg/config/compatibility.go:204 +0x108 github.com/replicate/cog/pkg/config.(*Config).validateAndCompleteCUDA(0x140001cf7d0) /Users/dominicbaggott/replicate/cog/pkg/config/config.go:274 +0xa1c github.com/replicate/cog/pkg/config.(*Config).ValidateAndCompleteConfig(0x140001cf7d0) /Users/dominicbaggott/replicate/cog/pkg/config/config.go:134 +0xb8 github.com/replicate/cog/pkg/config.GetConfig({0x0, 0x0}) /Users/dominicbaggott/replicate/cog/pkg/config/load.go:46 +0xe0 github.com/replicate/cog/pkg/cli.buildCommand(0x140001f3900, {0x104bf57b0, 0x0, 0x0}) /Users/dominicbaggott/replicate/cog/pkg/cli/build.go:28 +0x38 github.com/spf13/cobra.(*Command).execute(0x140001f3900, {0x104bf57b0, 0x0, 0x0}) /Users/dominicbaggott/.go/pkg/mod/github.com/spf13/cobra@v1.4.0/command.go:856 +0x668 github.com/spf13/cobra.(*Command).ExecuteC(0x140001f3680) /Users/dominicbaggott/.go/pkg/mod/github.com/spf13/cobra@v1.4.0/command.go:974 +0x410 github.com/spf13/cobra.(*Command).Execute(...) /Users/dominicbaggott/.go/pkg/mod/github.com/spf13/cobra@v1.4.0/command.go:902 main.main() /Users/dominicbaggott/replicate/cog/cmd/cog/cog.go:14 +0x94 The right response here is to change the version of CUDA you're specifying to one that's supported. We should handle this error more gracefully, and show a useful message to the user with that instruction. I found this when trying to build a model that specified cuda: "11.1", which used to be a supported version until b057e19. Here's a minimal cog.yaml to reproduce: build: gpu: true cuda: "11.1" python_version: "3.8" predict: "predict.py:Predictor" Ah. 11.1 should also probably work, and resolve to latest patch version. A lot of these issues, and things like #489, make me think we might want to step back and see if we can come up with a more robust overall design.
2025-04-01T04:35:21.011958
2023-07-26T17:10:21
1822831246
{ "authors": [ "zeke" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10299", "repo": "replicate/llama-chat", "url": "https://github.com/replicate/llama-chat/pull/11" }
gharchive/pull-request
link to streaming guide from README https://replicate.com/docs/streaming Tests failing, but I don't think it's from this README change, so I'm merging.
2025-04-01T04:35:21.082636
2024-05-19T11:02:58
2304554407
{ "authors": [ "jleibs", "traversaro" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10300", "repo": "rerun-io/cpp-example-opencv-eigen", "url": "https://github.com/rerun-io/cpp-example-opencv-eigen/pull/28" }
gharchive/pull-request
Add example of downloading rerun c++ lib via pixi To be honest, I am not fully convince about the duplication at the pixi level (so feel free to decline or heavily change the PR), but I found convenient to have a way to quickly validate that indeed the librerun-sdk conda package works fine. I also updated rerun version to 0.16.0 as the librerun-sdk conda-forge package is available for Windows only since 0.16.0 . Replaced with: https://github.com/rerun-io/cpp-example-opencv-eigen/pull/30
2025-04-01T04:35:21.083991
2023-09-05T07:43:51
1881397853
{ "authors": [ "Wumpf", "teh-cmc" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10301", "repo": "rerun-io/rerun", "url": "https://github.com/rerun-io/rerun/issues/3207" }
gharchive/issue
Implement missing roundtrip tests & API examples for C++ What the title says. Look for TODO(#3207) in the code. We've done all of them by now! actually still need to enable transform roundtrip and remove #3207 comment!
2025-04-01T04:35:21.097577
2024-10-23T12:28:31
2608428125
{ "authors": [ "Wumpf", "abey79", "gavrelina" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10302", "repo": "rerun-io/rerun", "url": "https://github.com/rerun-io/rerun/issues/7877" }
gharchive/issue
Releasable MVP for the Map view Initial PR: #6561 Absolute "must" for the first release: [x] Visible time range [x] https://github.com/rerun-io/rerun/issues/7849 [x] Interactive entities: selectable and hover tooltip (blocked on https://github.com/podusowski/walkers/issues/197) https://github.com/rerun-io/rerun/pull/7938 [x] https://github.com/podusowski/walkers/pull/221 [x] https://github.com/rerun-io/rerun/issues/7937 [x] https://github.com/rerun-io/rerun/pull/7972 [x] bp api + snippet/screenshot: https://github.com/rerun-io/rerun/pull/8027 [x] https://github.com/rerun-io/rerun/issues/7872 https://github.com/rerun-io/rerun/pull/8015 [x] #7968 [x] docs: https://github.com/rerun-io/rerun/pull/8084 [x] Update nuscene example: #8034 [x] How-to guide: very simple, mostly for discoverability: https://github.com/rerun-io/rerun/pull/8075 [x] Basic geoXXX example that is included in the viewer: https://github.com/rerun-io/rerun/pull/8044 [x] https://github.com/rerun-io/rerun/issues/8032 https://github.com/rerun-io/rerun/pull/8055 [x] https://github.com/rerun-io/rerun/pull/8078 [x] https://github.com/rerun-io/rerun/issues/8052 OK not to have for the first release: [ ] https://github.com/rerun-io/rerun/issues/8077 [ ] dogfood: update the air traffic data example to use the map view [ ] #8086 [ ] https://github.com/rerun-io/rerun/issues/7963 [ ] #8067 [ ] https://github.com/rerun-io/rerun/issues/7876 [ ] #8066 [ ] #8068 [ ] #8087 [ ] #8089 [ ] #8090 thoughts and ideas! https://github.com/podusowski/walkers/issues/210 add support to pan with the trackpad allow for transparency of the point styles (within a map context the layer of map actually has important information, so it's great to allow users flexibility to peek it once they need it) think of a seamless and easy flow to add mapbox API keys I'd propose to change the default style of dots to have some kind of tiny shadow under, because maps are usually very busy with lots of colors and signs already, and now the dots are easily blending in with the rest. so would be cool to think how can we can make it standing out better by default. not urgent, but a nice improvement with time. I can also imagine that being able to define different symbols and not only dots would come in handy. and a legend! and adding labels! (okay I will stop here, I think the #1 is the prio, then the API key flow, and the rest can be added as we go) everything in the must-have list got done!
2025-04-01T04:35:21.108968
2024-02-15T15:42:09
2136853774
{ "authors": [ "Wumpf" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10303", "repo": "rerun-io/rerun", "url": "https://github.com/rerun-io/rerun/pull/5207" }
gharchive/pull-request
Lift line vertex/strip count limitations What follow-up of #5192 Fixes #3076 Fixes #4844 technically there's still limits (see #5192 ) but they are machine dependent making this relatively hard to query. Overall I'd argue we removed the previous limits so there's no need for this in the original sense. (note that we haven't spend a lot of time optimizing the collection of lines, large amounts of lines don't perform all that well so far) Applies the same principles as on the previous PR that fixes the point cloud limits. The key difference here is that we use lines in a lot of places, making this a bigger refactor than originally assumed. ~The need to know both strip & vertex count for lines ahead of time is a bit problematic and isn't always as easy as it was with points. I had to introduce LineDrawableBuilderAllocator which is a very simplistic Vec-like allocator (minus the increase in size) for LineDrawableBuilder (previously called LineStripSeriesBuilder). I'm not super happy with this construct overall, but it's the best I could come up with in the short-term and things seem to be fairly robust and at least not overly complicated.~ ~In the future it would be nice to reconcile LineDrawableBuilderAllocator and LineDrawableBuilder into a single construct, likely still with the limitations that the size of a batch (think named unit with a transform) needs to be known ahead of time, which is practically always the case!~ Second iteration: There's now DataTextureSource (ideas for better names?) which is essentially a thing where you can throw data in and get a data texture out! It handles all the copies and dynamic sizings for you. This makes everything awesome because now we can handle reserve call just as an optimization without requiring them and without being on a bad path if you don't! <3 Checklist [x] I have read and agree to Contributor Guide and the Code of Conduct [x] I've included a screenshot or gif (if applicable) [ ] I have tested the web demo (if applicable): Using newly built examples: app.rerun.io Using examples from latest main build: app.rerun.io Using full set of examples from nightly build: app.rerun.io [x] The PR title and labels are set such as to maximize their usefulness for the next release's CHANGELOG [x] If applicable, add a new check to the release checklist! PR Build Summary Docs preview Examples preview Recent benchmark results Wasm size tracking I wonder if we could instead use a dynamic mode for the line allocator, i.e. that collects in RAM and push to VRAM once finished. yes, probably. But do we really want to fork the linebuilder right now :/ Hmmm actually.. maybe we can make the regular line builder growable by having several cpu-write-gpu-read buffers in there 🤔 We just need to be careful about this texture copy business. But conveniently the line build hasn't been ported all that far yet Alright! This is now worlds better than before and will make putting everything else in points & lines onto a cpu-write-gpu-read buffer a breeze!! I wish there was an easy way to write unit tests for this new allocator without shimming all wgpu calls 🤔
2025-04-01T04:35:21.117038
2020-11-19T22:53:11
866691409
{ "authors": [ "bsansouci", "cristianoc" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10304", "repo": "rescript-lang/rescript-vscode", "url": "https://github.com/rescript-lang/rescript-vscode/issues/118" }
gharchive/issue
Error parsing docs for Js.re.exec_ in stdlib Should I file this on bucklescript? We'll just remove special odoc parsing.
2025-04-01T04:35:21.125379
2018-11-06T15:18:07
377893723
{ "authors": [ "quasarchimaere", "sigpie" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10305", "repo": "researchstudio-sat/webofneeds", "url": "https://github.com/researchstudio-sat/webofneeds/pull/2499" }
gharchive/pull-request
Fix personas in re posts Closes #2471 Added persona selections to adhoc connections, i.e. connections started from a WhatsX-Post or from a shared Post-URI. Things to check: persona selection is only visible if the user is logged in persona selection is only visible where it's useful, that is creating a new post, not for any actions that do not create a new post the submit button (which is next to every text field) does behave as expected i find the button label rather long, i know this is our standard label but in this case we should figure out something else because its too long on desktop and def. too long in mobile. so if we think the missing padding is an issue we should see the labellength as an even more severe issue Padding has been added in all cases now. I agree that the labels are too long, but I couldn't think of a shorter label to use instead. Even if we shorten the "Anonymous" part, we'd still need to add the persona name (regardless of length) after a persona is selected. For now, the button label wraps and is displayed in two lines, but this needs some further investigation by @peacememories. An option could be to keep the label we had before instead of the "...as Anonymous" If a Persona is selected we could write "...as [Identicon]" (and show the identicon of the persona instead of the name, since the name would be visible in the selection menu anyway I can see keeping out the "Anonymously" as an option to make the labels shorter. Adding "as [Identicon]" would probably look weird though and might add other layouting issues. Without wrapping the text, it would still be too long for small displays (i.e. more than half the screen width). With wrapping the text, the button would need to be higher, which would also need to increase the height of the text field and the add details button. Also, I don't want to use a different format for displaying personas here compared to the publish button at post creation. Two elements with the same functionality should also have the same look.
2025-04-01T04:35:21.145919
2017-05-02T19:21:44
225794726
{ "authors": [ "jhermsmeier" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10306", "repo": "resin-io/resin-preload-image-script", "url": "https://github.com/resin-io/resin-preload-image-script/pull/42" }
gharchive/pull-request
README.md: Add requirements section This adds a section about requirements for running on each OS To Do: [ ] Add note about having to run bash preload.sh on Windows [ ] Expand on why git is listed as a requirement on Windows Connects To: #41 Closing in favor of #43
2025-04-01T04:35:21.147963
2015-09-14T20:25:48
106420872
{ "authors": [ "pcarranzav", "petrosagg" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10307", "repo": "resin-io/resin-wifi-connect", "url": "https://github.com/resin-io/resin-wifi-connect/issues/18" }
gharchive/issue
Cleanup the Edison version and move to a forked repo The edison version works, but can be refactored to also work with connman (currently uses hostapd) now that the kernel module issue it had is resolved. It should also be a separate repo (fork), to improve its visibility and allow making PR's from one to the other to update functionality. resin-wifi-connect is now based on hostapd for all devices so we'll keep the code here. Edison support is pending. We have to load and unload a kernel module like they do in their official images[1] but other than that it works. [1] https://git.yoctoproject.org/cgit/cgit.cgi/meta-intel-edison/tree/meta-intel-edison-distro/recipes-connectivity/hostapd/files/hostapd.service
2025-04-01T04:35:21.149761
2016-11-17T11:20:32
190018270
{ "authors": [ "floion" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10308", "repo": "resin-os/meta-resin", "url": "https://github.com/resin-os/meta-resin/pull/477" }
gharchive/pull-request
Migrate to poky morty version Connects to https://github.com/resin-os/resinos/issues/113 @agherzan @telphan @michal-mazurek This PR is replaced by https://github.com/resin-os/meta-resin/pull/523 Closing now
2025-04-01T04:35:21.157223
2011-11-11T00:09:20
2204648
{ "authors": [ "parndt", "resetexistence" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10309", "repo": "resolve/refinerycms", "url": "https://github.com/resolve/refinerycms/issues/1137" }
gharchive/issue
Bin refinerycms to generator Overhaul of how we used to generate a refinery cms app. The idea here is that we use Thor to delegate to different generators and just run them in succession instead of shelling out to the command line. Will review asap gem install refinerycms --version 2.0.0 --- 61 gems installed 15:01:46: ~$ which refinerycms /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/bin/refinerycms 15:01:47: ~$ refinerycms foobar create create README create Rakefile create config.ru create .gitignore create Gemfile create app create app/assets/images/rails.png create app/assets/javascripts/application.js create app/assets/stylesheets/application.css create app/controllers/application_controller.rb create app/helpers/application_helper.rb create app/mailers create app/models create app/views/layouts/application.html.erb create app/mailers/.gitkeep create app/models/.gitkeep create config create config/routes.rb create config/application.rb create config/environment.rb create config/environments create config/environments/development.rb create config/environments/production.rb create config/environments/test.rb create config/initializers create config/initializers/backtrace_silencers.rb create config/initializers/inflections.rb create config/initializers/mime_types.rb create config/initializers/secret_token.rb create config/initializers/session_store.rb create config/initializers/wrap_parameters.rb create config/locales create config/locales/en.yml create config/boot.rb create config/database.yml create db create db/seeds.rb create doc create doc/README_FOR_APP create lib create lib/tasks create lib/tasks/.gitkeep create lib/assets create lib/assets/.gitkeep create log create log/.gitkeep create public create public/404.html create public/422.html create public/500.html create public/favicon.ico create public/index.html create public/robots.txt create script create script/rails create tmp/cache create tmp/cache/assets create vendor/assets/stylesheets create vendor/assets/stylesheets/.gitkeep create vendor/plugins create vendor/plugins/.gitkeep /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/refinerycms-core-2.0.0/lib/generators/refinery/app/app_generator.rb:238:in `read': No such file or directory - foobar/Gemfile (Errno::ENOENT) from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/refinerycms-core-2.0.0/lib/generators/refinery/app/app_generator.rb:238:in `find_and_replace' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/refinerycms-core-2.0.0/lib/generators/refinery/app/app_generator.rb:177:in `generate_rails!' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/refinerycms-core-2.0.0/lib/generators/refinery/app/app_generator.rb:61:in `run!' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/task.rb:22:in `run' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/invocation.rb:118:in `invoke_task' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/invocation.rb:124:in `block in invoke_all' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/invocation.rb:124:in `each' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/invocation.rb:124:in `map' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/invocation.rb:124:in `invoke_all' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/group.rb:226:in `dispatch' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/thor-0.14.6/lib/thor/base.rb:389:in `start' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/gems/refinerycms-2.0.0/bin/refinerycms:5:in `<top (required)>' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/bin/refinerycms:19:in `load' from /Users/parndt/.rvm/gems/ruby-1.9.3-p0@refinery200/bin/refinerycms:19:in `<main>' Have to modify the find and replace function from the old script then - there's a built in Thor/Rails actions way to do it for generators
2025-04-01T04:35:21.160462
2024-05-11T15:15:21
2290954845
{ "authors": [ "maximdudai", "violetadev" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10310", "repo": "responsively-org/website", "url": "https://github.com/responsively-org/website/pull/49" }
gharchive/pull-request
UI elements - Navbar & Download button changed nav links hover style Star icon color when hover Download button styles Download button label -> Start Your Journey @maximdudai do you want to rebase this? I will merge it after. Thanks
2025-04-01T04:35:21.167711
2016-03-21T13:56:36
142355444
{ "authors": [ "exalted", "khalidabuhakmeh" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10311", "repo": "restful-routing/restful-routing", "url": "https://github.com/restful-routing/restful-routing/issues/89" }
gharchive/issue
Variable resource name Hi, I wonder if is it possible obtaining a route table similar to what's below? http://example.com/ http://example.com/{organization} http://example.com/{organization}/customers http://example.com/{organization}/customers/{id} The "interesting" thing above is that the URL for a given organization is given by it's name (as a parameter) rather than something like http://example.com/organizations/{id}. Thanks. I guess Rails (as an example) allows "empty" paths in routes, so we should be able to do something like: map.Resources<OrganizationsController>( x => { x.Path(""); // !!! x.Resources<CompaniesController>(); }); I'll have to think about this one, but off the top of my head I would just setup a Path and not use the Resources registration. :+1: Crazy thought, I haven't tried this, but maybe you could. map.Resources<OrganizationsController>( x => { x.As("{organization}"); x.Resources<CompaniesController>(); }); Note that that As method is used, and passes in a {organization} placeholder, instead of a string. @khalidabuhakmeh sadly, as it unfortunately happens in these cases, I had to move along without restful-routing in my project (actually to be 100% honest, I still use ApplicationController#RespondTo with FormatResult, which is a :star: ). This all is to say that I don't have a project to test your suggestions out and I am glad there may be workarounds (?). Thanks. That's a bummer, but I'm glad you found something positive out of the project. Good luck with your project.
2025-04-01T04:35:21.184611
2016-12-12T02:48:52
194879803
{ "authors": [ "arithmetric", "codecov-io", "fd0", "ibib", "kurin", "mholt", "tscs37" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10312", "repo": "restic/restic", "url": "https://github.com/restic/restic/pull/694" }
gharchive/pull-request
Backblaze B2 backend support This PR implements support for Backblaze B2 as a storage backend using kurin's blazer package. The code is based on the S3 backend. Like the S3 backend, the B2 account ID and key credentials should be specified by environment variables (B2_ACCOUNT_ID and B2_ACCOUNT_KEY), while the B2 bucket name and path prefix are specified through the repository URL in the format: b2:<bucket name> or b2:<bucket name>:<prefix path>. If you register with Backblaze B2, you should be able to run the full test suite against the B2 API at least once daily and stay within the free limits. By raising the caps to a modest amount (less than $1 total), you would be able to run the full test suite multiple times. Resolves #512. Current coverage is 50.36% (diff: 9.04%) Merging #694 into master will decrease coverage by 1.03% @@ master #694 diff @@ ========================================== Files 95 97 +2 Lines 7476 7664 +188 Methods 0 0 Messages 0 0 Branches 0 0 ========================================== + Hits 3843 3860 +17 - Misses 3084 3255 +171 Partials 549 549 Powered by Codecov. Last update 51cd78e...38043c3 Thanks for the review @tscs37! I've responded to your comments and made most of the suggested changes. I also pulled in your update to blazer. Ooo, this is timely. I'm looking into restic for my backups on B2 since I'm having lock issues with duplicity. Hope this merges soon! There is probably a bug somewhere in the code for extracting the prefix for the repository. I compiled restic with the changes of the PR and when using a bucket like b2:test-bucket:example it seems to try and use it for the bucket name, returning unable to open repo at b2:test-bucket:test: blazerb2.NewBucket: b2_create_bucket: 400: Invalid characters in bucketName: must be alphanumeric or '-' There also seems to be a problem indicating the upload progress, as uploading a larger repository seems to halt in progress for a while. Hey, thanks for the work so far! Does anyone of you have a contact at backblaze? Maybe we can ask them if they'd be willing to donate an account that can be used for integration testing? Backblaze B2 is free up to 10 GB. @tscs37 The repository location format that I implemented for B2 is b2:[bucket name]/[prefix]. So try b2:test-bucket/example instead of b2:test-bucket:example. @fd0 Thanks for the feedback. I'll submit a ticket to Backblaze support to inquire about a free or sponsored account. @mholt Thanks for your interest in this feature. The 10 GB free storage is more than enough for the tests, however one run of the e2e tests consumes 50% of the free daily download allocation (~500 MB) and 20% of the free requests (~500 class B and C requests). however one run of the e2e tests consumes 50% of the free daily download allocation (~500 MB) and 20% of the free requests (~500 class B and C requests). Holy smokes, that's a lot. Didn't realize how expensive those tests were. I'm sure this is obvious, but maybe in case some support rep doesn't understand the implications, make sure to tell them that if they can help raise the limits for a test account, that more people will be purchasing B2 service -- myself included. We can certainly lower the amount of requests/bytes for the CI tests for a Backblaze backend... @arithmetric Thanks for contacting Backblaze. I ran it again using b2:bucket/prefix and it works perfectly now. As a side note; I've found that the directory structure on backblaze does not reflect the on-disk format, making both compatible as much as possible would be a nice additional feature. That way existing backups can be uploaded and reused. What's left on this? Just a test account? Is there any way I can help? fwiw I made an API-breaking change to blazer to implement bucket attributes, but it's a small one and this patch vendors blazer anyway. Ping. :) Is there anything I can do to help get this merged? There is at least one merge conflict, and the backend API has changed since this code was submitted. Since there wasn't any activity for a few weeks, I'm tempted to close this PR for now. @mholt if you're interested in giving this a try and reworking the code, I'd be glad to assist! For reworking the whole PR I might need more time -- right now I have to get a new website and backend finished, but if this still needs help after that's done, I might be able to help, yeah. It just might be a month or two... Ok, then I'll close this PR for now, please let me know when/if you (or somebody else) have time and motivation to work on this. I promise I'll help :) I could poke at it, but it felt like this PR spent a long time in a state where it was basically done, but we didn't have a good story about how to run the integration tests. I don't want to complete a PR against the new backend only to have it sit for months in the same place. What do you need from backends in terms of tests, so that this isn't as big an obstacle next time? Ideally, I'd like to run the backend tests for each commit/PR, just as the other tests. For now we're running an sftp-server (for the sftp backend) and a minio s3 server (for the s3 backend) locally on Travis, and use that for tests. The upside is that we don't depend on any third-party services. I think it's okay to use a third-party service for running the CI tests (I suspect that there will be more backends like backblaze in the future which we cannot just run locally). I understood that there is a test account with some free space for backblaze that may be suitable to run the test suite against for each commit/PR. What are the constraints for using the free backblaze account? We can certainly reduce the amount of data for each test, if that is still a concern. @arithmetric you said that you contacted backblaze, did they respond? Source Free Tier: Storage: 10GB Upload: Free Download: 1GB/Day Transactions: 2500 free Downloads/day 2500 other Transactions/day Hm, I think that the free tier may work, if we reduce the amount of data a bit. Let's continue the discussion in #512. I've submitted a new request with Backblaze.
2025-04-01T04:35:21.194719
2022-12-06T02:10:35
1477964683
{ "authors": [ "olivierodo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10313", "repo": "restqa/restqa", "url": "https://github.com/restqa/restqa/issues/288" }
gharchive/issue
🚀 [FEATURE]: Generate mock from curl command - Plugin HTTP Mock 👀 Background The plugin Http Mock is allowing RestQA to. mock external http dependencies. ✌️ What is the actual behavior? While working in with RestQA if we want to mock an external dependency we need to create them as yaml file. Which could be cumbersome... 🕵️‍♀️ How to reproduce the current behavior? Install RestQA npm i -D @restqa/restqa Install the plugin http mock npm i -D @restqa/plugin-http-mock initiate the project: $(npm bin)/restqa run Add the plugin to your project https://dashboard.restqa.io/#/documentation/http-mock Create your first mock 🤞 What is the expected behavior? It would be great to be able to generate mock more easily 😎 Proposed solution. Having a command on RestQA or in the plugin allowing to generate mock from a curl command. Example: If I want to generate a mock representing the api running on localhost:3000/test i could write the command: restqa generate curl localhost:3000/test --plugin http-mock 🙏 Would you be willing to submit a PR? Yes Code of Conduct [X] I agree to follow this project's Code of Conduct In order to to avoid dependencies with the plugin @restqa/plugin-http-mock let use the following: npx @restqa/plugin-http-mock curl localhost:3000/test
2025-04-01T04:35:21.198091
2023-04-29T23:34:49
1689718883
{ "authors": [ "allomov", "keymastervn" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10314", "repo": "retailzipline/pg_generated_column_support", "url": "https://github.com/retailzipline/pg_generated_column_support/pull/2" }
gharchive/pull-request
Update pg gem version in gemspec to support latest verstions of the gem Hello! It works fine for me with gem "pg", "~> 1.4.3". Think that update makes sense. Thank you! Same issue
2025-04-01T04:35:21.236737
2023-05-02T14:39:29
1692554858
{ "authors": [ "Axelen123", "oSumAtrIX" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10315", "repo": "revanced/ARSCLib", "url": "https://github.com/revanced/ARSCLib/issues/2" }
gharchive/issue
Fork EOL This fork was originally made because the patcher required some patches in order to work. The only patch left right now is the one I made to fix some performance issues with the apk writer on android devices. The upcoming archive2 backend does not have the performance issues of the old java.util.zip-based backend. This fork will no longer be needed after archive2 is released and patcher would be able to use upstream ARSCLib. Can you link anything from upstream to this issue for tracking? There are no issues upstream tracking archive2 development How come it is subject to be introduced? An issue or PR should be opened upstream so that it can be tracked there as well as referenced here.
2025-04-01T04:35:21.253374
2017-03-17T06:08:00
214916210
{ "authors": [ "brendensoares", "notzippy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10316", "repo": "revel/revel", "url": "https://github.com/revel/revel/issues/1139" }
gharchive/issue
Auto route method to controller action I'd like to be able to define a route that translates all HTTP methods to controller actions based on convention like: * /api/sessions Sessions instead of having to list out GET, POST, PUT, DELETE, etc. It should be working that way, see manual here http://revel.github.io/manual/routing.html Can you confirm this does not work ? Closing due to no response from reported
2025-04-01T04:35:21.256250
2018-09-29T08:39:28
365111974
{ "authors": [ "mrarrowb" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10317", "repo": "revel/revel", "url": "https://github.com/revel/revel/issues/1375" }
gharchive/issue
Unable to run myapp in Windows Hi, I am very new to GO programming language and enjoying it completely. But while starting to work on web application development using Revel framework, i created the application using "revel new myapp", which created the folder alright. That said, when i am running the command "revel run myapp", i am getting the below error, could anybody please help. Error Message: Running run Revel executing: run a Revel application DEBUG 13:46:09 revel server.go:27: RegisterServerEngine: Registered engine section=server name=go DEBUG 13:46:09 revel template_engine.go:45: Registered template engine loaded section=template go=nil LOG15_ERROR="Normalized odd number of arguments by adding nil" INFO 13:46:09 app revel.go:170: app.conf: No mode found: {"mode":"dev", "specialUseFlag":false} section=system ERROR 13:46:09 harness.go:231: Could not start application error="revel/harness: app died reason: exit status 1" closing this issue
2025-04-01T04:35:21.260097
2022-06-30T15:26:25
1290302813
{ "authors": [ "SoMuchForSubtlety" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10318", "repo": "reviewdog/reviewdog", "url": "https://github.com/reviewdog/reviewdog/issues/1207" }
gharchive/issue
github-pr-check fails to complete if there are too many findings Updating the check run fails with the following error message: UpdateCheckRun failed: {"message":"Invalid request. Only 65535 characters are allowed; 66618 were supplied.","documentation_url":"https://docs.github.com/rest/reference/checks#update-a-check-run"} At some point there was an attempt to fix this in fe3996e1bd87c07ce52a705453e517c9e1153531, but it seems like the change was not sufficient. Currently, the summary is limited to 150 findings, but if the text for the individual findings is very long, this can still be too much. https://github.com/reviewdog/reviewdog/blob/abcd1e0f45b73cd4d5c624b5e64870556d8631b0/doghouse/server/doghouse.go#L20-L26 https://github.com/reviewdog/reviewdog/blob/abcd1e0f45b73cd4d5c624b5e64870556d8631b0/doghouse/server/doghouse.go#L216-L231 Maybe a better approach would be to stop appending after the byte threshold is reached: func (ch *Checker) summaryFindings(name string, checks []*filter.FilteredDiagnostic) []string { var lines []string lines = append(lines, "<details>") lines = append(lines, fmt.Sprintf("<summary>%s (%d)</summary>", name, len(checks))) lines = append(lines, "") for _, c := range checks { nextLine := githubutils.LinkedMarkdownDiagnostic(ch.req.Owner, ch.req.Repo, ch.req.SHA, c.Diagnostic) if len(lines) + len(nextLine) >= 65535 { lines = append(lines, "... (Too many findings. Dropped some findings)") break } lines = append(lines, nextLine) } lines = append(lines, "</details>") return lines } I think it's also worth considering, if the text field of the output wouldn't be a better place for all the findings. In the example from the GitHub docs^1, the summary is much shorter. "output": { "title": "Mighty Readme report", "summary": "There are 0 failures, 2 warnings, and 1 notice.", "text": "You may have some misspelled words on lines 2 and 4. You also may want to add a section in your README about how to install your app.", "annotations_count": 2, "annotations_url": "https://api.github.com/repos/github/hello-world/check-runs/4/annotations" }, I'd be willing to open a PR with the proposed fix if there are no objections from a maintainer.
2025-04-01T04:35:21.280734
2024-12-18T23:17:32
2748921647
{ "authors": [ "reysic", "softworkz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10319", "repo": "reysic/AzureKeyVaultExplorer", "url": "https://github.com/reysic/AzureKeyVaultExplorer/issues/6" }
gharchive/issue
One-Click Link Replacement Hi @reysic, thanks a lot for your intention to continue this project. Do you already have a plan for setting up a new OneClick link? sw You’re welcome, @softworkz! And yes, I do intend to continue using ClickOnce for deployment, and I have a plan on how I will set that up. I’ll share more info soon; I’ll leave this issue open to track. Thanks! Awesome, thanks again!
2025-04-01T04:35:21.291047
2024-04-29T11:40:39
2268805414
{ "authors": [ "Aherys", "Packetdancer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10320", "repo": "reznok/GMCAbilitySystem", "url": "https://github.com/reznok/GMCAbilitySystem/pull/56" }
gharchive/pull-request
Simple compability for engine compilation + crash fix added the UPROPERTY Category to allow Plugin compilation with engine Fixed a crash in HandleTaskHeartbeat : If you let a task live long enough, in case of a hitch or a low FPS, the task seem to be destroyed but keeped in the map RunningTasks. The check will avoid the crash, clean the list and throw a log message to avoid error shadowing in future développement. For what it's worth, I'm going to recommend holding off on merging this in; the changes for the crash fix conflict with the rewrite to how attributes are replicated in pull request 55. (The addition of categories to all UPROPERTY entries to allow GMAS to be compiled as an engine plugin is a change still worth merging in separately.) Understood, on the need we could split this pull request in two part. I think we will also do this for future pull request to simplify revision, instead of bundling features/fix/enhancements. Pull request close, switched to dev branch and crash change will be re-implemented with PR55
2025-04-01T04:35:21.293263
2016-05-06T10:26:20
153423258
{ "authors": [ "kacperryniec", "rfennell" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10321", "repo": "rfennell/vNextBuild", "url": "https://github.com/rfennell/vNextBuild/issues/9" }
gharchive/issue
Make the GenerateReleaseNotes task for on non-Windows agents This will require migrating to Node/Javascript. This should not be a major issue for the REST based business logic, but how do we write the template? Currently the template is in effect markdown with embedded PowerShell, could we use the same model or will be need Node specific templates? That would be absolutely awesome! Especially nowadays when Xamarin is becoming more and more popular and VSTO is strongly supporting non-windows builds via for example OS X agents. Have a look at my nww`crossplatform extension. See this blog post for background
2025-04-01T04:35:21.300185
2016-07-25T19:36:35
167447497
{ "authors": [ "Hrxn", "phihag", "sfxworks" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10322", "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/10164" }
gharchive/issue
Youtube-dl bug Trying to download an awesome video. Told me to report it here. youtube-dl https://www.youtube.com/watch?v=fdOcOtPqcIs --verbose [debug] System config: [] [debug] User config: [] [debug] Command-line args: ['https://www.youtube.com/watch?v=fdOcOtPqcIs', '--verbose'] [debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8 [debug] youtube-dl version 2014.08.05 [debug] Python version 2.7.9 - Linux-4.4.13-v7+-armv7l-with-debian-8.0 [debug] Proxy map: {} [youtube] Setting language [youtube] fdOcOtPqcIs: Downloading webpage [youtube] fdOcOtPqcIs: Downloading video info webpage [youtube] fdOcOtPqcIs: Extracting video information [youtube] fdOcOtPqcIs: Encrypted signatures detected. WARNING: unable to extract html5 player; please report this issue on http://yt-dl.org/bug [youtube] {22} signature length 43.40, html5 player None ERROR: Signature extraction failed: Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 479, in _decrypt_signature video_id, player_url, s File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 356, in _extract_signature_function raise ExtractorError('Cannot identify player %r' % player_url) ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vflIB5TLK/base.js'; please report this issue on https://yt-dl.org/bug . Be sure to call youtube-dl with the --verbose flag and include its complete output. Make sure you are using the latest version; type youtube-dl -U to update. ; please report this issue on https://yt-dl.org/bug . Be sure to call youtube-dl with the --verbose flag and include its complete output. Make sure you are using the latest version; type youtube-dl -U to update. Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 479, in _decrypt_signature video_id, player_url, s File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 356, in _extract_signature_function raise ExtractorError('Cannot identify player %r' % player_url) ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vflIB5TLK/base.js'; please report this issue on https://yt-dl.org/bug . Be sure to call youtube-dl with the --verbose flag and include its complete output. Make sure you are using the latest version; type youtube-dl -U to update. Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/youtube_dl/YoutubeDL.py", line 516, in extract_info ie_result = ie.extract(url) File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 170, in extract return self._real_extract(url) File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 860, in _real_extract encrypted_sig, video_id, player_url, age_gate) File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 489, in _decrypt_signature u'Signature extraction failed: ' + tb, cause=e) ExtractorError: Signature extraction failed: Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 479, in _decrypt_signature video_id, player_url, s File "/usr/lib/python2.7/dist-packages/youtube_dl/extractor/youtube.py", line 356, in _extract_signature_function raise ExtractorError('Cannot identify player %r' % player_url) ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vflIB5TLK/base.js'; please report this issue on https://yt-dl.org/bug . Be sure to call youtube-dl with the --verbose flag and include its complete output. Make sure you are using the latest version; type youtube-dl -U to update. ; please report this issue on https://yt-dl.org/bug . Be sure to call youtube-dl with the --verbose flag and include its complete output. Make sure you are using the latest version; type youtube-dl -U to update. [debug] youtube-dl version 2014.08.05 As @Hrxn has correctly identified, your version of youtube-dl is ancient. See our FAQ on how to update.
2025-04-01T04:35:21.316849
2017-06-09T15:08:54
234856414
{ "authors": [ "bentheguitarman", "cyberjacob" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10323", "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/13327" }
gharchive/issue
Hotstar Problem What is the purpose of your issue? [x ] Bug report (encountered problems with youtube-dl) [ ] Site support request (request for adding support for a new site) [ ] Feature request (request for a new functionality) [ ] Question [ ] Other Single video: youtube-dl.exe http://www.hotstar.com/tv/malcolm-in-the-middle/12117/grandma-sues/1000114289 I have been downloading from Hotstar the show "Malcolm in The Middle" and am now encountering a problem this is what comes up in the log Warning: Failed to download m3u8 information: HTTP Error 502: Bad Gateway ERROR: No video formats found; please report this issue on https:yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl - U to update I have updated Youtube-dl and tried an Indian proxy (wasnt needed for some reason) but won't work either way. Please add the -v flag to your command line you run youtube-dl with, copy the whole output and insert it here. C:\Users\Ben\Downloads\Compressed\ffmpeg-20170503-a75ef15-win64-static\bin>youtube-dl.exe -v http://www.hotstar.com/tv/malcolm-in-the-middle/12117/grandma-sues/1000114289 [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: ['-v', 'http://www.hotstar.com/tv/malcolm-in-the-middle/12117/grandma-sues/1000114289'] [debug] Encodings: locale cp1252, fs mbcs, out cp437, pref cp1252 [debug] youtube-dl version 2017.06.05 [debug] Python version 3.4.4 - Windows-10-10.0.15063 [debug] exe versions: ffmpeg N-85750-ga75ef15, ffprobe N-85750-ga75ef15 [debug] Proxy map: {} [HotStar]<PHONE_NUMBER>: Downloading JSON metadata [HotStar]<PHONE_NUMBER>: Downloading JIO JSON metadata [HotStar]<PHONE_NUMBER>: Downloading m3u8 information Im unsure how to copy the rest as the program shuts straight after, I am not the most tech savvy person when it comes to command line stuff This is my BAT file in Notepad, just added the -v youtube-dl.exe -v http://www.hotstar.com/tv/malcolm-in-the-middle/12117/grandma-sues/1000114289 If you add another line at the end of the file with pause (just on it's own, nothing else) the command window should stay open for you to copy from it. It'll wait for you to press any key before finishing the BAT and closing. C:\Users\Ben\Downloads\Compressed\ffmpeg-20170503-a75ef15-win64-static\bin>youtube-dl.exe -v http://www.hotstar.com/tv/malcolm-in-the-middle/12117/grandma-sues/1000114289 [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: ['-v', 'http://www.hotstar.com/tv/malcolm-in-the-middle/12117/grandma-sues/1000114289'] [debug] Encodings: locale cp1252, fs mbcs, out cp437, pref cp1252 [debug] youtube-dl version 2017.06.05 [debug] Python version 3.4.4 - Windows-10-10.0.15063 [debug] exe versions: ffmpeg N-85750-ga75ef15, ffprobe N-85750-ga75ef15 [debug] Proxy map: {} [HotStar]<PHONE_NUMBER>: Downloading JSON metadata [HotStar]<PHONE_NUMBER>: Downloading JIO JSON metadata [HotStar]<PHONE_NUMBER>: Downloading m3u8 information WARNING: Failed to download m3u8 information: HTTP Error 502: Bad Gateway ERROR: No video formats found; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Traceback (most recent call last): File "C:\Users\dst\AppData\Roaming\Build archive\youtube-dl\rg3\tmp6q2mjw76\build\youtube_dl\YoutubeDL.py", line 761, in extract_info File "C:\Users\dst\AppData\Roaming\Build archive\youtube-dl\rg3\tmp6q2mjw76\build\youtube_dl\extractor\common.py", line 433, in extract File "C:\Users\dst\AppData\Roaming\Build archive\youtube-dl\rg3\tmp6q2mjw76\build\youtube_dl\extractor\hotstar.py", line 89, in _real_extract File "C:\Users\dst\AppData\Roaming\Build archive\youtube-dl\rg3\tmp6q2mjw76\build\youtube_dl\extractor\common.py", line 1056, in _sort_formats youtube_dl.utils.ExtractorError: No video formats found; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. C:\Users\Ben\Downloads\Compressed\ffmpeg-20170503-a75ef15-win64-static\bin>pause Press any key to continue . . . I am wondering if perhaps it is an issue on Hotstar side of things as error 502 Bad gateway I am off to bed anyway, hopefully it is a minor issue back again, still have the same problem anyone able to help at all. If it helps I have new info, using express vpn Mumbai server I cant actually watch the show on hotstar either but if I choose the Chenai one I can watch it in browser (firefox) but still wont let me download the stream (error 502) just in a rush as the show I want to download will be gone on june 30th .
2025-04-01T04:35:21.329637
2018-07-30T04:09:55
345604562
{ "authors": [ "Leolexis", "bato3", "dstftw" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10324", "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/17105" }
gharchive/issue
Parse error Make sure you are using the latest version: run youtube-dl --version and ensure your version is 2018.07.29. If it's not, read this FAQ entry and update. Issues with outdated version will be rejected. [x] I've verified and I assure that I'm running youtube-dl 2018.07.29 Before submitting an issue make sure you have: [x] At least skimmed through the README, most notably the FAQ and BUGS sections [x] Searched the bugtracker for similar issues including closed ones [x] Checked that provided video/audio/playlist URLs (if any) are alive and playable in a browser What is the purpose of your issue? [x] Bug report (encountered problems with youtube-dl) [ ] Site support request (request for adding support for a new site) [ ] Feature request (request for a new functionality) [x] Question [ ] Other If the purpose of this issue is a bug report, site support request or you are not completely sure provide the full verbose output as follows: Add the -v flag to your command line you run youtube-dl with (youtube-dl -v <your command line>), copy the whole output and insert it here. It should look similar to one below (replace it with your log inserted between triple ```): [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: ['-v'] [debug] Encodings: locale cp1251, fs mbcs, out cp866, pref cp1251 [debug] youtube-dl version 2018.07.29 [debug] Python version 3.4.4 (CPython) - Windows-10-10.0.17134 [debug] exe versions: none [debug] Proxy map: {} ... <end of log> Description of your issue, suggested solution and other information The issue is I keep getting the parse error with every website I'm trying to use C:\Users\я\Desktop\dl>youtube-dl --all-subs --skip-download --hls-prefer-native http://www.nbc.com/world-of-dance/video/the-qualifiers-5/3749244 [NBC] 3749244: Downloading JSON metadata [ThePlatform] 3749244: Downloading SMIL data ERROR: 3749244: Failed to parse XML (caused by ParseError('not well-formed (invalid token): line 8, column 42',)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. C:\Users\я\Desktop\dl>youtube-dl --all-subs --skip-download https://www.cbs.com/shows/elementary/video/qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C/elementary-meet-your-maker/ [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading XML ERROR: qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Failed to parse XML (caused by ParseError('not well-formed (invalid token): line 8, column 42',)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Any help? Works for me. iff still don't work for you use --write-pages and try find reason in last file: Saving request to: (or paste content here https://pastebin.com/) [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: [u'--all-subs', u'--skip-download', u'--hls-prefer-native', u'http://www.nbc.com/world-of-dance/video/the-qualifiers-5/3749244', u'--proxy', u'***', u'-v'] [debug] Encodings: locale cp1250, fs mbcs, out cp852, pref cp1250 [debug] youtube-dl version 2018.07.29 [debug] Git HEAD: b50478675 [debug] Python version 2.7.14 (CPython) - Windows-10-10.0.17134 [debug] exe versions: ffmpeg 4.0, ffprobe 4.0, phantomjs 2.1.1 [debug] Proxy map: {u'***'} [NBC] 3749244: Downloading JSON metadata [ThePlatform] 3749244: Downloading SMIL data [ThePlatform] 3749244: Downloading m3u8 information [ThePlatform] 3749244: Downloading JSON metadata [debug] Default format spec: bestvideo+bestaudio/best [info] Writing video subtitles to: The Qualifiers 5-3749244.en.tt [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: [u'--all-subs', u'--skip-download', u'--proxy', u'***', u'-v', u'https://www.cbs.com/shows/elementary/video/qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C/elementary-meet-your-maker/'] [debug] Encodings: locale cp1250, fs mbcs, out cp852, pref cp1250 [debug] youtube-dl version 2018.07.29 [debug] Git HEAD: b50478675 [debug] Python version 2.7.14 (CPython) - Windows-10-10.0.17134 [debug] exe versions: ffmpeg 4.0, ffprobe 4.0, phantomjs 2.1.1 [debug] Proxy map: {***} [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading XML [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading StreamPack SMIL data [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading m3u8 information [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading RTMP SMIL data [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading HLS_AES SMIL data [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading m3u8 information WARNING: Failed to download m3u8 information: HTTP Error 403: Forbidden [CBS] qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C: Downloading JSON metadata [debug] Default format spec: bestvideo+bestaudio/best [info] Writing video subtitles to: Elementary - Meet Your Maker-qzV5zwpnktuopB_ERsrYvoUL4GGFCX9C.en.ttml Post the output with --dump-pages.
2025-04-01T04:35:21.333094
2015-07-25T07:39:12
97205547
{ "authors": [ "Pat38", "dstftw" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10325", "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/6355" }
gharchive/issue
Support for extractor "Channel 4" Hi everybody, When I try to download this show : youtube-dl --all-subs http://www.channel4.com/programmes/dispatches/on-demand/59665-005 the program returns this error : [generic] 59665-005: Requesting header WARNING: Falling back on generic information extractor. [generic] 59665-005: Downloading webpage [generic] 59665-005: Extracting information ERROR: Unsupported URL: http://www.channel4.com/programmes/dispatches/on-demand/59665-005 Do you plan to add "Channel 4" support to the list of extractors ? Thank you in advance for your answer, Yours faithfully Patrick. Duplicate of #2050.
2025-04-01T04:35:21.341477
2016-06-17T12:30:46
160877698
{ "authors": [ "RingoTheDog", "TRox1972", "rrthomas", "yan12125" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10326", "repo": "rg3/youtube-dl", "url": "https://github.com/rg3/youtube-dl/issues/9814" }
gharchive/issue
Time to stop describing youtube-dl as "small"? Make sure you are using the latest version: run youtube-dl --version and ensure your version is 2016.06.16. If it's not read this FAQ entry and update. Issues with outdated version will be rejected. [x] I've verified and I assure that I'm running youtube-dl 2016.06.16 Before submitting an issue make sure you have: [x] At least skimmed through README and most notably FAQ and BUGS sections [x] Searched the bugtracker for similar issues including closed ones What is the purpose of your issue? [ ] Bug report (encountered problems with youtube-dl) [ ] Site support request (request for adding support for a new site) [ ] Feature request (request for a new functionality) [ ] Question [x] Other Description of your issue, suggested solution and other information Is it time to stop describing youtube-dl as "small"? I counted the current version with cloc, as installed from the webupd8 PPA, and it comes to a shade under 80kLOC in just under 650 files. I don't think that qualifies as small, either in lines of code or in complexity of installation (e.g. cloc itself is about 10kLOC, but ships as a single Perl file, so it's sort of small). Suggested change: remove the word "small" from the project description on GitHub, from https://rg3.github.io/youtube-dl/ and from any other corresponding places. I note that the README in particular does not seem (any more?) to describe youtube-dl as "small". youtube-dl is great; thanks for working on it! Good idea! Good point. On http://rg3.github.io/youtube-dl/ it says that youtube-dl is a small command-line program to download videos from YouTube.com and a few more sites. Maybe change 'few' to 'many' or something. change small to awesome ;-)
2025-04-01T04:35:21.343398
2015-07-02T09:25:50
92592684
{ "authors": [ "rgbkrk", "vmalyi" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10327", "repo": "rgbkrk/atom-script", "url": "https://github.com/rgbkrk/atom-script/issues/494" }
gharchive/issue
Can't enter values through console when using raw_input() in Python I have a python script which processes user input with help of following code block: def process_user_input(): print('Enter text') sys.stdout.flush() user_input = raw_input() ... When I start this script with help of CTRL+SHIFT+B in Atom editor, it runs but I'm unable to enter any values in the console. Environment: Win7x64 Atom 1.0.0 <EMAIL_ADDRESS> We don't handle STDIN sadly. We'd need someone to add support. Closing as duplicate of #73, #280 #385.
2025-04-01T04:35:21.370043
2022-11-12T20:07:23
1446586673
{ "authors": [ "NA3" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10328", "repo": "rgl/windows-vagrant", "url": "https://github.com/rgl/windows-vagrant/issues/25" }
gharchive/issue
error when attempting to rsync a synced folder BOX successfully built! vagrant box add -f windows-2019-amd64 windows-2019-amd64-vsphere.box ==> box: Box file was not detected as metadata. Adding it directly... ==> box: Adding box 'windows-2019-amd64' (v0) for provider: box: Unpacking necessary files from: file:///home/goad/windows-vagrant/windows-2019-amd64-vsphere.box ==> box: Successfully added box 'windows-2019-amd64' (v0) for 'vsphere'! vagrant box list windows-2019-amd64 (vsphere, 0) ❯ cd example ❯ vagrant up ==> vagrant: You have requested to enabled the experimental flag with the following features: ==> vagrant: ==> vagrant: Features: typed_triggers ==> vagrant: ==> vagrant: Please use with caution, as some of the features may not be fully ==> vagrant: functional yet. Bringing machine 'default' up with 'vsphere' provider... ==> default: Setting custom memory: 4096 ==> default: Setting custom cpu count: 4 ==> default: Calling vSphere CloneVM with the following settings: ==> default: -- Template VM: Kore/vm/Modeles/Packer/windows-2019-amd64-vsphere ==> default: -- Target VM: Kore/vm/GOAD/windows-2019-vagrant-example ==> default: New virtual machine successfully cloned ==> default: Waiting for the machine to report its IP address... default: Timeout: 240 seconds default: IP: <IP_ADDRESS> ==> default: Waiting for machine to boot. This may take a few minutes... default: WinRM address: <IP_ADDRESS>:5985 default: WinRM username: vagrant default: WinRM execution_time_limit: PT2H default: WinRM transport: negotiate ==> default: Machine booted and ready! ==> default: Setting hostname... ==> default: Waiting for machine to reboot... ==> default: Rsyncing folder: /home/goad/windows-vagrant/example/ => /vagrant There was an error when attempting to rsync a synced folder. Please inspect the error message below for more info. Host path: /home/goad/windows-vagrant/example/ Guest path: /vagrant Command: "rsync" "--verbose" "--archive" "--delete" "-z" "--copy-links" "--no-owner" "--no-group" "-e" "ssh -p 22 -o LogLevel=FATAL -o ControlMaster=auto -o ControlPath=/tmp/vagrant-rsync-20221112-199115-1lwg5d2 -o ControlPersist=10m -o IdentitiesOnly=yes -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -i '/home/goad/.vagrant.d/insecure_private_key'" "--exclude" ".vagrant/" "/home/goad/windows-vagrant/example/" "vagrant@<IP_ADDRESS>:/vagrant" Error: rsync: connection unexpectedly closed (0 bytes received so far) [sender] rsync error: unexplained error (code 255) at io.c(235) [sender=3.1.3] can i skip the folder sync? Regard. i got it config.vm.synced_folder '.', '/vagrant', disabled: true
2025-04-01T04:35:21.382631
2017-03-31T15:46:17
218541518
{ "authors": [ "GregRocket", "loftgren" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10332", "repo": "rgraciano/echo-sonos", "url": "https://github.com/rgraciano/echo-sonos/issues/145" }
gharchive/issue
Help finding options.example.js Can someone please tell me where to find the file options.example.js? Thanks very much in advance! Sorry, I actually know where options.example.js is located... what i really need is to be able to find options.js. I assume I am looking locally? Step 5 is killing me! Any help is greatly appreciated! @loftgren Just rename the file options.example.js to options.js
2025-04-01T04:35:21.390106
2022-07-30T09:27:36
1322946786
{ "authors": [ "cmccomb", "schungx" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10334", "repo": "rhaiscript/rhai-sci", "url": "https://github.com/rhaiscript/rhai-sci/issues/5" }
gharchive/issue
Checking the first element of array In some of the array functions, I see code that checks the first element and then assumes the rest of the array to be of the same type (usually by checking whether we're working with an INT or FLOAT array): if arr[0].is::<INT>() { : } else if arr[0].is::<FLOAT>() { : } ... This is perhaps necessary for statically-typed languages, but Rhai is dynamic so this restriction is rarely necessary as the users may expect to be able to freely mix INT with FLOAT. For example, we can consider to do this instead: Assume the array is an INT if and only if all elements are INT. Otherwise, convert to FLOAT array. This is consistent with treatment of FLOAT in Rhai, which is: a + b is only INT if and only if a and b are both INT. let (ints, floats) = arr.iter().fold((0, 0), |(i, f), x| { if x.is::<INT>() { (i + 1, f) } else if x.is::<FLOAT>() { (i, f + 1) } else { (i, f) } }); if ints + floats < arr.len() { // raise error as some elements are not numeric } else if floats == 0 { let i_vec = arr.iter().map(|x| x.as_int().unwrap()).collect::<Vec<_>>(); } else { let f_vec = arr .iter() .map(|x| { if x.is::<FLOAT>() { x.as_float().unwrap() } else { x.as_int().unwrap() as FLOAT } }) .collect::<Vec<_>>(); } Thanks for providing the sample code! This check-convert block is one of several that shows up frequently, so I will address this over time with some other updates as well. A slight alternative can be used to early return when you encounter an error or a FLOAT... let is_all_int = arr .iter() .try_fold((), |_, x| { if x.is::<FLOAT>() { Err(false) // Err(false) = some FLOAT } else if x.is::<INT>() { Ok(()) // Ok(()) = all INT } else { Err(true) // Err(true) = type error } }) .or_else(|err| match err { Ok(()) => true, Err(false) => false, Err(true) => Err(......), })?;
2025-04-01T04:35:21.391701
2016-10-28T15:07:37
185947379
{ "authors": [ "serj-lotutovici", "vanniktech" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10335", "repo": "rharter/auto-value-moshi", "url": "https://github.com/rharter/auto-value-moshi/issues/67" }
gharchive/issue
Consider moving @MoshiAdapterFactory into it's own artifact It'd be awesome if @MoshiAdapterFactory would be in it's own artifact so that one does not have to use the following: provided 'com.ryanharter.auto.value:auto-value-moshi:0.4.1' to add the annotation to the classpath. The big disadvantage is that all of the other classes are also added to the class path (AutoValue components, possible Guava etc). Butterknife does something similar. There's an annotation artefact, that is shared between the runtime library and the annotation processor.
2025-04-01T04:35:21.393997
2015-05-05T11:26:35
73287147
{ "authors": [ "miminar", "rhatdan" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10336", "repo": "rhatdan/docker", "url": "https://github.com/rhatdan/docker/pull/51" }
gharchive/pull-request
Fixed push of unqualified registry Local image with unqualified name had to be fully qualified before a push to default registry. Moved check for official repository to daemon side because client doesn't know which registry is the default. Merged.
2025-04-01T04:35:21.404886
2024-04-19T04:39:13
2252069994
{ "authors": [ "mecsu", "rhertogh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10337", "repo": "rhertogh/yii2-oauth2-server", "url": "https://github.com/rhertogh/yii2-oauth2-server/issues/21" }
gharchive/issue
Support Token Revocation Is your feature request related to a problem? Please describe. I need an endpoint to revoke token Describe the solution you'd like Follow this spec RFC7009 to implement revocation endpoint Describe alternatives you've considered I have to extend this controller rhertogh\Yii2Oauth2Server\controllers\web\Oauth2ServerController, then implement new action Revoke Hi @mecsu, Thanks for the suggestion, please have a look at 305862a6 Could you test it on the master branch? The endpoint is available on /oauth2/revoke by default. Hi @rhertogh, Thanks for implementing this feature. I have tested call endpoint /oauth2/revoke to revoke refresh_token but it return error: { "error": "Bad Request", "error_description": "yii\\web\\BadRequestHttpException: Unable to resolve thetokenparameter to a valid token type ..." } And error message in log: [error][rhertogh\Yii2Oauth2Server\controllers\web\server\Oauth2RevokeAction::parseTokenAsRefreshToken] The client specified thetoken_type_hintas "refresh_token", however the server is unable to parse thetokenas such: Defuse\Crypto\Exception\WrongKeyOrModifiedCiphertextException: Ciphertext has invalid hex encoding. in .../vendor/defuse/php-encryption/src/Crypto.php:311 Can you help check this issue? @mecsu Could you send me the exact request body (for security, make sure the tokens in there are no longer valid)? Hi @rhertogh, I tested revoke refresh_token successfully. But I have tried revoke access_token, it return error: { "error": "Bad Request", "error_description": "yii\\web\\BadRequestHttpException: Unable to resolve thetokenparameter to a valid token type ..." } I have debugged, found issue within function rhertogh\Yii2Oauth2Server\controllers\web\server\Oauth2RevokeAction::parseTokenAsAccessToken, it return ["clientIdentifier" => null]. protected function parseTokenAsAccessToken(Oauth2Module $module, string $token, string $tokenTypeHint) { try { ... $accessTokenClaims = $accessToken->claims(); $accessTokenIdentifier = $accessTokenClaims->get('jti'); // issue with below statement, cause no have claim "client_id" $clientIdentifier = $accessTokenClaims->get('client_id'); return [ 'clientIdentifier' => $clientIdentifier, 'accessTokenIdentifier' => $accessTokenIdentifier ]; } catch (\Throwable $e) { if ($tokenTypeHint === 'access_token') { Yii::getLogger()->log( 'The client specified the `token_type_hint` as "access_token", however the server is unable to parse the `token` as such: ' . $e, $module->getElaboratedHttpClientErrorsLogLevel(), __METHOD__ ); } unset($e); } return null; } @mecsu The client_id is a new claim in the access token. So any old access token won't include it. Does it work with a newly generated access token? @mecsu The client_id is a new claim in the access token. So any old access token won't include it. Does it work with a newly generated access token? Yes, it work with new access token. It's great. Many thanks.
2025-04-01T04:35:21.436332
2016-01-12T09:13:13
126131619
{ "authors": [ "hekonsek" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10338", "repo": "rhiot/rhiot", "url": "https://github.com/rhiot/rhiot/issues/386" }
gharchive/issue
Add support for self-documenting commands Each CMD command should provide a help information about itself. Done.
2025-04-01T04:35:21.457384
2022-09-10T19:17:54
1368709287
{ "authors": [ "Azuka", "hsblhsn", "riandyrn" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10341", "repo": "riandyrn/otelchi", "url": "https://github.com/riandyrn/otelchi/issues/10" }
gharchive/issue
Add filter for ignoring certain endpoints Thanks for this project: I've been using it for a couple months now. Would it be possible to ignore a healthcheck endpoint due to the frequency of calls? Elsewhere it seems the pattern is to accept function filters that return a boolean: https://github.com/open-telemetry/opentelemetry-go-contrib/issues/896 I'd be happy to open a pull request if you think the idea has merit. Hello, @Azuka Yeah, I think this would be great. Looking forward to your PR. Hey @Azuka, Is there any update on that? If no, I can open one 😄 Thanks 🙏
2025-04-01T04:35:21.508765
2022-11-02T06:37:33
1432581148
{ "authors": [ "richard1122" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10343", "repo": "richard1122/chromium", "url": "https://github.com/richard1122/chromium/pull/12" }
gharchive/pull-request
WebMessagePayloadView 3998611: WIP: Refactor WebMessagePayload to WebMessagePayloadView in blink | https://chromium-review.googlesource.com/c/chromium/src/+/3998611 Replaced by #13
2025-04-01T04:35:21.512183
2022-06-06T06:52:17
1261429005
{ "authors": [ "conanliuhuan", "pitw", "richardtallent", "tony19", "zhanglongqi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10344", "repo": "richardtallent/vite-plugin-singlefile", "url": "https://github.com/richardtallent/vite-plugin-singlefile/issues/33" }
gharchive/issue
Error [ERR_REQUIRE_ESM]: require() of ES Module ...\node_modules\vite-plugin-singlefile\dist\index.js from ...\vite.config.js not supported. add "type": "module", to your package.json Duplicate of #23 A workaround is to install 0.7.1: npm i -D<EMAIL_ADDRESS> Please try 0.10.0-beta and let me know if it helps. It should allow use in both CommonJS and ESM environments. Got same error; With 0.10.0-beta it works 👍 0.10.0 has been released,
2025-04-01T04:35:21.533516
2021-10-11T14:45:37
1022796914
{ "authors": [ "DISAPPEARED13" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10345", "repo": "richzhang/PerceptualSimilarity", "url": "https://github.com/richzhang/PerceptualSimilarity/issues/82" }
gharchive/issue
Dropout or not if i use this loss function to calculate my model loss, should I set 'use_dropout' as False to get stable output? Thanks a lot. : ) use_dropout should be True, since it was trained with dropout. Setting it in eval mode will make things stable. Thanks for replying!! :D I use this for my model and it doesn't decrease, now I guess maybe I didn't set loss_model.eval() and calculate it properly. And.. I am sorry that I've got another problem to bother you, if I use 3D patch data to calculate it for example [2, 1, 32, 256, 256] should i split it into a shape like [1, 256, 256] of 2*32 arraies, get the loss each little array and sum it up? but I found this way will cost lots of GPU memory : (
2025-04-01T04:35:21.537564
2022-04-22T07:57:03
1211964090
{ "authors": [ "Zhairest", "joannanananana" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10346", "repo": "richzhang/colorization-pytorch", "url": "https://github.com/richzhang/colorization-pytorch/issues/25" }
gharchive/issue
'SIGGRAPHGenerator' object has no attribute 'model' when running on MAC(python3.5.4),I encountered the following problem:'SIGGRAPHGenerator' object has no attribute 'model' I wonder how to fix it? Traceback (most recent call last): File "test.py", line 40, in model.setup(opt) File "/Users/joanna/Desktop/colorization/colorization-pytorch-master/models/base_model.py", line 42, in setup self.load_networks(opt.which_epoch) File "/Users/joanna/Desktop/colorization/colorization-pytorch-master/models/base_model.py", line 136, in load_networks self.__patch_instance_norm_state_dict(state_dict, net, key.split('.')) File "/Users/joanna/Desktop/colorization/colorization-pytorch-master/models/base_model.py", line 116, in __patch_instance_norm_state_dict self.__patch_instance_norm_state_dict(state_dict, getattr(module, key), keys, i + 1) File "/Users/joanna/opt/anaconda3/envs/color/lib/python3.5/site-packages/torch/nn/modules/module.py", line 576, in getattr type(self).name, name)) AttributeError: 'SIGGRAPHGenerator' object has no attribute 'model' me too. did you make it out in the end?
2025-04-01T04:35:21.542041
2023-07-24T09:40:53
1818037916
{ "authors": [ "ricklamers", "soulteary" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10347", "repo": "ricklamers/gpt-code-ui", "url": "https://github.com/ricklamers/gpt-code-ui/issues/62" }
gharchive/issue
OpenAI API generated code depends on automatic installation ⚠️ Please check that this feature request hasn't been suggested before. [X] I searched previous Ideas in Discussions didn't find any similar feature requests. [X] I searched previous Issues didn't find any similar feature requests. 🔖 Feature description After the code is generated by the OpenAI API, if there is no local dependency on Python or other languages, it will be automatically installed. ✔️ Solution You only need to make some simple adjustments to the code, use pre-script or monitor whether it is an execution error due to lack of dependencies during execution, install it, and try again. ❓ Alternatives No response 📝 Additional Context No response Acknowledgements [X] My issue title is concise, descriptive, and in title casing. [X] I have searched the existing issues to make sure this feature has not been requested yet. [X] I have provided enough information for the maintainers to understand and evaluate this request. Agree this would be cool to add. Duplicate of https://github.com/ricklamers/gpt-code-ui/discussions/6
2025-04-01T04:35:21.561192
2019-02-14T17:36:57
410413706
{ "authors": [ "halsten", "kainjow", "karstenBriksoft" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10348", "repo": "ridiculousfish/HexFiend", "url": "https://github.com/ridiculousfish/HexFiend/issues/189" }
gharchive/issue
Integrate Kaitai with HexFiend for Templates Any chance you can integrate Kaitai to improve on template writing and also the extensibility of HexFiend. It's a nice utility that I use a lot, but this Tcl choice of yours is rather interesting. You know Python also ships with macOS and so does a lot of other things too. Would be nice to see this added to the tool. --halsten It's been brought up before, see https://github.com/ridiculousfish/HexFiend/issues/89#issuecomment-354649108 It's also GPL3. Everyone’s going to have their favorite language. What made Tcl stand out was it’s simple syntax. A lot of templates don’t look like source code. It also treats everything as a string. Python would require conversions, and doesn’t appear to support 64-bit unsigned integers out of the box. Also I’m unsure if the built-in Python in macOS can be directly embedded. Saying all that, if you want to add support for a new language that’s already part of the Mac I wouldn’t be applaud to it. The code is also designed to separate the template language implementation from the language itself. Kaitai defines a Yaml based structure for describing binary data. There's also a kaitai structure compiler that converts the structures into code that can be integrated with other software that needs to parse the particular kinds of data. That last part is not very interesting for HexFiend, but the first part is. I've created a Yaml decoder (based on libYaml, which can be installed via homebrew: brew install libyaml). If present, the Kaitai TemplateController will be used for .ksy files in the template folder. It's not fully working yet, but it's a nice start (see my fork at: https://github.com/karstenBriksoft/HexFiend) While I find your work commendable, I wonder if instead of rewriting all that logic, it'd be better to add support to HexFiend to somehow support one of the languages that Kaitai already supports, either through IPC or direct scripting. Then there would be no duplicate logic, just bindings to HF. I haven't looked to see if this would even be doable though.
2025-04-01T04:35:21.569150
2021-08-12T14:49:44
968980360
{ "authors": [ "DCCoder90", "rafaelcanical", "riedeljan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10349", "repo": "riedeljan/insomnia-universal-git", "url": "https://github.com/riedeljan/insomnia-universal-git/issues/6" }
gharchive/issue
Buggy branch fetching When the setup ist completed first the modal has to be submitted and re-opened before a branch can be selected. Same bug occurs when creating a new branch from within the setup. If you want I can tackle this. Have some time this weekend that I'll be wrapping up my PR, I can include a fix in for this as well. That‘s awesome, but I‘ve already got the fix in place, just need to finalize it ☺️ Von meinem iPhone gesendet Am 13.08.2021 um 13:22 schrieb Ernest Mallett @.***>:  If you want I can tackle this. Have some time this weekend that I'll be wrapping up my PR, I can include a fix in for this as well. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android. How is the status on this? Seems like dropdown with branches never gets any branch humm Something I can help with? Hi! Didn't really work on this for a while since and would need to have a look. Last time I checked branches fetched fine 🤔 Can you describe what exactly happens on your end and possibly provide output from the dev console? @rafaelcanical Yes sure :) First I opened GitLab - Setup and set all my settings. But dropdown doesn't fill with branches. This is what console returns: On network tab, this request https://gitlab.com/users/sign_in returns the following:
2025-04-01T04:35:21.643810
2023-11-19T00:12:49
2000631875
{ "authors": [ "MalaySheth", "nehahimane" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10352", "repo": "rilling/OpenTracksConcordia", "url": "https://github.com/rilling/OpenTracksConcordia/issues/223" }
gharchive/issue
Subtask 6: Implement the logic to delay the tracking for selected time. Change the code to implement the delay for selected time in seconds. https://github.com/rilling/OpenTracksConcordia/issues/184 Duplicates #221
2025-04-01T04:35:21.659883
2024-07-02T18:48:04
2386920896
{ "authors": [ "GuillaumeGomez", "softdevca" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10353", "repo": "rinja-rs/rinja", "url": "https://github.com/rinja-rs/rinja/pull/41" }
gharchive/pull-request
Specific errors when macro, import or extends used outside top level More specific errors when macro, import or extends are used outside top level. This was helpful when finding a stray {% import %}. Looks good, thanks! Please add UI tests (in testing/tests/ui). I opened https://github.com/rinja-rs/rinja/pull/42, so once merged, all you'll have to do is update UI test added (and also improve the code added by the PR). See #42.
2025-04-01T04:35:21.662794
2018-02-15T12:49:37
297434389
{ "authors": [ "Omranic", "saqueib" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10354", "repo": "rinvex/addresses", "url": "https://github.com/rinvex/addresses/issues/16" }
gharchive/issue
Unable to install on Laravel 5.6.3 Its looks great package but I wasn't able to install on my laravel app. Using version ^0.0.2 for rinvex/addresses ./composer.json has been updated Loading composer repositories with package information Updating dependencies (including require-dev) Your requirements could not be resolved to an installable set of packages. Problem 1 - Installation request for rinvex/addresses ^0.0.2 -> satisfiable by rinvex/addresses[v0.0.2]. - rinvex/addresses v0.0.2 requires rinvex/cacheable dev-develop -> satisfiable by rinvex/cacheable[dev-develop] but these conflict with your requirements or minimum-stability. Installation failed, reverting ./composer.json to its original content. You can now use https://github.com/rinvex/addresses/tree/v0.0.3 which has full support for Laravel v5.6+ @Omranic thanks but I am still getting following: Updating dependencies (including require-dev) Your requirements could not be resolved to an installable set of packages. Problem 1 - Installation request for rinvex/addresses ^0.0.3 -> satisfiable by rinvex/addresses[v0.0.3]. - rinvex/addresses v0.0.3 requires rinvex/cacheable dev-develop -> satisfiable by rinvex/cacheable[dev-develop] but these conflict with your requirements or minimum-stability. Here is my composer "require": { "php": ">=7.0.0", "conceptlz/thunderbolt": "1.0.0", "fideloper/proxy": "~4.0", "gerardojbaez/laraplans": "^2.1", "laravel/framework": "5.6.*", "laravel/passport": "~5.0", "laravel/tinker": "~1.0", "laravelcollective/html": "^5.5", "spatie/laravel-activitylog": "^2.5", "yajra/laravel-datatables": "^1.0", "rinvex/addresses": "^0.0.3" }, "require-dev": { "filp/whoops": "~2.0", "fzaninotto/faker": "~1.4", "mockery/mockery": "~1.0", "phpunit/phpunit": "~7.0" },
2025-04-01T04:35:21.666402
2021-06-27T15:36:57
930959003
{ "authors": [ "TitanFighter", "albizeka", "alfanzain", "amanpatel", "beyondverage0908", "micdobro", "rifton007", "sharkfin009" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10355", "repo": "riophae/vue-treeselect", "url": "https://github.com/riophae/vue-treeselect/issues/469" }
gharchive/issue
Is this project no longer maintained? Is this project no longer maintained? No new release, no accepted PRs, no responses from @riophae. vue-treeselect is a fantastic project - it would be a big loss if it was left to die... :( you are correct - no longer maintained It's very annoying to use a package and and see it is no longer mainained. A simple message in readme file "this pacakge is not maintained" would be the minimum. Unfortunately, this is often happening in ecosystem of Vue, too many package dead. please someone give alternative of this package I've found this one. https://www.npmjs.com/package/vue3-treeselect the vue3-treeselect is alse not maintained I've forked this package and updated a bit surrounding warnings and errors. It works for me (at the moment), but don't expect super things. thanks @amanpatel I have it running here @amanpatel did u test it, also how do we install this fork of yours?
2025-04-01T04:35:21.734422
2023-04-26T17:25:51
1685435138
{ "authors": [ "allenjbaum", "ssecatchseagate" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10356", "repo": "riscv-non-isa/riscv-arch-test", "url": "https://github.com/riscv-non-isa/riscv-arch-test/issues/337" }
gharchive/issue
Incorrect macro for register save in arch_test.h Error: unrecognized opcode `sigupd( x1,x1)' arch_test.h: .macro RVTEST_SAVE_GPRS BASEREG REG_SV_ADDR // optionally save GPRs .option push .option norvc .set offset,0 LA( \BASEREG, \REG_SV_ADDR) //this destroys basereg, but saves rest SIGUPD( \BASEREG, x1) ... Changing these to RVTEST_SIGUPD allowed compile to proceed Sigh - if it isn't tested, it's borken. Will be fixed in the next merge. Because registers may hold addresses that will different between DUT and reference models, we expect that this will only be used to debug failing code. We could add an extra parameter to this to mark which registers do/don't hold addresses, and skip storing if its an address to make it more useful On Wed, Apr 26, 2023 at 10:26 AM ssecatchseagate @.***> wrote: Error: unrecognized opcode `sigupd( x1,x1)' arch_test.h: .macro RVTEST_SAVE_GPRS BASEREG REG_SV_ADDR // optionally save GPRs .option push .option norvc .set offset,0 LA( \BASEREG, \REG_SV_ADDR) //this destroys basereg, but saves rest SIGUPD( \BASEREG, x1) ... Changing these to RVTEST_SIGUPD allowed compile to proceed — Reply to this email directly, view it on GitHub https://github.com/riscv-non-isa/riscv-arch-test/issues/337, or unsubscribe https://github.com/notifications/unsubscribe-auth/AHPXVJSBIT2PQWW5FTBAPRDXDFLC3ANCNFSM6AAAAAAXMX4BOA . You are receiving this because you are subscribed to this thread.Message ID: @.***> Hi Allen, I do think this is useful. While, all these in theory should get echo'd into the signature file already as part of the regular checking, an additional check is useful that other registers didn't corrupted, since it's "free".* And if I'm going to be debugging a test by putting this in halfway, I don't want false negatives if I'm already debugging stuff, so I wanted it working already for the "easy case" *Locally, the past version, I had masked out any registers that held test only values to ensure that they matched. -Stacey From: Allen Baum @.> Sent: Wednesday, April 26, 2023 4:40 PM To: riscv-non-isa/riscv-arch-test @.> Cc: Stacey Secatch @.>; Author @.> Subject: Re: [riscv-non-isa/riscv-arch-test] Incorrect macro for register save in arch_test.h (Issue #337) This message has originated from an External Source. Please use proper judgment and caution when opening attachments, clicking links, or responding to this email. Sigh - if it isn't tested, it's borken. Will be fixed in the next merge. Because registers may hold addresses that will different between DUT and reference models, we expect that this will only be used to debug failing code. We could add an extra parameter to this to mark which registers do/don't hold addresses, and skip storing if its an address to make it more useful On Wed, Apr 26, 2023 at 10:26 AM ssecatchseagate @.***> wrote: Error: unrecognized opcode `sigupd( x1,x1)' arch_test.h: .macro RVTEST_SAVE_GPRS BASEREG REG_SV_ADDR // optionally save GPRs .option push .option norvc .set offset,0 LA( \BASEREG, \REG_SV_ADDR) //this destroys basereg, but saves rest SIGUPD( \BASEREG, x1) ... Changing these to RVTEST_SIGUPD allowed compile to proceed — Reply to this email directly, view it on GitHub https://github.com/riscv-non-isa/riscv-arch-test/issues/337, or unsubscribe https://github.com/notifications/unsubscribe-auth/AHPXVJSBIT2PQWW5FTBAPRDXDFLC3ANCNFSM6AAAAAAXMX4BOA . You are receiving this because you are subscribed to this thread.Message ID: @.***> — Reply to this email directly, view it on GitHubhttps://github.com/riscv-non-isa/riscv-arch-test/issues/337#issuecomment-1524126284, or unsubscribehttps://github.com/notifications/unsubscribe-auth/A3R2BO4US2VVYK55CFZ7TCLXDGP7BANCNFSM6AAAAAAXMX4BOA. You are receiving this because you authored the thread.Message ID: @.***> Seagate Internal
2025-04-01T04:35:21.738655
2023-09-12T06:06:58
1891738088
{ "authors": [ "KYG-yaya573142", "KaiYG", "cmuellner", "kito-cheng" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10357", "repo": "riscv-non-isa/riscv-c-api-doc", "url": "https://github.com/riscv-non-isa/riscv-c-api-doc/issues/51" }
gharchive/issue
Should the macro __riscv_c and __riscv_compressed defined for Zca? Should the macro __riscv_c (and __riscv_compressed) defined when only Zca is enabled? Also, seems GCC and Clang behave differently in this case, e.g., for -march=rv32i_zca, GCC: undefined __riscv_c, while defined __riscv_compressed LLVM: undefined both. Related questions: https://github.com/riscv-non-isa/riscv-asm-manual/issues/91 https://github.com/riscv/riscv-isa-manual/pull/1121 __riscv_compressed was deprecated, but I guess we should consider redefine that as defined if any extension provided compressed instruction is available? We have C and Zca now, but not guarantee we won't have more independent Zc* extensions . cc @a4lg As the Zc spec v1.0.4-2 added the rule that C implies Zca, I think existence of the compressed instructions can be reported by __riscv_zca now. Thanks for the update and closing the ticket!
2025-04-01T04:35:21.752089
2023-05-29T02:18:04
1729883262
{ "authors": [ "mjosaarinen", "nibrunieAtSi5" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10358", "repo": "riscv/riscv-crypto", "url": "https://github.com/riscv/riscv-crypto/pull/325" }
gharchive/pull-request
[code-samples] Fixing links to Rivos' repo links where not pointing to the correct branch It would be good if eventually the code samples could be built from code in riscv official repos. @kdockser , @egouriou-rivos could one of you review / approve this PR to fix some links ? Apologies for ignoring this review request. No worries, maybe some of those links should now be updated to point towards riscv upstream repos rather than rivos' ? Thanks. Update the links to the last good ones.
2025-04-01T04:35:21.797430
2024-06-08T10:04:29
2341582099
{ "authors": [ "grabbou" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10359", "repo": "rise-tools/rise-tools", "url": "https://github.com/rise-tools/rise-tools/pull/84" }
gharchive/pull-request
refactor: join playground and playground-mobile into one package Gets rid of separate package, removes some legacy code. Fixed a few bugs Remove patch-package as Nate fixed upstream Tamagui issue (thank you <3) Refactored storage management Thanks! Merging and going to send two follow-ups tomorrow: rename packages to rise-tools create "experimental" section (after we merge models - to avoid conflicts)
2025-04-01T04:35:21.803944
2022-10-26T08:56:12
1423671363
{ "authors": [ "TennyZhuang", "hengm3467", "tabVersion" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10360", "repo": "risingwavelabs/risingwave", "url": "https://github.com/risingwavelabs/risingwave/issues/6042" }
gharchive/issue
Tracking: Watermark RFC PR: risingwavelabs/rfcs#2 Implementation design doc: https://github.com/risingwavelabs/rfcs/blob/46123694776083788a08b880ab1fa27eda542b76/rfcs/0016-watermark-operators-explained.md Source [x] #6952 Frontend [x] #6704 [x] #6070 [x] Derive watermark columns [ ] #6452 Stream [x] #6043 [x] #6046 [x] #6060 [x] #6050 [x] #8444 [x] Watermark derivation in window functions [x] #6268 [x] #6472 [x] #6112 [x] #8106 [x] #6813 [x] #7811 [x] #8375 After some discussions, we can also clean the state in DynamicFilter executors. will we handle watermark events from the upstream in CDC scenes? @TennyZhuang Do you think we can prepare external docs about watermark?
2025-04-01T04:35:21.815442
2023-11-09T08:59:37
1985143379
{ "authors": [ "Rossil2012", "yezizp2012" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10361", "repo": "risingwavelabs/risingwave", "url": "https://github.com/risingwavelabs/risingwave/pull/13341" }
gharchive/pull-request
feat: support ALTER [TABLE | [MATERIALIZED] VIEW | SOURCE | SINK | CONNECTION | FUNCTION] <name> SET SCHEMA <schema_name> syntax I hereby agree to the terms of the RisingWave Labs, Inc. Contributor License Agreement. What's changed and what's your intention? Resolve #13159 [x] ALTER TABLE [x] ALTER MATERIAL VIEW [x] ALTER VIEW [x] ALTER [SOURCE | SINK] [x] ALTER FUNCTION [x] ALTER CONNECTION [x] Tests [x] Tests for ALTER [FUNCTION] Checklist [x] I have written necessary rustdoc comments [ ] I have added necessary unit tests and integration tests [ ] I have added fuzzing tests or opened an issue to track them. (Optional, recommended for new SQL features #7934). [ ] My PR contains breaking changes. (If it deprecates some features, please create a tracking issue to remove them in the future). [x] All checks passed in ./risedev check (or alias, ./risedev c) [ ] My PR changes performance-critical code. (Please run macro/micro-benchmarks and show the results.) [ ] My PR contains critical fixes that are necessary to be merged into the latest release. (Please check out the details) Documentation [x] My PR needs documentation updates. (Please use the Release note section below to summarize the impact on users) Release note If this PR includes changes that directly affect users or other significant modifications relevant to the community, kindly draft a release note to provide a concise summary of these changes. Please prioritize highlighting the impact these changes will have on users. As ALTER SET SCHEMA changes the SchemaId, we need to decide how to pass the information from meta to frontend. There are 2 options here: Overwrite schema_id in proto. And since the old schema_id is absent in the proto, frontend needs to iterate over all schemas in the database and check who owns the [table | view | ...]_id to pin down the original schema. Add a new field optional uint32 new_schema_id in proto. Method 1 keeps the proto definition clean but introduce overehead for iteration, but since the no. of schemas in database are generally not huge and the frequencies of calling ALTER SET SCHEMA is relatively low, the overhead might be acceptable. There is no overhead in Method 2 but introducing a new field may be ugly. Which way will you suggest here? @yezizp2012 As ALTER SET SCHEMA changes the SchemaId, we need to decide how to pass the information from meta to frontend. There are 2 options here: Overwrite schema_id in proto. And since the old schema_id is absent in the proto, frontend needs to iterate over all schemas in the database and check who owns the [table | view | ...]_id to pin down the original schema. Add a new field optional uint32 new_schema_id in proto. Method 1 keeps the proto definition clean but introduce overehead for iteration, but since the no. of schemas in database are generally not huge and the frequencies of calling ALTER SET SCHEMA is relatively low, the overhead might be acceptable. There is no overhead in Method 2 but introducing a new field may be ugly. Which way will you suggest here? @yezizp2012 I prefer the first one. There won't be too many schemas and the altering will be very low-frequency. Good job! I will review it after my vacation.
2025-04-01T04:35:21.816968
2024-05-09T14:23:38
2287805653
{ "authors": [ "CAJan93", "tabVersion" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10362", "repo": "risingwavelabs/risingwave", "url": "https://github.com/risingwavelabs/risingwave/pull/16667" }
gharchive/pull-request
DO NOT MERGE: Notes about the RW + TiDB presentation in London What's changed and what's your intention? Overview of what I would like to present in London Please correct me if I am wrong Looking forward to it! See you in London.
2025-04-01T04:35:21.833533
2019-09-27T14:49:42
499490392
{ "authors": [ "MasterGroosha", "MaximPVladimirovich", "MeirionHughes", "PlanetDestroyyer", "bendeg", "jingle1000", "marko-36", "pvrooyen" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10363", "repo": "ritwickdey/vscode-live-server", "url": "https://github.com/ritwickdey/vscode-live-server/issues/576" }
gharchive/issue
Resource blocked due to MIME type mismatch I'm submitting a... [ ] Regression (a behavior that used to work and stopped working in a new release) [x] Bug report <!-- Please search GitHub for a similar issue or PR before submitting --> [ ] Feature request [ ] Documentation issue or request [ ] Other Current behavior I'm trying a simple bootstrap template. No JS and CSS are loaded because error: "Resource «http://<IP_ADDRESS>:5500/css/dashboard.css» blocked due to MIME type mismatch («text/html») (X-Content-Type-Options: nosniff)" I tried following this advice, however errors are gone and no CSS/JS are loaded. Opening the same "test" page in browser without live-server works and renders fine. This is how I include CSS: <link rel="stylesheet" type="text/css" href="../css/bootstrap.min.css" > Environment Browser: - [x] Firefox version 68 For Tooling issues: - Live Server: 5.6.1 - Platform: Windows 10 - Visual Studio Code: vsCodium 1.38.1 I am having this error as well :/ Same issue here... The problem is live-server is serving css with mime-type incorrectly set as text/html instead of text/css. Go to networks tab (In Chrome dev tools). Click on one of the red .css files that failed to load. Click on Response Headers and confirm Content-Type: text/html; charset=utf-8 I had this. The problem is the live server isn't serving relative to the index.html, its doing it relative to the workspace / package. i.e. I have /packages/foo/dist/index.html /packages/foo/dist/css/bar.css open index.html while vscode is open on the root folder and the browser will error that files are incorrect or it can't find js scripts. Temporary solution is to open vscode directly onto the /dist folder, start the live-server and then it will work fine. Had the same issue : As my working directory is a direct subfolder of the root directory on the web site (/rootdir/workingdir) , in Vscode, I have to open it locally by choosing the folder that is directly above it as the working directory, example : Locally : ../../../AAAA/workingdir/ => I have to open "AAAA" directory in Vscode, not "workingdir" Problem solved. Make sure the path is correct. changing this: href="css/style.css" to this: href="/css/style.css" worked! I also did not specify a type. href="css/style.css" vs. href="/css/style.css" might be the problem. --or-- In VSCODE, you might have the workspace directory set wrong, which causes this: Make sure that "path" link to the root of the site. In my case, it was pointing to the whole project site, one level above the actual website directory. i m having same issue in my flask application anyone got any solution??
2025-04-01T04:35:21.838718
2015-11-11T05:49:09
116270582
{ "authors": [ "mhlavacka", "ritz078" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10364", "repo": "ritz078/embed.js", "url": "https://github.com/ritz078/embed.js/issues/48" }
gharchive/issue
Instagram support Any plans to add Instagram support? I see there is Instagram mentioned in the code but not in the official support Its already there . The readme is just a bit outdated as currenly i am busy with a newer version of embed.js
2025-04-01T04:35:21.857146
2022-03-05T20:43:36
1160442170
{ "authors": [ "hermet", "projectitis" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10365", "repo": "rive-app/rive-tizen", "url": "https://github.com/rive-app/rive-tizen/pull/69" }
gharchive/pull-request
Fix issue with nested save and restore Hi @hermet I found that it is possible for Rive to nest calls to save/restore, so the current solution is not suitable. Looking at the rive/skia solution, they use a stack. So I have updated to a similar solution. Some of my test code looked like this: cout << "About to save" << endl; _renderer->save(); _renderer->transform(transform); artboard->draw(_renderer); _renderer->restore(); cout << "Finished restore" << endl; And calling this once results in something like: About to save save save save restore save restore save restore save restore restore restore Finished restore So nesting of these calls is found in practice :) Here is the riv file that I use for testing juice.zip @hermet @JSUYA Do you see any issue with this PR that I need to change? I was wondering if you wanted to avoid std::stack? @projectitis thanks, will check soon :) Thank you :)
2025-04-01T04:35:21.858973
2022-09-06T12:23:37
1363203994
{ "authors": [ "dreadedhamish", "rix1337" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10366", "repo": "rix1337/FeedCrawler", "url": "https://github.com/rix1337/FeedCrawler/issues/654" }
gharchive/issue
Is this suitable for music? I've set up FeedCrawler, added sites, and everything seems to be working fine (added sites are added, some are fine, a few with SSL errors etc...) Searches however don't work - as as I've dug deeper into the app it looks like it is created and usable just for video? Hey @dreadedhamish And thanks for taking interest in this project. FeedCrawler is not internationalized and I lack the resources to implement this. Only German users are able to use this project. Regarding your question: music is out of scope.
2025-04-01T04:35:21.903724
2015-01-21T04:42:39
54977142
{ "authors": [ "glycerine", "rjohnsondev" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10367", "repo": "rjohnsondev/go-safe-browsing-api", "url": "https://github.com/rjohnsondev/go-safe-browsing-api/issues/6" }
gharchive/issue
can full hashes be downloaded in bulk, in advance? Hi Richard, Thank you for this really useful go binding! It is really nice. I'm trying out the v3 adaption that github.com/kgilonne/go-safe-browsing-api provides. One problem: I get a ton of false positives from MightBeListed(), and IsListed() is too slow. Is it possible to pull down the full hashes every 1800 minutes (or the update frequency that google requests), and to otherwise remain offline while having accurate (full) hash matching? Thanks! Jason Hey Jason, Glad you find it useful; I'll be looking to merge in the v3 fork at some point when I have time. Interesting thought, while I can't see this being explicitly forbidden in a quick review of the docs, it isn't really intended behaviour from google's standpoint as outlined in the spec. Google is also quite explicit about full hashes not returning a positive for the IsListed unless there has been an update in the last 45 mins. From a practicality point of view, there seem to be around 300,000 prefix hashes around at any one time. The full length lookup call is designed to allow for a couple of hash lookups at once, but I'd be confident there would be a limit to the volume allowed with a single request. If however you managed to get a hundred prefix responses each request you could feasibly get the full DB with 3,000 requests. I would expect you to hit the backoff behaviour requests at some point trying to do that however. I'm also unsure on how much data this would end up being, I would speculate that it's a lot ;). (This is all assuming the new protocol follows the same basic behaviour as v2) All up, I'm not completely sure what you are trying to do, however unless you are running a list of URLs with a really, really high bad-URL hit ratio I think you would still be better off using MightBeListed, letting the matches trickle in through an async IsListed call for interesting matches over time, then re-running the full list again. Some hacking to the library here https://github.com/rjohnsondev/go-safe-browsing-api/blob/master/safebrowsing.go#L364 will allow you to extend the 45 mins to something longer if it is taking more than say 30 mins to retrieve your full hashes. Thanks, Richard Thanks for your thoughts Richard. The primary problem I'm having is that the number of links I want to scan (6K-8K per second) is just too much for a network roundtrip to google to keep up. I've asked on the google group for the api, but it doesn't seem to be attended. I may try to email the original devs directly for their advice. I don't mind updating every 30 minutes at all, since that can be done in the background in parallel. But with IsListed() now I'm only getting 10-20 links scanned per second, even with using MightBeListed() first--there are just that many collisions in the 32-bit hash prefix space. Anyway. I appreciate your suggestions. I'll let you know if I figure out a solution. Hi Richard, I consulted with one of the safe-browsing-api developers, and he suggested a batching strategy [below]. Is this viable? Could you advise on/suggest how to impliment such batching? Thanks! Jason ---------- Forwarded message ---------- From: Noé Lutz Date: Wed, Jan 28, 2015 at 4:42 PM To: Jason Hi Jason, Even at 10K links per second you should only see an estimated 1-2 hits per second. That's assuming the database contains 500K entries. Is that roughly what you're seeing? Hash lookups are extremely efficient and they can also be batched. Can you somehow put matching URLs in a separate queue and then batch hash server requests to our frontends? There is really no good way to download the entire list of full hashes. You could (technically) issue a hash server request whenever you get an update to pre-lengthen all the hash prefixes that you just got in an update. I don't think this would violate our policy but it would certainly not be the usual way of using the protocol. Hope that helps. noe. Yup, the full-length hash lookup is here https://github.com/rjohnsondev/go-safe-browsing-api/blob/master/safebrowsing.go#L453 this is usually called with the generated hashes for a matching URL. Cool, I'll give that a try. Thanks Richard.
2025-04-01T04:35:21.936068
2017-07-19T09:59:25
243986702
{ "authors": [ "kfirufk", "lucab" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10368", "repo": "rkt/rkt", "url": "https://github.com/rkt/rkt/issues/3744" }
gharchive/issue
stage1: enter should enter the application mount namespace too Helllo... I'm using php docker image as the base image for my container, it's in order to add additional extensions to php. everything works fine.. the image is created.. but when I'm running it and I execute 'rkt enter ID', I notice that /sys, /proc and /sys are not mounted. I'm testing it on Container Linux by CoreOS 1465.0.0 (Ladybug) with rkt version 1.27.0. when I use the PHP code to browse these directories I do see the volumes mounted properly and I can see their content. this issue evolved from an in issue I opened on acbuild: https://github.com/containers/build/issues/322 in general when I test the images that I created, it's easier for me to have these volumes mounted when I use 'rkt enter' to re-configure,modify the container for testing before re-creating it. (Capturing my comment in the other ticket) This probably need some patching on rkt enter side to detect there is a nested mount namespace and to enter it. The tricky part is that this additional mount namespace is optional (systemd creates it) and it must be detected by inspecting the top pid of your app, which is unknown. As a workaround if you just need /proc and /sys, after rkt-entering you can either mount them or nsenter the mount-ns of your app, depending on which binary you have available.
2025-04-01T04:35:21.956605
2017-07-25T21:57:20
245547654
{ "authors": [ "flyinb" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10369", "repo": "rlamasb/Firebase.Xamarin", "url": "https://github.com/rlamasb/Firebase.Xamarin/issues/32" }
gharchive/issue
"Response status code does not indicate success: 400 (Bad Request)." i am able to get a token, but I am getting a generic 400 error with the following code: var authProvider = new FirebaseAuthProvider(new FirebaseConfig("--My API Key--")); var auth = await authProvider.SignInWithEmailAndPasswordAsync(username, password); var firebase = new FirebaseClient("https://mydbname.firebaseio.com/"); var user = new LocalUser(); var xx = await firebase.Child("Users").Child(username).WithAuth(auth.FirebaseToken).PostAsync(user,false).ConfigureAwait(true); and the LocalUser is defined as: public class LocalUser { public string SomeData = "MoarData"; } i think i sorted it. Doesn't like "." or "@" symbols in the key. Would be nice to see the inner exception? OR a detailed message?
2025-04-01T04:35:21.964836
2023-04-11T14:44:46
1662656011
{ "authors": [ "DrBu7cher", "rlopzc" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10370", "repo": "rlopzc/logger_exporter", "url": "https://github.com/rlopzc/logger_exporter/issues/11" }
gharchive/issue
[bug] Exporting logs failed - JSON invalid This error occurred when I set it up as described in the README but without the LoggerJSON. 16:40:03.460 [error] [LoggerExporter] Exporting 2 logs failed. JSON too large or invalid 16:40:03.461 [error] [LoggerExporter] Error. Check the LoggerExporter configuration. Response: {:ok, %Finch.Response{status: 400, body: "loghttp.PushRequest.Streams: []*loghttp.Stream: unmarshalerDecoder: Value looks like Number/Boolean/None, but can't find its end: ',' or '}' symbol, error found in #10 byte of ...|\"}]}} \"]]}]}|..., bigger context ...|0:01 GMT\\\"}, {\\\"content-length\\\", \\\"274\\\"}]}} \"]]}]}|...\n", headers: [{"content-type", "text/plain; charset=utf-8"}, {"x-content-type-options", "nosniff"}, {"date", "Tue, 11 Apr 2023 14:40:03 GMT"}, {"content-length", "274"}]}} config: config :logger, backends: [:console, LoggerExporter.Backend] config :logger, LoggerExporter, app_name: "my_app", environment_name: config_env(), metadata: [:request_id], host: "http://localhost:3100" Did I miss a fallback option if the logged message is not valid JSON? I cannot turn this message off easily, since the library reporting that is not developed by me...and this might not be the only case where I could/am going to encounter such a situation...could the lib just drop invalid JSON messages? Or am I misunderstanding the message? According to this documentation: https://grafana.com/docs/loki/latest/api/#push-log-entries-to-loki the timestamp should actually be a string, not a number. I've created a fix. I'm going to test that it works and let you know here @DrBu7cher
2025-04-01T04:35:21.982715
2022-07-01T06:22:03
1290966458
{ "authors": [ "YvesW", "rmaes4" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10372", "repo": "rmaes4/decimal-prices", "url": "https://github.com/rmaes4/decimal-prices/issues/2" }
gharchive/issue
Update to RL 1.8.25 Your plugin failed to build with the 1.8.25 changes. I haven't looked into it too closely, but if I remember correctly, getVarcIntValue should be used instead of getVar. The SetVar methods also changed IIRC. You can probably check https://github.com/runelite/runelite/commit/87b41de60e8e8973b15ff95d4ca873b16b28c354 to find the changes. Would be lovely if you could update it; thanks in advance! @YvesW Thank you for bringing this to my attention. I have been taking a break from this game ever since I maxed. I was not aware that this plugin broke. I will try to fix this plugin this weekend as I should have some spare time. Thank you very much, that'd be awesome! I also lately had the problem that I'd get two dots ("..") while only entering one (i.e. the result would be "6..5k" while I only entered "6.5k"). So if you're going through your code and notice something that could cause that, then you know it might happen while playing. If you don't notice anything, I'll try test ingame once the plugin can be build and report back! @YvesW I am fixing the plugin now. I cannot seem to reproduce the double decimal issue. Could you please include specific information as to how that issue occurs? @YvesW I have submitted a PR to Runelite for the updated plugin here: https://github.com/runelite/plugin-hub/pull/2955 In addition to fixing the plugin I have also altered it to prevent entering more than one decimal. This should prevent the issue you were describing. @rmaes4 My apologies, I didn't see your responses earlier. Sadly I didn't check for example my client.log or any other logs when it happened, so I don't think I have too much more info. I'm using the US QWERTY international keyboard layout and while I was trying to enter the decimal prices, it would oddly enough always duplicate the dot. I'm not using the numpad and I'm able to enter single dots perfectly fine in other programs and in other places in the game. E.g. typing "1.2k" in public chat is no problem, but while setting the price, it would always result in "1..2k". If I remember correctly, a singular backspace would also delete both dots, which made it basically impossible for me to enter decimal prices. However, I'm not entirely sure anymore about the backspace part, so take that with a grain of salt. I currently can't test it since the plugin is disabled in the hub, although I could load your plugin during the weekend via IntelliJ if you'd like me to. Very much looking forward to enabling the plugin again; thank you for your work! If I find any weird bugs, I'll report it again. @YvesW Once the Runelite team makes the new version available, please test it and report back. I have modified the code to never allow more than one decimal for the price entry (I could not think of any reason why someone would need more than one). Hopefully this will fix the issue you have described! @rmaes4 I just noticed that NFC has merged the PR. Works great, thank you very much! A couple of my clanmates were already complaining that they were missing your plugin, so I'd like to thank you from them as well! While testing I noticed that e.g. "6.k" (no decimals) and "1.234" (no k or m) results in setting it to 1. However, I'd argue that's a feature and not a bug, since those have to be typos by the user. So with that, I'll close the issue! @YvesW Thank you for confirming that the plugin works. I'll admit that 6.k is an edge case I had not considered. At this point I have had no complaints or issues related to that scenario so I am inclined to leave it alone. Cheers and happy scaping!
2025-04-01T04:35:21.994792
2024-09-22T08:24:37
2540823158
{ "authors": [ "cameronr", "tbkj98" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10373", "repo": "rmagatti/auto-session", "url": "https://github.com/rmagatti/auto-session/issues/381" }
gharchive/issue
[FEATURE] Is your feature request related to a problem? Please describe. I am working with multiple directories/projects in a single NeoVim session. I am using NeoVim tabs to manage different projects in single session. I manually set the tab name using Tabby,nvim From the screenshot attached above, I've renamed the tab to "config". After saving the session and restoring it, The auto-session plugin restores everything(according to my requirement). but it does not restores the tab name Describe the solution you'd like I would like the tab names to be restored as well upon session restore. Additional context If anyone uses other solutions for the mentioned workflow. Please share, that would be great help. I have time to contribute to the project. According to the tabby.nvim docs, you just need to make sure you have globals in your sessionoptions: Save and restore in session You can save and restore tab layout and tab names in session, by adding word tabpages(for layout) and globals(for tab names) to vim.opt.sessionoptions. This is a valid sessionoptions: vim.opt.sessionoptions = 'curdir,folds,globals,help,tabpages,terminal,winsize' Can you double check your sessionoptions? Thank you for the answer. I was missing "globals" from sessionopts, but even after including globals, it did not work. Hmm, it worked for me in my testing. Did you make sure to save a new session after adding globals to your sessionoptions (sessions created before adding globals won't have the tabnames)? Can you post the following: What set sessionoptions? returns The contents of a session file from a session with a custom tab name created after you added globals? Hmm, that's very strange. auto-session uses native neovim sessions under the hood so it really is odd that it works when you usemks/source. Can you post the the following: The full contents of a session you create manually with mks that does save tab names The full contents of a session that auto-session creates that does not save tab names I'm trying to figure out if the issue is in saving the tab names to the session file or in restoring the names. Also, if you have your nvim configuration is publicly available that might also help me track down the issue. manual.txt auto-session.txt I've shared both of them and named the files accordingly. Interesting, the auto-session.txt session has an empty set of tab names (let TabbyTabNames = "[]") while the manual one is filled out correctly. It looks like those sessions are from two different projects, I wonder if that's related in some way (e.g. maybe it depends on what buffers you have open in each tab). Can you do the following: Set up the tabs (with names) how you want run :mks manual.vim to save a manual session caled manual.vim run :SessionSave test to have auto-session save a session called test.vim in ~/.local/share/nvim/sessions/ Exit vim to have auto-session automatically save the session in ~/.local/share/nvim/sessions/ with cwd as the name (with / escaped as %2F) Post the value for TabbyTabNames for all three sessions in this issue? Another possibility is a timing issue between when Tabby sets TabbyTabNames and when auto-session saves the session. Can you post both your auto-session config and your Tabby config to this issue? Thanks for the help Glad it's working. For posterity, do you know what ended up being the issue?
2025-04-01T04:35:22.021482
2018-03-03T17:35:12
302016747
{ "authors": [ "arispati", "dusadrian" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10375", "repo": "rndme/download", "url": "https://github.com/rndme/download/issues/61" }
gharchive/issue
error: download is not a function I don't know how to describe it properly, and neither do I have a MRE. But it seems that if the main function is called "download", that creates an error as described in the title. I have renamed your function to something else (e.g. "downloadit") and then it worked. No idea what might be triggering this error (I am using OS X version 10.9.5 and Chrome Version 64.0.3282.167), but I hope this might be of help somehow. I've the same issue, using NPM to install downloadjs. // Downloadjs window.download = require("downloadjs")(data, strFileName, strMimeType); end the result is Uncaught (in promise) ReferenceError: download is not defined Actually, that was my bad. I had another object called "download" in the global space (that I had forgotten about, and that cased the error in my case. Which explains why it worked when I renamed the function. Sorry for the noise, great library.
2025-04-01T04:35:22.051722
2019-10-19T09:41:33
1105120310
{ "authors": [ "rustatian", "wolfy-j" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10376", "repo": "roadrunner-server/roadrunner", "url": "https://github.com/roadrunner-server/roadrunner/issues/925" }
gharchive/issue
[💡FEATURE REQUEST]: Profiler Module based on xhprof Integrate https://github.com/tideways/php-xhprof-extension as the source of profiling data and pass this data into rr service via RPC. Store data in local db (BoltDB) or memory and share information about application performance via RPC or expose the port. Need UI. Thanks for the FR 👍🏻. Our PHP team created the official tool to collect profiling/debug information: https://github.com/buggregator/spiral-app, so, others are not planned (but for sure, can be added by the community).
2025-04-01T04:35:22.052963
2023-10-24T06:52:51
1958626578
{ "authors": [ "roannav", "sagarb2003" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10377", "repo": "roannav/matrix-movies-website-hacktoberfest", "url": "https://github.com/roannav/matrix-movies-website-hacktoberfest/pull/113" }
gharchive/pull-request
added hovering in links in footer https://github.com/roannav/matrix-movies-website-hacktoberfest/assets/99736036/6043b7dc-e904-45a9-9485-1650d01991de REVIEW AND MERGE IT UNDER HACKTOBERFEST2023 @all-contributors please add @sagarb2003 for code and design
2025-04-01T04:35:22.053906
2014-08-23T17:40:26
40980683
{ "authors": [ "rob-brown" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10378", "repo": "rob-brown/RBStoryboardLink", "url": "https://github.com/rob-brown/RBStoryboardLink/issues/40" }
gharchive/issue
Swift version? Should RBStoryboardLink be converted to be Swift-exclusive, have a Swift version, or remain as-is? At this point I don't see any benefits from a Swift version or two different versions. Keeping an ObjC version will guarantee both ObjC and Swift compatibility.
2025-04-01T04:35:22.057560
2015-06-22T04:16:10
90005287
{ "authors": [ "khendry", "richardPFisk" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10379", "repo": "robb/Underscore.m", "url": "https://github.com/robb/Underscore.m/issues/65" }
gharchive/issue
Published Pod is old Thanks for a great library. Would it be possible to publish an updated version 0.3.0? This would then allow easy use of functions such as groupBy. The current version, 0.2.1, does not contain groupBy. Please let me know if I can help. Thanks again. Working on it. Hopefully very soon. 0.3.0 has been released.
2025-04-01T04:35:22.059833
2023-03-30T09:16:22
1647228674
{ "authors": [ "kristieelim" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10380", "repo": "robbdasailer/Techlabs_Group09_Backend", "url": "https://github.com/robbdasailer/Techlabs_Group09_Backend/issues/7" }
gharchive/issue
Please enable cors for the netlify domain There is an issue based on cors in the netlify domain but not in localhost:3000. Please enable cors for the netlify domain: https://tafel-route.netlify.app/ As enabling cors for two domains seems more complicated than our deadline allows, please enable cors only for localhost:3000. This is because we need it more than the netlify domain for now. app.use(cors({ origin: 'https://localhost:3000/' }));
2025-04-01T04:35:22.102903
2019-02-28T17:24:37
415736836
{ "authors": [ "mcandre", "robfig" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:10381", "repo": "robfig/cron", "url": "https://github.com/robfig/cron/issues/177" }
gharchive/issue
Feature: Name and remove jobs For more dynamic workflows, would be really helpful to be able to assign names to jobs, and to delete named jobs. As a workaround, one could manage a map of names to individual cron.Cron instances, and manually Start() and Stop() each Cron instance as needed. This is supported on the v2 branch, and I'm trying to find the best way to support this for v3. Please let me know if the current approach on v2 or v3 has any problems for you.