id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1202090124
aws lambda-edge https://docs.aws.amazon.com/lambda/latest/dg/lambda-edge.html https://github.com/nuxt/nuxt.js/issues/11700 Hi, I wonder if you have any ETA for this preset? @tresko I created the PR exactly at the moment you commented. LOL! The current escape hatch is to wrap an existing preset for Lambda@Edge. An example of wrapping a aws-lambda preset is shown in the following code. import { URLSearchParams } from "url"; import * as nitro from "./index.mjs"; export const handler = async (event) => { const request = event.Records[0].cf.request; const queryStringParameters = Object.fromEntries( new URLSearchParams(request.querystring).entries() ); const response = await nitro.handler({ path: request.uri, queryStringParameters, httpMethod: request.method, headers: normalizeIncomingHeaders(request.headers), body: request.body, }); return { status: response.statusCode, headers: normalizeOutgoingHeaders(response.headers), body: response.body, }; }; function normalizeIncomingHeaders(headers) { return Object.fromEntries( Object.entries(headers).map(([key, keyValues]) => [ key, keyValues.map((kv) => kv.value).join(","), ]) ); } function normalizeOutgoingHeaders(headers) { return Object.fromEntries( Object.entries(headers).map(([key, values]) => [ key, values.split(",").map((value) => ({ value })), ]) ); } Saving this code as wrapper.mjs and setting the Lambda entry point as wrapper.handler should work. NOTE: require nodejs_14_x or nodejs_16_x @WinterYukky Wow, amazing! Thank you! I saw on your PR that you are using Nuxt3 on lambda@edge. Am I correct? Can you share a wrapper example for Nuxt 3? @tresko Yes. I'm using Nuxt3 on lambda@edge. I have created and published a simple project for you. It is a site that when accessed via /<anime title> calls Free's API and SSRs the results. The AWS configuration is written by AWS CDK in the cdk directory. I have included a few comments as you may not be familiar with the AWS CDK, but if you have any questions please ask. I have included the deployment procedure at the end of README.md so you can try it if you like. ref: https://github.com/WinterYukky/nuxt3-lambda-edge-example Thank you! @pi0 any plan to support this feature in the near future? Up, there are currently two PRs for this. It is very frustrating that @pi0 is not even explaining why/what was wrong with the first one/ what are the technical difficulties/ what are bigger priorities 😢 There's 2 PR Open for lambda-edge, but they are not the same. We could create 2 different presets, lambda-edge and lambda-edge-cdk https://github.com/unjs/nitro/pull/1075 => lambda-edge https://github.com/unjs/nitro/pull/240 => lambda-edge-cdk There's also the option to have a single preset and a config flag such as aws-lambda-edge.cdk = true The non cdk version is useful to integrate with AWS deployment frameworks such as SST Tracked in Nitro : https://github.com/unjs/nitro/issues/133 Tracked here in SST : https://github.com/serverless-stack/sst/issues/2314 Nuxt SST support PR : https://github.com/serverless-stack/sst/pull/2989 Working demo : https://github.com/serverless-stack/sst/issues/2314#issuecomment-1596223785 Hello, I can see that this development has been stopped for a while. Is there any intention to finish it? I wanted to submit a PR for this, but there are many of them already. If anybody wants a working solution based on the current opened PRs, you can use it from my fork: "nitropack": "git://github.com/AlbertSabate/nitro#df9eed697da2d7f09fa9a4a4f9d008fe11201df1" -> Updated to current v2.8.1 And then use the preset: preset: 'aws-lambda-edge',. Let me know if I can help speed this up to have it on the official repo. But, for now, I won't submit anything as the credits should go for the already opened ones. Hello, I can see that this development has been stopped for a while. Is there any intention to finish it? I wanted to submit a PR for this, but there are many of them already. If anybody wants a working solution based on the current opened PRs, you can use it from my fork: "nitropack": "git://github.com/AlbertSabate/nitro#df9eed697da2d7f09fa9a4a4f9d008fe11201df1" -> Updated to current v2.8.1 And then use the preset: preset: 'aws-lambda-edge',. Let me know if I can help speed this up to have it on the official repo. But, for now, I won't submit anything as the credits should go for the already opened ones. I think this hasn't been reviewed yet because it's low on the priority list. I left comments for @pi0 in #1557 about the 2 possible approaches. However I'm pretty sure SST has evolved since, so we would need to pin the compatible versions. An approach like #1075 would be easier to review and merged in, but that wouldn't be deployable by Nitro alone. Hello @Hebilicious, I have reviewed both PRs, and it will be very easy to get merged #1075; there are a few lines that need to be updated to make it work with the latest nitro version. After merging this, you can finish branch #1557, which seems way more complicated and, as you said, may need to be updated. For example, in my case, I use it to build a Solidjs app. In the new version 0.4.2, they have implemented Vinxi, which uses nitropack to build a solidjs app. Then I have my own CDK deployment script. So, having #1075 will be plug-and-play, while #1557 is irrelevant for my use case. This is an example. I'd be very happy to share my CDK script with you if that helps you finish #1557, but from what I see, the bottleneck comes from SST. My opinion on the matter: Step 1: Prepare and merge #1075. I can help you if necessary Step 2: Finish and merge the CDK as a separate preset since it is almost done. I can also give you a hand with this. This could make use of the previous handler. Step 3: SST: I need help understanding this point, as SST already provides everything for deploying an app. But it can come from step 3 because it can make a gain of the handler of step 1 and base the resource creation based on step 2. Having those three split may seem too much preset, but it will be clear, as my understanding. Let me know if you want me to help with anything! :)
gharchive/issue
2022-04-12T16:29:03
2025-04-01T06:46:07.505860
{ "authors": [ "AlbertSabate", "Hebilicious", "WinterYukky", "anjali89r", "ennioVisco", "pi0", "tresko" ], "repo": "unjs/nitro", "url": "https://github.com/unjs/nitro/issues/79", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2747157711
Support arrays to be passed to withQuery Describe the feature Currently, withQuery only allows objects to be passed, which means that you cannot do things like: example.com/path?foo=1&foo=2&foo=3 because objects cannot have multiple keys with the same name. It would be good to allow arrays to be passed to withQuery. Additional information [X] Would you be willing to help implement this feature? I saw that you are able to pass the object value an array which will solve my use case. i will close this issue
gharchive/issue
2024-12-18T08:36:46
2025-04-01T06:46:07.508112
{ "authors": [ "jwanner83" ], "repo": "unjs/ufo", "url": "https://github.com/unjs/ufo/issues/274", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1204969970
how to use the code? How can I run the code? hi @mlvl36667, sorry for the ultra late reply. can I still help you or are you good by now? @unnmdnwb3 I'd still be interested.
gharchive/issue
2022-04-14T20:56:04
2025-04-01T06:46:07.518752
{ "authors": [ "mlvl36667", "unnmdnwb3" ], "repo": "unnmdnwb3/rational-exchange", "url": "https://github.com/unnmdnwb3/rational-exchange/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2216330928
inspector conflicts with vite-plugin-html and cannot access /__unocss UnoCSS version 0.58.3 Describe the bug inspector conflicts with vite-plugin-html and cannot access /__unocss Reproduction https://stackblitz.com/edit/unocss-unocss-eorvyy?file=vite.config.ts System Info No response Validations [X] Read the Contributing Guidelines. [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. [X] Check that this is a concrete bug. For Q&A open a GitHub Discussion or join our Discord Chat Server. [X] The provided reproduction is a minimal reproducible example of the bug. Can you provide more specific information, such as how to reproduce this problem and your expected results? I wanted to implement @unocss/inspector, but the vite plugin used in the project, vite-plugin-html, blocked access to localhost:5173/__unocss You can take a look at this stackblitz : https://stackblitz.com/edit/unocss-unocss-eorvyy?file=vite.config.ts same here 我也遇到了,求解 我是将vite-plugin-html插件调整到 production 环境下就可以了, 因为vite开发环境下支持 env 变量输入到html
gharchive/issue
2024-03-30T08:32:02
2025-04-01T06:46:07.525285
{ "authors": [ "1514100951", "Simon-He95", "alpacachen" ], "repo": "unocss/unocss", "url": "https://github.com/unocss/unocss/issues/3669", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1646095644
FreeBSD:14:amd64 libpthread-stubs not found I'm getting the following error trying to install the unifi controller on my pfsense box (version 2.7.0-DEVELOPMENT (amd64)): Package libXdmcp-1.1.3 already installed. Fetching : . done pkg: archive_read_open_filename(/tmp/.XXXXX): Unrecognized archive format Failed to install the following 1 package(s): https://pkg.freebsd.org/FreeBSD:14:amd64/latest/ Steps: In pfSense GUI go to Diagnostics -> Command Prompt and enter in the following command: fetch -o - https://git.io/j7Jy | sh -s I get the following output: sh: mongod: not found Mounting new filesystems... done. Removing discontinued packages... Usage: pkg lock [-lqy] [-a|[-Cgix] <pkg-name>] pkg lock --has-locked-packages pkg unlock [-lqy] [-a|[-Cgix] <pkg-name>] For more information see 'pkg help lock'. Usage: pkg delete [-DfnqRy] [-Cgix] <pkg-name> ... pkg delete [-Dnqy] -a For more information see 'pkg help delete'. done. Installing required packages... packagesite.pkg 6590 kB 3525 kBps 02s x packagesite.yaml.sig x packagesite.yaml.pub x packagesite.yaml Package png-1.6.39 already installed. Package brotli-1.0.9,1 already installed. Package freetype2-2.12.1_2 already installed. Package fontconfig-2.14.2,1 already installed. Package alsa-lib-1.2.2_1 already installed. Package mpdecimal-2.5.1 already installed. Package python37-3.7.16_2 already installed. Package libfontenc-1.1.4 already installed. Package mkfontscale-1.2.1 already installed. Package dejavu-2.37_1 already installed. Package giflib-5.2.1 already installed. Package xorgproto-2022.1 already installed. Package libXdmcp-1.1.3 already installed. Fetching : . done pkg: archive_read_open_filename(/tmp/.XXXXX): Unrecognized archive format Failed to install the following 1 package(s): https://pkg.freebsd.org/FreeBSD:14:amd64/latest/ Looking at the actual script (https://raw.githubusercontent.com/gozoinks/unifi-pfsense/master/install-unifi/install-unifi.sh) I see that it's trying to AddPkg libpthread-stubs next. My "ABI" is "FreeBSD:14:amd64". So I checked the https://pkg.freebsd.org/FreeBSD:14:amd64/latest/packagesite.pkg and there is no libpthread-stubs in \packagesite\packagesite.yaml... So I guess my next possible step is to manually install libpthread-stubs. But I'm not sure which version to install. But even if I did manually install it I would need to create a new script and remove that line from it. What version of libpthread-stubs should I be installing? What command should I use to install it? How important is libpthread-stubs? (should it be removed from this install script) I've just encountered the same problem, while looking for the packages I found this: https://forums.freebsd.org/threads/devel-libpthread-stubs-port-has-been-deleted-no-consumers-left-and-never-supported-pthread-stubs-in-libc-on-freebsd.88350/ Based on this I downloaded the install-unifi.sh file, commented out the AddPkg libpthread-stubs line and re-ran the script locally. From what I can tell it seems that have installed okay. My backup restored and my devices have been re-adopted. Clients are showing as expected and the topology map has updated. OK Yes I did the same thing. I had to download the .sh file. Comment out the libpthread-stubs. Save it. Upload it to the pfSense Router. Use chmod u+e xxx.sh to be able to run it. SSH into the server, and run the .sh file (if I tried to do it in the GUI it would time out before it finished. It's possible that it still finished, but I couldn't see the output). Ugh, slightly a bit roundabout.
gharchive/issue
2023-03-29T15:53:52
2025-04-01T06:46:07.532535
{ "authors": [ "drohack", "pelstob" ], "repo": "unofficial-unifi/unifi-pfsense", "url": "https://github.com/unofficial-unifi/unifi-pfsense/issues/308", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
1364650993
[All][TextBox] Placeholder Text not Vertically Centered Current behavior When using a TextBox with MaterialOutlineTextBoxStyle the placeholder text is not vertically centered. Expected behavior When using a TextBox with MaterialOutlineTextBoxStyle the placeholder text should be vertically centered. How to reproduce it (as minimally and precisely as possible) Download MyApp.zip Open solution Start project Notice Placeholder Text in the TextBox is not vertically centered Environment Nuget Package: Uno.WinUI Uno.Material.WinUI Package Version(s): 4.5.0-dev.802 2.3.0-dev.12 Affected platform(s): [x] iOS [x] Android [ ] WebAssembly [ ] UWP [x] WinUI [ ] MacOS Anything else we need to know? Not fixed in 2.3.0-dev.20 Syncing with @Soap-141 at the moment to make sure of the nuget package version and the sample app he used for his tests So after syncing with @Soap-141 regarding the vertical alignment issue for the TextBox, Thomas did a sample app but there were no font files added to the sample so it was using the default font. Uno.Themes for Material is using Roboto font and the project is using EncodeSansCondensed font. So I'm trying to make sure placeholder and editable text are properly vertically aligned regarding what font is used at the end @jeromelaban, @kazo0 with all of the changes with the font lately I'm not sure if this issue is still relevant should be covered within #974 @jhanvi03 I will let you verify the Material TextBox (that you can test with Gallery Canary) to see if we still have some vertical alignment issues please. If so please leave more details for the impacted platforms. If not we can close this issue. Text seems to be align on iOS, Android and WinUI. I have opened another issue for Icon not align with the text https://github.com/unoplatform/Uno.Themes/issues/1443 Text seems to be align on iOS, Android and WinUI. I have opened another issue for Icon not align with the text #1443 Thanks @jhanvi03! I am closing this issue then, and we will use https://github.com/unoplatform/Uno.Themes/issues/1443 to track the alignment issue when there is an icon.
gharchive/issue
2022-09-07T13:16:56
2025-04-01T06:46:07.543535
{ "authors": [ "Soap-141", "Xiaoy312", "agneszitte", "agneszitte-nventive", "jhanvi03" ], "repo": "unoplatform/Uno.Themes", "url": "https://github.com/unoplatform/Uno.Themes/issues/842", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1398313105
[LoadingView] Use of Dispatcher on WinUI Current behavior https://github.com/unoplatform/uno.toolkit.ui/blob/main/src/Uno.Toolkit.UI/Controls/LoadingView/CompositeLoadableSource.cs#L124 We use the Dispatcher property on WinUI which is not on Windows WinUI project. Expected behavior We have to use the DispatcherQueue which is available on all platfroms. How to reproduce it (as minimally and precisely as possible) Use the LoadingView on Windows with a WinUI project Environment Affected platform(s): [ ] iOS [ ] Android [ ] WebAssembly [ ] WebAssembly renders for Xamarin.Forms [x] Windows [ ] Build tasks Visual Studio: [ ] 2017 (version: ) [ ] 2019 (version: ) [ ] for Mac (version: ) Relevant plugins: [ ] Resharper (version: ) Anything else we need to know? @Xiaoy312 @nickrandolph Not sure who is working on it right now :) I can take a look once I get toolkit building again - bad merge i have this fixed in my branch, not sure when it will be up/merged: https://github.com/unoplatform/uno.toolkit.ui/commit/32fadceabd433530b424cf9e386e1f2bd91650eb likely wont be in the same pr. it is kinda out of scope with runtime tests? btw CI is in a bad state today, so i cant just make a quick pr for that now :/ @nickrandolph you can just grab the code too Thanks. Will need to modify it slightly as we can't call getforcurrentthread from background thread as it returns null
gharchive/issue
2022-10-05T20:03:51
2025-04-01T06:46:07.550405
{ "authors": [ "Xiaoy312", "dr1rrb", "nickrandolph" ], "repo": "unoplatform/uno.toolkit.ui", "url": "https://github.com/unoplatform/uno.toolkit.ui/issues/355", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2336689072
fix(splash): apply workaround for layered background bug GitHub Issue (If applicable): closes unoplatform/uno#16725 PR Type What kind of change does this PR introduce? Bugfix Refactoring (no functional changes, no api changes) What is the current behavior? Opacity applied from fade out animation caused multiple colors layers to be displayed in splash screen. What is the new behavior? Workarounded by disabling the animation. PR Checklist Please check if your PR fulfills the following requirements: [x] Tested code with current supported SDKs [ ] Tested the changes where applicable: [x] WinUI [ ] iOS [ ] Android [ ] WASM [ ] MacOS [x] Updated the documentation as needed: [ ] General Doc Update [ ] Controls Doc Update [ ] Extensions Doc Update [ ] controls-styles.md [ ] lightweight-styling.md (LightWeight Styling Resource Keys) [ ] Runtime Tests and/or UI Tests for the changes have been added (for bug fixes / features) (if applicable) [x] Contains NO breaking changes [x] Associated with an issue (GitHub or internal) [x] Commits must be following the Conventional Commits specification. Other information Background existing in both resizetizer generated splashscreen image and in the wrapping grid. @agneszitte backport needed?
gharchive/pull-request
2024-06-05T19:33:35
2025-04-01T06:46:07.558446
{ "authors": [ "Xiaoy312", "kazo0" ], "repo": "unoplatform/uno.toolkit.ui", "url": "https://github.com/unoplatform/uno.toolkit.ui/pull/1151", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1416763452
[Map] DM-Liandri-Classic I want to ask to unset variations of this map. One is made to replicate the classic look and feel of it's UT99 version (hence the "Classic" suffix), while the other is made to be in line with UT3 more modern maps. URL: https://unrealarchive.org/maps/unreal-tournament-3/deathmatch/L/dm-liandri-classic_c5f76c72.html Hash: c5f76c72b396765c50e4d4ed64b33358f6326544 Current name: DM-Liandri-Classic makes sense, unset variation
gharchive/issue
2022-10-20T14:45:49
2025-04-01T06:46:07.603066
{ "authors": [ "PootisKorn", "shrimpza" ], "repo": "unreal-archive/unreal-archive-data", "url": "https://github.com/unreal-archive/unreal-archive-data/issues/1832", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
2341109247
Using custom Llama-3 tokenizer Hi, I have trained my custom Llama tokenizer and I was wondering how/if it would be feasible to train Llama-3 with Unclothe but use my tokenizer instead? You can load the tokenizer afterwards and separately after if that helps
gharchive/issue
2024-06-07T20:23:21
2025-04-01T06:46:07.606656
{ "authors": [ "MikeMpapa", "danielhanchen" ], "repo": "unslothai/unsloth", "url": "https://github.com/unslothai/unsloth/issues/607", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
723765454
Make a GitHub action GitHub action that downloads hyperlink, builds it and runs it with certain args. Original intent was to just install rust, compile the stuff and put it into GHA cache. Unfortunately you can't use actions inside actions: https://github.com/actions/runner/issues/646, and docker builds are also not cached (see #6) Perhaps we need proper releases after all. done!
gharchive/issue
2020-10-17T14:02:18
2025-04-01T06:46:07.608629
{ "authors": [ "untitaker" ], "repo": "untitaker/hyperlink", "url": "https://github.com/untitaker/hyperlink/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1841594882
Improve: Index finalizer method Implementing the __del__ method allows to add custom freeing logic to the object when it is finalized. From python docs object.__del__(self) Called when the instance is about to be destroyed. This is also called a finalizer or (improperly) a destructor. x.__del__() ... is only called when x’s reference count reaches zero. :tada: This PR is included in version 1.2.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-08-08T16:04:18
2025-04-01T06:46:07.611541
{ "authors": [ "AleksandrKent", "ashvardanian" ], "repo": "unum-cloud/usearch", "url": "https://github.com/unum-cloud/usearch/pull/194", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1987934685
How to build the site reproducibly? What I want: Get any mkdocs site (sources) that uses the "encryptcontent" plugin https://github.com/unverbuggt/mkdocs-encryptcontent-plugin/blob/version3/documentation/mkdocs.yml is definitely a suitable example Run mkdocs build --site-dir site1 Run mkdocs build --site-dir site2 Compare the directories and have them be identical (with diff -r site1 site2). What I get instead: Huge diffs https://github.com/mkdocs/regressions/actions/runs/6826747067/job/18567340779 Why I want this: I check many random mkdocs sites and plugins to ensure that any change to mkdocs itself does not disrupt their function. So in my case the two mkdocs build commands actually use a different version of mkdocs but that's irrelevant here. This would allow me to add this plugin to my test coverage. Otherwise I won't be able to do that. It's probably obvious to the plugin's author why this happens and it looks very intentional. The content is encrypted with random salt every time? https://github.com/search?q=repo%3Aunverbuggt%2Fmkdocs-encryptcontent-plugin+get_random_bytes(&type=code I would propose to add an environment variable such as MKDOCS_ENCRYPTCONTENT_CONSTANT_SALT that would be able to replace these random calls if present, or something like that. The content is encrypted with random salt every time? yes, kind of: the intialization vector (called IV) of every AES encrypted sting is randomized every time. This measure ensures, that the same plain text leads to different ciphertext. Also, all AES keys are randomized every build, but this isn't exactly a security measure (at least I can't think of a reason that would strictly require this). It's more a measure to ensure that the AES keys are random, as they remain secret (and are decrypted though the KDF keys). So the only way to safely encrypt the pages and make sure that every build produces the same output would be to save all IVs and all AES keys. As this would require a huge amount of effort it's not really something I'd invest time into. If you only do this for testing (to check if mkdocs and all of it's plugins produce the same output), then I'd go for an insecure test mode that would set all IVs to a fixed value and also set all AES keys to different but deterministic values. Also display huge warnings to never upload this page anywhere. Is this something that would help you do what you want? Thanks! Regarding the last paragraph: yes that is exactly what I'm asking for I've pushed changes that should lead to constant ciphertext output when insecure_test: true is defined. Please try it by cloning/installing the current development version and adding the following under the plugin configuration of mkdocs.yml: - encryptcontent: insecure_test: true Great! Thank you very much. It worked perfectly. https://github.com/mkdocs/regressions/actions/runs/6840922769/job/18600695060 Now just one more thing- I am actually interested in using this exact MkDocs site for my testing: https://github.com/unverbuggt/mkdocs-encryptcontent-plugin/blob/version3/documentation/mkdocs.yml -ideally without creating a fork of it just for the purposes of setting this config option in mkdocs.yml https://github.com/mkdocs/regressions/commit/6763245b4baa444bcf43837daadac74c3cf97099 To try this out, for now I added a workaround to edit the file on the fly, but if it could be avoided, that would be great. So if you could please additionally edit this file https://github.com/unverbuggt/mkdocs-encryptcontent-plugin/blob/version3/documentation/mkdocs.yml so that it includes such a config: - encryptcontent: insecure_test: !ENV [MKDOCS_ENCRYPTCONTENT_INSECURE_TEST, false] -that would really help me. yes, sure. Thanks!
gharchive/issue
2023-11-10T16:17:49
2025-04-01T06:46:07.621723
{ "authors": [ "oprypin", "unverbuggt" ], "repo": "unverbuggt/mkdocs-encryptcontent-plugin", "url": "https://github.com/unverbuggt/mkdocs-encryptcontent-plugin/issues/54", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1124991401
Remove shell script There is a shell script remanent leftover from testing in the repo [/script.sh] that needs to be removed. Opened #6
gharchive/issue
2022-02-05T17:49:57
2025-04-01T06:46:07.638250
{ "authors": [ "elohmrow", "natereid72" ], "repo": "upbound/platform-ref-azure", "url": "https://github.com/upbound/platform-ref-azure/issues/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
983001455
New resource of type "http" Is your feature request related to a problem? Please describe. I'm trying to build an updatecli workflow that retrieve the maven latest version, and I would like to add a condition that checks if the tarball of the latest version exists (same idea as checking if a Docker image is published but it's a file at a given URL). There are no way to do it, except using a condition of type file and a transformer that empties the buffer but it's hacky AND slow. Describe the solution you'd like I would like a new resource of type http that would allow the following: Source: Retrieve the body or the header of an URL Condition: Check if an URL exists (e.g. answer a certain type of HTTP code that could be defined) or if it has a given header defined (key or value) Target: POST/PUT the source to change an URL Describe alternatives you've considered I could use the "shell" (#264) resource with curl orwget Additional context That's indeed a good suggestion I think @mavimo was interested to have a look to this one. Instead of having a new HTTP plugin, we could also reuse the file plugin and just check if a http query return 200 I think @mavimo was interested to have a look to this one. Instead of having a new HTTP plugin, we could also reuse the file plugin and just check if a http query return 200 Using the file resource for http feels weird : its goal is to retrieve content of a file, while http can (should?) be specified for a custom verb, a map of headers, and its return as a source could be the body or a header for instance : different use cases @dduportal Good point @dduportal Good point I’m highly motivated to start a new resource ! UX Proposal for this new resource: name: End to end test of the 'http' resource kind pipelineid: "e2e/http" # Sources of type http returns either the body or a header sources: # Returns the content of the 'maven-metadata.xml' file content as source (multi-line, direct HTTP/200) getJenkinsWarArtifactMetadatas: kind: http spec: url: https://repo.jenkins-ci.org/releases/org/jenkins-ci/main/jenkins-war/maven-metadata.xml # Returns the content of the 'checksums.txt' file content as source (multi-line, following redirections HTTP/3xx) getUpdatecli0.65.1Checksums: kind: http spec: url: https://github.com/updatecli/updatecli/releases/download/v0.65.1/checksums.txt # HTTP/302 (GitHub redirection to raw.githubusercontent.com) ## Source fails due to HTTP/3xx redirect disabled by user # getUpdatecli0.65.1Checksums: # kind: http # spec: # url: https://github.com/updatecli/updatecli/releases/download/v0.65.1/checksums.txt # HTTP/302 (GitHub redirection to raw.githubusercontent.com) # followredirects: false # Default is true # Returns the content of the Header 'Location' (e.g. the target of the HTTP redirect) getRedirectLocationForUpdatecliLinuxAmd64Archive: kind: http spec: url: https://github.com/updatecli/updatecli/releases/download/v0.65.1/updatecli_Linux_arm64.tar.gz response: header: Location ## Source fails with an error: the URL returns HTTP/404 (same for server-side errors such as HTTP/5xx) # failOnHttpError: # kind: http # spec: # url: https://do.not.exists.com/unknown_resource # Returns the content of the 'maven-metadata.xml' file from the private URL (custom headers for the request) getFromPrivateArtifact: kind: http spec: url: https://private.maven.repo/releases/org/jenkins-ci/main/jenkins-war/maven-metadata.xml request: headers: Authorization: 'Bearer Xcjkdhvcjidcdscdcplcmdz' Accept: 'application/xml' conditions: # Returns 'true' if the specified URL returns HTTP/1xx, HTTP/2xx or HTTP/3xx checkForURL: kind: http spec: url: https://repo.jenkins-ci.org/releases/org/jenkins-ci/main/jenkins-war/maven-metadata.xml ## Conditions returns 'false' as the URL returns HTTP/404 ## If there is an HTTP/500 (server side error) then the conditions fails with an ERROR (different than returning false which "skips" the pipeline) # checkForNonExistingUrl: # kind: http # spec: # url: https://do.not.exists.com/unknown_resource # Returns 'true' as the URL exists checkForURL: kind: http spec: url: https://repo.jenkins-ci.org/releases/org/jenkins-ci/main/jenkins-war/maven-metadata.xml # Returns 'true' if the specified URL returns HTTP/1xx, HTTP/2xx or HTTP/3xx to the custom request (custom verb and headers) checkWithCustomRequest: kind: http spec: url: https://private.maven.repo/releases/org/jenkins-ci/main/jenkins-war/maven-metadata.xml request: verb: HEAD headers: Authorization: 'Bearer Xcjkdhvcjidcdscdcplcmdz' # Returns 'true' if the response header 'Location' contains the same value as the source 'remoteUrl' getRedirectLocationForUpdatecliLinuxAmd64Archive: kind: http sourceid: remoteUrl spec: url: https://github.com/updatecli/updatecli/releases/download/v0.65.1/updatecli_Linux_arm64.tar.gz response: header: Location # Returns 'true' if the response header 'Location' contains the value 'https://google.com' getRedirectLocationForUpdatecliLinuxAmd64Archive: kind: http disablesourceinput: true spec: url: https://github.com/updatecli/updatecli/releases/download/v0.65.1/updatecli_Linux_arm64.tar.gz response: header: Location assertvalue: https://google.com # Returns 'true' if the response code is HTTP/302 getUpdatecli0.65.1Checksums: kind: http spec: url: https://github.com/updatecli/updatecli/releases/download/v0.65.1/checksums.txt # HTTP/302 (GitHub redirection to raw.githubusercontent.com) response: code: 302 # No target => @mavimo @olblak would this proposal cover your use cases? => For the jenkins-infra project we would benefit of this resource for the following use cases: Checking for package availabilities: chocolatey (for instance https://github.com/jenkins-infra/packer-images/blob/1255193338032ef4439022f2e954ba75eb5f00df/updatecli/updatecli.d/trivy.yaml#L35) Checking for Maven binary on download sites (for instance https://github.com/jenkins-infra/packer-images/blob/1255193338032ef4439022f2e954ba75eb5f00df/updatecli/updatecli.d/maven.yml#L37) Checking for NodeJS binary (for instance https://github.com/jenkins-infra/packer-images/blob/1255193338032ef4439022f2e954ba75eb5f00df/updatecli/updatecli.d/nodejs.yml#L36) Retrieving checksum files to verifiy downloads (as we tend to have GPG keys in our repositories) @dduportal Your proposal is very nice and would be interesting. In my case, I was looking for something simpler :D just checking that a file exsit on a scm repository :p
gharchive/issue
2021-08-30T16:56:59
2025-04-01T06:46:07.648303
{ "authors": [ "dduportal", "olblak" ], "repo": "updatecli/updatecli", "url": "https://github.com/updatecli/updatecli/issues/268", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
191158773
Fix empty subscription bug Fix #14 Add a unit test for the erroneous behaviour Enable some tests that were left out by mistake in a previous commit Clean up environment variables that are set during tests in order to not interfere with other tests Thanks for spotting those 2 nonsensical variable names. They slipped by somehow after a refactor... LGTM
gharchive/pull-request
2016-11-23T00:05:04
2025-04-01T06:46:07.652350
{ "authors": [ "AlexisMontagne", "mihaitodor" ], "repo": "upfluence/sensu-client-go", "url": "https://github.com/upfluence/sensu-client-go/pull/15", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2646968509
🛑 OS Ticket is down In 843ede4, OS Ticket (https://support.upmin.edu.ph) was down: HTTP code: 0 Response time: 0 ms Resolved: OS Ticket is back up in 85a3bdd after 34 minutes.
gharchive/issue
2024-11-10T08:00:55
2025-04-01T06:46:07.656412
{ "authors": [ "upmin-dev" ], "repo": "upmin-dev/UP-time", "url": "https://github.com/upmin-dev/UP-time/issues/134", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2744746277
🛑 OS Ticket is down In 4e6f566, OS Ticket (https://support.upmin.edu.ph) was down: HTTP code: 0 Response time: 0 ms Resolved: OS Ticket is back up in 8492008 after 47 minutes.
gharchive/issue
2024-12-17T12:00:54
2025-04-01T06:46:07.659162
{ "authors": [ "upmin-dev" ], "repo": "upmin-dev/UP-time", "url": "https://github.com/upmin-dev/UP-time/issues/846", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
141971781
Added support for change default baseURL. Hi, I changed BCLBackend.m and BCLAdminBackend.m files to be able of override the default baseURL for client API via Info.plist needed in case of a self-hosted BeaconControl environment (my case). And added support to cancel default local notification presented in case of an action being performed in background-mode. Let me know if it's ok. Best regards. Hi Estevão. Thanks for contributing. Good point on that with dynamic changing base url and support for local notifications. Cheers
gharchive/pull-request
2016-03-18T20:37:49
2025-04-01T06:46:07.660704
{ "authors": [ "estevaolucas", "isanth" ], "repo": "upnext/BeaconControl_iOS_SDK", "url": "https://github.com/upnext/BeaconControl_iOS_SDK/pull/1", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2125348723
Mistral Endpoint Integration Similar to OpenAI, Anthropic, Cohere, etc., support Mistral Endpoints to be used as evaluator LLM Implemented in #558
gharchive/issue
2024-02-08T14:46:33
2025-04-01T06:46:07.676557
{ "authors": [ "Dominastorm", "sourabhagr" ], "repo": "uptrain-ai/uptrain", "url": "https://github.com/uptrain-ai/uptrain/issues/513", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2384335947
Sky: Security Model How does Sky prevent nodes from falsifying their identity when making requests? What is our long-term strategy towards closing any loopholes in this? Does Sky filter out all script tags and include enough built-in components for UI devs to get by? What is Sky's attitude towards using iframes for sandboxing? Will Shrubbery eventually place all nodes at their own subdomains so that they can handle their own client-side authentication? Solved by #101.
gharchive/issue
2024-07-01T17:03:46
2025-04-01T06:46:07.721983
{ "authors": [ "hanfel-dovned", "tiller-tolbus" ], "repo": "urbit/shrub", "url": "https://github.com/urbit/shrub/issues/78", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
287694949
URF .NET Core Hi, Thanks for the great project. I was looking for the github url of "URF .NET Core", but could not find it, so can you please provide it. also, i noticed that: https://www.nuget.org/packages/URF.Core/1.0.0 is unlisted. https://github.com/urfnet/URF.Core beta has been released, closing this.
gharchive/issue
2018-01-11T08:02:13
2025-04-01T06:46:07.729303
{ "authors": [ "lelong37", "reader-man" ], "repo": "urfnet/URF.NET", "url": "https://github.com/urfnet/URF.NET/issues/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1068130674
Mac M1 Laptops not Supported Environment: OS: Mac M1 Laptop - Monterey 12.0.1 Java version Azul Zulu 15.0.5 usb4java version 1.3.0 Bug description When running this library on the new M1 Max Macbook Pros you get the following exception: Caused by: org.usb4java.LoaderException: Native library not found in classpath: /org/usb4java/darwin-aarch64/libusb4java.dylib This doesn't seem to be a supported CPU architecture. Try building the source for this CPU architecture. Hope that helps.
gharchive/issue
2021-12-01T08:53:11
2025-04-01T06:46:07.775519
{ "authors": [ "ScottPierce", "rupeshsaxena" ], "repo": "usb4java/usb4java", "url": "https://github.com/usb4java/usb4java/issues/86", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
727114294
Standardize backend responses We will be switching the format of the data being sent back by the backend from the current disorganized and inconsistent mess to a consistent format. This will include revamping how calculation etc data is returned to become a matrix of data plus layers of arrays over it. released in 2.3.9, "testing" is not really relevant. any bugs or additional features shoudl be new issues
gharchive/issue
2020-10-22T07:07:55
2025-04-01T06:46:07.777408
{ "authors": [ "devowit" ], "repo": "usc-isi-i2/t2wml", "url": "https://github.com/usc-isi-i2/t2wml/issues/264", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
659770568
Adding Galaxy Digital script Links Issue link #147 Documentation links http://api2.galaxydigital.com/volunteer/docs/ Changes: Adds a script to sync the Galaxy Digital data for Delaware How To Test: Run the script against a dev airtable base Ran the script several times in dev and prod, the data synced successfully and did not crate duplicate records on re-run
gharchive/pull-request
2020-07-18T00:32:58
2025-04-01T06:46:07.782788
{ "authors": [ "michaelogren" ], "repo": "usdigitalresponse/neighbor-express", "url": "https://github.com/usdigitalresponse/neighbor-express/pull/165", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2539799765
Feature: Grant Notes feed #3428 Ticket #3428 Description Adds Grant Notes input and feed for Grant Activity sidebar Includes some adjustments for Grant Followers modal to align on styling/implementation Minor adjustments to BE service includes: adding lead ahead approach for pagination mirroring grant followers (next indicating additional set) adding flag to indicate note revision Screenshots / Demo Video Testing Automated and Unit Tests [x] Added Unit tests Manual tests for Reviewer [ ] Added steps to test feature/functionality manually Checklist [x] Provided ticket and description [x] Provided screenshots/demo [ ] Provided testing information [x] Provided adequate test coverage for all new code [x] Added PR reviewers @TylerHendrickson Changes from https://github.com/usdigitalresponse/usdr-gost/pull/3471 are merged and PR can be reviewed
gharchive/pull-request
2024-09-20T23:07:28
2025-04-01T06:46:07.787557
{ "authors": [ "greg-adams" ], "repo": "usdigitalresponse/usdr-gost", "url": "https://github.com/usdigitalresponse/usdr-gost/pull/3532", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2308759049
🛑 Apply page iframe is down In d048609, Apply page iframe (https://eop-fra.secure.force.com/digitalservice/) was down: HTTP code: 0 Response time: 0 ms Resolved: Apply page iframe is back up in 1a99e8f after 1 hour.
gharchive/issue
2024-05-21T17:19:30
2025-04-01T06:46:07.790308
{ "authors": [ "rachellanman" ], "repo": "usds/uptime-monitoring", "url": "https://github.com/usds/uptime-monitoring/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1369693097
Create needextract or needextend as target link We have a usecase where we import information from an external tool and automatically create need objects for the imported elements. Just like with needimport, the imported elements are then linked from other existing needs or extended using needextend or by adding textual contents around a needextract in different rST pages. For example, import-page.rst: .. needtoolimport:: :filter: type == 'tool-element' another-page.rst: .. needextend:: TOOL_ELEMENT_01 :+implements: REQ_SW_01 .. needextract:: :filter: "TOOL_ELEMENT_01" in id By default, when the imported elements are linked in this way, the HTML navigation always points to the page where the needs are imported (which is a page with a large number of imported elements). It would be nice to have a feature where a needextract or needextend can be declared as the link target for a need and not the original (imported) need. This helps to navigate to the need which has more details than the "raw" imported ones. Points to consider needextract or needextend could filter multiple elements. How to make sure target links are created for single elements needextract could extract the same need in multiple pages. How to mark only a single needextract as target link Nice idea. For needextract this is possbile, but not for needextend, becauseito does not write any output to the HTML. So it can't be a link target. For needextract I suggest a new option: is_target. Which is just a flag, no values are allowed. If is_target is set, the target-element (node) of the original need gets deactivated/removed. And the created need (the one by needextract ) gets a target-node with the same ID as the original one. is_target is used for all needs, which are created/filtered by needextract. If multiple needextract are using is_target for the same need, the first one wins (maybe a warning can be written). "First one", because the algorithm checks if the original need still has a target-node. If not, another needextract must have "stolen" it already. ah, you are right on needextend that it does not have an HTML representation and cannot be used. I like your proposol of is_target. Let's explore that futher. Any updates regarding this issue? Was blocked by it... Not really. Some tests were made to reset the link-target in Sphinx. But somehow Sphinx does not really accept these changes and is still using the original links/references. Hi, the solution with an is_target flag for needextract would be sufficient and very useful from my point of view. Can we expect that any time soon? Sorry to say, but not from my side, as the available time is quite low. But I would happily support everybody, who jumps in and tries to make this possible.
gharchive/issue
2022-09-12T11:05:17
2025-04-01T06:46:07.798482
{ "authors": [ "Slaaaash", "danwos", "georgsey", "twodrops" ], "repo": "useblocks/sphinx-needs", "url": "https://github.com/useblocks/sphinx-needs/issues/689", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2135137640
Update cli/overview.md Update of the parameters due to the newly available options (tests-only, bail) Closing this since it has already been updated in the documentation.
gharchive/pull-request
2024-02-14T20:20:07
2025-04-01T06:46:07.799792
{ "authors": [ "chrisnagel", "helloanoop" ], "repo": "usebruno/bruno-docs", "url": "https://github.com/usebruno/bruno-docs/pull/41", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1088120718
Add tracing and metrics Experimental tracing and metrics support for the service-api. Leaving this in draft for the future where we have a tracing backend.
gharchive/pull-request
2021-12-24T03:27:31
2025-04-01T06:46:07.802562
{ "authors": [ "smlx" ], "repo": "uselagoon/lagoon-ssh-portal", "url": "https://github.com/uselagoon/lagoon-ssh-portal/pull/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1465471656
[MacOS 13.1][Python 3.11]: build fails on clickhouse_driver Hi! First of all, thanks for publishing such a great framework for community. Recently, i've tried to build userver using https://userver.tech/d1/d03/md_en_userver_tutorial_build.html#macos and faced with a problem about "longintrepr.h" header, which was moved in python 3.11. CMake command just like in tutorial, i.e.: cmake -DCMAKE_BUILD_TYPE=Release -DCMAKE_EXPORT_COMPILE_COMMANDS=ON \ -DUSERVER_NO_WERROR=1 -DUSERVER_CHECK_PACKAGE_VERSIONS=0 \ -DUSERVER_FEATURE_REDIS_HI_MALLOC=1 \ -DUSERVER_FEATURE_CRYPTOPP_BLAKE2=0 -DUSERVER_DOWNLOAD_PACKAGE_CRYPTOPP=1 \ -DUSERVER_FEATURE_CLICKHOUSE=0 \ -DUSERVER_FEATURE_RABBITMQ=0 \ -DOPENSSL_ROOT_DIR=$(brew --prefix openssl@1.1) \ -DUSERVER_PG_INCLUDE_DIR=$(pg_config --includedir) -DUSERVER_PG_LIBRARY_DIR=$(pg_config --libdir) \ -DUSERVER_PG_PKGLIB_DIR=$(pg_config --pkglibdir) -DUSERVER_PG_SERVER_INCLUDE_DIR=$(pg_config --includedir-server) \ .. So we were stuck on cmake command on building wheels for clickhouse-driver by this: Building wheels for collected packages: clickhouse-driver Building wheel for clickhouse-driver (setup.py): started Building wheel for clickhouse-driver (setup.py): finished with status 'error' error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [104 lines of output] /Users/zaharkarlin/Libs/userver/build_release/venv-userver-testenv/lib/python3.11/site-packages/setuptools/config/setupcfg.py:508: SetuptoolsDeprecationWarning: The license_file parameter is deprecated, use license_files instead. warnings.warn(msg, warning_class) running bdist_wheel running build running build_py creating build creating build/lib.macosx-13-arm64-cpython-311 creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/blockstreamprofileinfo.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/log.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/queryprocessingstage.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/protocol.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/client.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/opentelemetry.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/clientinfo.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/reader.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/result.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/connection.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/context.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/defines.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/writer.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/errors.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/progress.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/block.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/readhelpers.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/types.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/available.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/writer.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/escape.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/compat.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/helpers.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams copying clickhouse_driver/streams/compressed.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams copying clickhouse_driver/streams/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams copying clickhouse_driver/streams/native.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/result.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/block.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/helpers.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/lz4hc.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/lz4.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/zstd.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/base.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/extras.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/connection.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/errors.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/cursor.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/service.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/floatcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/uuidcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/arraycolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/datecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nestedcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/util.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/enumcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/intervalcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/lowcardinalitycolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/tuplecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/intcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/boolcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/ipcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/stringcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nothingcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/mapcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copyi Running setup.py clean for clickhouse-driver ng clickhouse_driver/columns/nullablecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/decimalcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/datetimecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/exceptions.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/simpleaggregatefunctioncolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/base.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nullcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/service.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/floatcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/datecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/lowcardinalitycolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/tuplecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/intcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/stringcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/datetimecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/base.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy running build_ext building 'clickhouse_driver.bufferedreader' extension creating build/temp.macosx-13-arm64-cpython-311 creating build/temp.macosx-13-arm64-cpython-311/clickhouse_driver clang -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX13.sdk -I/Users/zaharkarlin/Libs/userver/build_release/venv-userver-testenv/include -I/opt/homebrew/opt/python@3.11/Frameworks/Python.framework/Versions/3.11/include/python3.11 -c clickhouse_driver/bufferedreader.c -o build/temp.macosx-13-arm64-cpython-311/clickhouse_driver/bufferedreader.o clickhouse_driver/bufferedreader.c:209:12: fatal error: 'longintrepr.h' file not found #include "longintrepr.h" ^~~~~~~~~~~~~~~ 1 error generated. error: command '/usr/bin/clang' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for clickhouse-driver Failed to build clickhouse-driver Installing collected packages: clickhouse-driver, aio-pika, pytest-aiohttp, yandex-taxi-testsuite Running setup.py install for clickhouse-driver: started Running setup.py install for clickhouse-driver: finished with status 'error' error: subprocess-exited-with-error × Running setup.py install for clickhouse-driver did not run successfully. │ exit code: 1 ╰─> [106 lines of output] /Users/zaharkarlin/Libs/userver/build_release/venv-userver-testenv/lib/python3.11/site-packages/setuptools/config/setupcfg.py:508: SetuptoolsDeprecationWarning: The license_file parameter is deprecated, use license_files instead. warnings.warn(msg, warning_class) running install /Users/zaharkarlin/Libs/userver/build_release/venv-userver-testenv/lib/python3.11/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. warnings.warn( running build running build_py creating build creating build/lib.macosx-13-arm64-cpython-311 creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/blockstreamprofileinfo.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/log.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/queryprocessingstage.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/protocol.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/client.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/opentelemetry.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/clientinfo.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/reader.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/result.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/connection.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/context.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/defines.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/writer.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/errors.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/progress.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/block.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver copying clickhouse_driver/readhelpers.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/types.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/available.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings copying clickhouse_driver/settings/writer.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/settings creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/escape.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/compat.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util copying clickhouse_driver/util/helpers.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/util creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams copying clickhouse_driver/streams/compressed.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams copying clickhouse_driver/streams/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams copying clickhouse_driver/streams/native.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/streams creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/result.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/block.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy copying clickhouse_driver/numpy/helpers.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/numpy creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/lz4hc.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/lz4.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/zstd.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression copying clickhouse_driver/compression/base.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/compression creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/extras.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/connection.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/errors.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi copying clickhouse_driver/dbapi/cursor.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/dbapi creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/service.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/floatcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/uuidcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/arraycolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/datecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nestedcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/util.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/enumcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/intervalcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/lowcardinalitycolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/tuplecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/intcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/boolcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/ipcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/stringcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nothingcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/mapcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nullablecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/decimalcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/datetimecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/exceptions.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/simpleaggregatefunctioncolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/base.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns copying clickhouse_driver/columns/nullcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns creating build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/service.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/floatcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/datecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/__init__.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/lowcardinalitycolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/tuplecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/intcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/stringcolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/datetimecolumn.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy copying clickhouse_driver/columns/numpy/base.py -> build/lib.macosx-13-arm64-cpython-311/clickhouse_driver/columns/numpy running build_ext building 'clickhouse_driver.bufferedreader' extension creating build/temp.macosx-13-arm64-cpython-311 creating build/temp.macosx-13-arm64-cpython-311/clickhouse_driver clang -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX13.sdk -I/Users/zaharkarlin/Libs/userver/build_release/venv-userver-testenv/include -I/opt/homebrew/opt/python@3.11/Frameworks/Python.framework/Versions/3.11/include/python3.11 -c clickhouse_driver/bufferedreader.c -o build/temp.macosx-13-arm64-cpython-311/clickhouse_driver/bufferedreader.o clickhouse_driver/bufferedreader.c:209:12: fatal error: 'longintrepr.h' file not found #include "longintrepr.h" ^~~~~~~~~~~~~~~ 1 error generated. error: command '/usr/bin/clang' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: legacy-install-failure × Encountered error while trying to install package. ╰─> clickhouse-driver note: This is an issue with the package mentioned above, not pip. hint: See above for output from the failure. CMake Error at cmake/UserverTestsuite.cmake:71 (message): Failed to install testsuite dependencies Call Stack (most recent call first): testsuite/SetupUserverTestsuiteEnv.cmake:16 (userver_venv_setup) CMakeLists.txt:107 (include) I think that this https://github.com/mymarilyn/clickhouse-driver/blob/42783d1bf1a6138ee2fec363c312e91c942a99bc/clickhouse_driver/bufferedreader.c#L208 in clickhouse_driver sources might be incorrect and fixed in https://github.com/mymarilyn/clickhouse-driver/blob/master/clickhouse_driver/bufferedreader.c (after cythonizing files). Please, help us to bypass this problem and build successfully on Mac OS. Hi! It seems like python3.11 issue is already fixed here in clickhouse-driver and the new version 0.2.5 was published on November 27 (Yes, the same day you wrote this issue, but few hours later :blush: ). Can you rerun your build аnd show clickhouse-driver version if it fails? You can find it in %your_build_root%/venv-userver-testenv/lib/python3.8/site-packages/clickhouse_driver-*/METADATA @DmitriySud thanks for linked issue, seems that it is a root-cause. that's how my METADATA version info looks like: clickhouse_driver-0.2.5.dist-info/METADATA Metadata-Version: 2.1 Name: clickhouse-driver Version: 0.2.5 Looks like solved to me. Please reopen if the issue appears again
gharchive/issue
2022-11-27T15:13:19
2025-04-01T06:46:07.818934
{ "authors": [ "DmitriySud", "Ladence", "apolukhin" ], "repo": "userver-framework/userver", "url": "https://github.com/userver-framework/userver/issues/206", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1289678252
🛑 Msite is down In d728700, Msite ($MSITE_URL) was down: HTTP code: 403 Response time: 814 ms Resolved: Msite is back up in 6210987.
gharchive/issue
2022-06-30T06:49:17
2025-04-01T06:46:07.822166
{ "authors": [ "rbudiharso" ], "repo": "usetada/status-page", "url": "https://github.com/usetada/status-page/issues/360", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1997266081
演習問題バグ修正 (todo) addbutton が新しい todo を作ったときに disabled になっていなかった マージします。
gharchive/pull-request
2023-11-16T16:32:19
2025-04-01T06:46:07.927722
{ "authors": [ "Fridge0" ], "repo": "ut-code/utcode-learn", "url": "https://github.com/ut-code/utcode-learn/pull/604", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2369169950
Chore: Dev container support Create Dev container configuration files to make a consistent dev environment. 試過一下swagger可以正常跑起來,應該是沒問題 這個看要不要處理一下,但應該沒差 還有README.md看要不要加一下中文w vscode with wsl => https://learn.microsoft.com/zh-tw/windows/wsl/tutorials/wsl-vscode) 在連接至 WSL 之後,若使用的是 Windows 作業系統,要開啟位於 C 或是其他磁碟/位置的目錄,請至 /mnt 底下尋找 若使用的是 VS code,啟動後會自動安裝 ESLint 及 MongoDB 等工具擴充套件 新增中文版 README,並新增 Dev container 的執行指引
gharchive/pull-request
2024-06-24T04:24:24
2025-04-01T06:46:07.937632
{ "authors": [ "Chun-Cheng", "Jeff92316046" ], "repo": "utaipei-sa/api.reserve.utsa", "url": "https://github.com/utaipei-sa/api.reserve.utsa/pull/53", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
902326124
Add comments to main.rs Start for #14 I have only commented main.rs, so you can check the commenting style, and let me know if you have something different in mind. If this is fine, after this I'll make a PR of 2-3 files at once as you said in #14. I have also started a high-level outerview documentation, which will help understand the overall flow and components. I think as I am writing this not knowing much about container runtimes or youki, this will be helpful for new contributors. Let me know if any changes are required. Hey, Thanks! Can you tell exactly where the changes are required? I kept docker in the flow diagram as it it not final anyways, and in comments I only mentioned docker, as I think currently it is only tested on docker. Let me know what changes are required and I will update the PR. @YJDoc2 I apologize for the confusing explanation. I have commented on it. Basically, if you don't show a concrete example (flow diagram, etc.), I hope you will use the term higher-level container runtime instead of docker. @utam0k Hey, I have made the requested changes, please review.
gharchive/pull-request
2021-05-26T12:37:26
2025-04-01T06:46:07.940893
{ "authors": [ "YJDoc2", "utam0k" ], "repo": "utam0k/youki", "url": "https://github.com/utam0k/youki/pull/38", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1501311892
utmctl in a renamed UTM.app fails with applicationNotFound Describe the issue When testing a new UTM beta version, I usually rename UTM.app to UTM-{version}.app and keep the previous version installed. Doings so with version 4.1.2 and running utmctl returns an applicationNotFound error. Configuration UTM Version: 4.1.2 macOS Version: 13.1 Mac Chip (Intel, M1, ...): M1 Crash log $ ln -s /Applications/UTM-4.1.2.app/Contents/MacOS/utmctl ~/.local/bin/utmctl $ utmctl list 2022-12-17 11:01:08.544 utmctl[655:1692025] Can't find app with file path /Applications/UTM.app Error: applicationNotFound Does it work if you use the full path directly without a symlink? Nope, same issue: $ /Applications/UTM-4.1.2.app/Contents/MacOS/utmctl list 2022-12-21 09:27:49.349 utmctl[29618:3569256] Can't find app with file path /Applications/UTM.app Error: applicationNotFound
gharchive/issue
2022-12-17T10:24:46
2025-04-01T06:46:07.966791
{ "authors": [ "jrjsmrtn", "osy" ], "repo": "utmapp/UTM", "url": "https://github.com/utmapp/UTM/issues/4816", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1842197621
ISO refuses to open I was trying to open an iso file for windows xp on utm but when i clicked it nothing would happen. I am using iPadOS 16.6 with UTM SE. I dont know if it has to do with my siging method because i use Scarlet to install apps :/ https://github.com/utmapp/UTM/issues?q=is%3Aissue+is%3Aclosed+Scarlet
gharchive/issue
2023-08-08T22:37:42
2025-04-01T06:46:07.968280
{ "authors": [ "GGamer11247", "osy" ], "repo": "utmapp/UTM", "url": "https://github.com/utmapp/UTM/issues/5546", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2002378915
Add new methods and fix old ones This PR adds 2 new methods to hubspot: getContact and getContactProperty and also fixes an old method updateAccount while changing the signature to better match hubspot instead of AC Closing in favour of #51
gharchive/pull-request
2023-11-20T14:52:35
2025-04-01T06:46:07.969193
{ "authors": [ "PineappleIOnic" ], "repo": "utopia-php/analytics", "url": "https://github.com/utopia-php/analytics/pull/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
710568681
Add Linux Desktop on your browser as option? I'm very intrigued in the Jupyter Desktop Server referenced in the welcome page to the Jupyterhub service. I'm currently running a VPN+VNC combo to connect to a remote teaching server for a lab course. Something like this could make life a lot easier for my students. How do I make this happen? I suppose that first you would need to install a vpn on xfce if xfce doesn't ship with it. @yuvipanda is this correct? @michaelmack it's actually already installed! Check out utoronto.2i2c.cloud/hub/user-redirect/desktop. You can also access it from under 'New' button in the default Jupyter Notebook view. We can install pretty much any desktop application you want. I'll be very excited if you can use this in your class - will help us improve the service for everyone! Let us know what you would need, and we can talk about how to access them? Currently everyone's memory limits are set for interactive python use, so maybe not big enough for desktop use fully yet. Can easily increase limits though. @ntaback jupyter-desktop-server already runs a VNC client, so nothing more would be needed really. @michaelmack heya! Just wanted to see if you had any time to look at it. Would love some feedback, so it can be made usable for your students :) @michaelmack heya! Just wanted to see if you had any time to look at it. Would love some feedback, so it can be made usable for your students :)
gharchive/issue
2020-09-28T20:49:59
2025-04-01T06:46:07.972264
{ "authors": [ "michaelmack", "ntaback", "yuvipanda" ], "repo": "utoronto-2i2c/jupyterhub-deploy", "url": "https://github.com/utoronto-2i2c/jupyterhub-deploy/issues/32", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1091818187
Any plans to split this into a separate library? Hey uttarayan :) I'd love to know if you plan to split mctl-rs into a separate library, so I (and potentially others) can use it in my own notifier program to get the status of mpd. Oh that seems like a good Idea I'll split the main into lib.rs and main.rs and keep everything in the same crate. Oh that seems like a good Idea I'll split the main into lib.rs and main.rs and keep everything in the same crate. Thank you, I look forward to using it, let me know if you'd like any help. Thank you, I look forward to using it, let me know if you'd like any help. Sure. I'll let you know If I need help. Sure. I'll let you know If I need help. Off-topic, can I have a go at having PlayerInfo expose even more data, such as volume or state? Off-topic, can I have a go at having PlayerInfo expose even more data, such as volume or state? I'd also love to add some kind of way to notify the caller if the requested player is not running. I'd also love to add some kind of way to notify the caller if the requested player is not running. Yeah I actually just use It to bind the keys to the media keys in my keyboard so I didn't add It when I was writing. I should have added some descriptive errors/messages. You can go ahead and open a PR if you want. Yeah I actually just use It to bind the keys to the media keys in my keyboard so I didn't add It when I was writing. I should have added some descriptive errors/messages. You can go ahead and open a PR if you want. Sure, I'm working on it right now :) Sure, I'm working on it right now :)
gharchive/issue
2022-01-01T13:32:27
2025-04-01T06:46:07.978083
{ "authors": [ "grtcdr", "uttarayan21" ], "repo": "uttarayan21/mctl-rs", "url": "https://github.com/uttarayan21/mctl-rs/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1642571588
du: Update the GNU error message to our for tests/du/threshold.sh We are doing a better job. Make it FAIL => PASS Warning: Congrats! The gnu test tests/du/threshold is no longer failing!
gharchive/pull-request
2023-03-27T18:07:21
2025-04-01T06:46:07.979366
{ "authors": [ "sylvestre" ], "repo": "uutils/coreutils", "url": "https://github.com/uutils/coreutils/pull/4656", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1845925193
Some updates of deps Some updates that renovate had issues with because of cargo deny. Why do you include clap but not tempfile? There are open renovate PRs for is_terminal and tempfile because of cargo-deny and rustix. There is no renovate issue with clap. I must've gotten tempfile and some other name confused. I added it. I did clap to anticipate a renovate PR, because it is out of date (a bit). We could specify a patch version for clap in Cargo.toml, then we would get more renovate PRs. I think that's unrelated. Renovate works on Cargo.lock versions as well. We just have a limit on Renovate PRs I think
gharchive/pull-request
2023-08-10T21:30:42
2025-04-01T06:46:07.982141
{ "authors": [ "cakebaker", "tertsdiepraam" ], "repo": "uutils/coreutils", "url": "https://github.com/uutils/coreutils/pull/5148", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2164994607
Implement program last last - show a listing of last logged in users Hey there, @sylvestre. I'm looking to expand my knowledge with Rust, and was wanting to try taking this on. I've done a bit of poking around in the OpenBSD version of last.c, as well as looking at pinky.rs (which also uses utmpx), and utmpx.rs. I was wondering a couple of things, just for clarification. For starters, do you think this is the right direction for getting resources to get a good implementation down? Would the move here be to add the crate for the coreutils implementation of utmpx.rs and use that to generate the list? Is the util-linux repo meant to also be used for other operating systems (Windows, *BSD, Mac) like the coreutils repo? was wanting to try taking this on Great :) For starters, do you think this is the right direction for getting resources to get a good implementation down? Sounds good to me. Would the move here be to add the crate for the coreutils implementation of utmpx.rs and use that to generate the list? Probably, though I don't know. The dependency itself is already defined (uucore), you simply have to enable the utmpx feature. Is the util-linux repo meant to also be used for other operating systems (Windows, *BSD, Mac) like the coreutils repo? Yes, it's a long-term goal and we currently run things on Windows and MacOS in the CI. But it's fine if it's Linux-only. Hope that helps and maybe @sylvestre will chime in when he returns from vacation. Awesome! Thanks for the info, and I'll see what I can get done on this. @Puffy1215 I haven't seen anything from @kanielrkirby . I think you can go ahead :) Ah, I had gotten a bit busy, sorry about that! If you'd like @Puffy1215, I could share the implementation I have currently, if only just for some inspiration / somewhere to start. I think the main point of interest at the time was that I needed to override the way utmpx does the DNS host resolution, as I believe the data might be structured differently on BSD systems. i.e., there's a host column, and an ipv6 column, and if the host just has a kernel, then it uses the ipv6 (or something like that, it's been a little bit since I had a chance to work on this). So that might be a good issue to make for the utmpx / uucore repo/crate, to get better parity with GNU. Though I'm not sure how it interacts with other tooling in util-linux or coreutils, so I can't be positive that's doable. Hey there @Puffy1215. I'm happy to pick this one back up if you lost interest. Just let me know if you still wanted to work on this one. Happy hacking :) @kanielrkirby I had a partially working implementation for last, so I will try to make a pull request for it this week. It was missing some functionality, but looking through the other utilities that seems to be fine.
gharchive/issue
2024-03-02T22:03:11
2025-04-01T06:46:07.991424
{ "authors": [ "Puffy1215", "cakebaker", "kanielrkirby", "sylvestre" ], "repo": "uutils/util-linux", "url": "https://github.com/uutils/util-linux/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2230494010
lscpu: fix never_loop lint error This PR fixes an error from the never_loop clippy lint and should make the "cargo clippy" job in the CI pass (at least on Ubuntu). It also fixes two "unused variable" warnings. Changes since last push: none, just a rebase
gharchive/pull-request
2024-04-08T07:42:16
2025-04-01T06:46:07.993145
{ "authors": [ "cakebaker" ], "repo": "uutils/util-linux", "url": "https://github.com/uutils/util-linux/pull/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
248483247
(too) rigid check of program-id. xxxx (NO DOT) Once again - same "erro" category like previous issue. We do have several programs where the name of the program-id is not terminated with a DOT. (Sure the grammar insists on a dot - but most compilers dont check/insist on a DOT to terminate the name ...) thus the parser runs into the next error: Parsing file AAA02.cbl. line 5:7 mismatched input 'ENVIRONMENT' expecting {COMMON, DEFINITION, INITIAL, IS, LIBRARY, DOT_FS} Collecting units in file AAA02.cbl. Cobollines with the incorrect declaration. 000010 IDENTIFICATION DIVISION. 000050 PROGRAM-ID. AAA02 000060******************************************************************07.05.15 * not DOT after the name of the program-id !!! Suggestion: though the parser is acting correctly - reality shows that most of the declarations like author or date-written or remarks etc (till input-output section) should be parsed with a rather error-forgiving algorithm. I have relaxed the grammar for identification division entries such as the program id in 048043e2c08957201815844d713bdb21ec3a042c. Now, a DOT_FS is not required. Added a unit test. Thanks!
gharchive/issue
2017-08-07T17:48:19
2025-04-01T06:46:08.065392
{ "authors": [ "Reinhard-Prehofer", "uwol" ], "repo": "uwol/cobol85parser", "url": "https://github.com/uwol/cobol85parser/issues/33", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
532588318
mandatory minion and clients can set image Signed-off-by: Ricardo Mateus rmateus@suse.com What does this PR change? ISSUE: https://github.com/SUSE/spacewalk/issues/10235 Changes: Client, minion and sshminion renamed for names not tight with image version in use Image version to used in client, minion and sshminion loaded from a varible. Default value is: "sles12sp4" We have the need to run test suites with different images for minions and clients. It was not possible since the image was hardcode to use "sles12sp4". This will imply changing the minions name in all terraform-test-runer configuration files which use cucumber-testsuite module to run. Thanks for the nice work Ricardo. LGTM, but please test. Eric if I understand correctly, you would test those changes in your branch for uyuni test suite refactor to use opensuse, and also update terraform-test-runner. Did I miss understood? Thanks I'm fine, as far we still keeping all the possible clients in modules/libvirt/controller/variables.tf ( I need it for QAM purposes) Thanks for the nice work Ricardo. LGTM, but please test. @Bischoff if I understand correctly, you would test those changes in your branch for uyuni test suite refactor to use opensuse, and also update terraform-test-runner. Did I miss understood? Thanks About testing: I just meant it looks good to me, but I can't guarantee I did not miss some error, so this PR would need some testing. The openSUSE on Uyuni is a different problem. I had it work, but stopped before doing the real switch by changing the runner, and DHCP and DNS records. I had PRs ready for that, but due to all recent refactorings, they probably needs some love and care again. I'll probably recontact you about that. Issue is spacewalk#9798 . @rjmateus can you please take care of the final round of testing and merging (no hurries)? Sorry to bother on last minute :trollface: Now I noticed we will use sles12sp4 as default image. I was thinking... could make sense to use sles15 if @Bischoff already did the work to use them as our default clients in the testsuite? P.S. Can be done in a future PR, for sure. Sure @srbarrios. I will make that change, test it and if everything goes well merge it @srbarrios @Bischoff @moio This PR doesn't make sense anymore since name and image are now a parameter for all clients/minions. What we can do is change "variable names" for hosts in cucumber_testsuite, which should be addressed in https://github.com/SUSE/spacewalk/issues/10730 I will close this PR and delete the associated branch. @srbarrios @Bischoff @moio This PR doesn't make sense anymore since name and image are now a parameter for all clients/minions. Correct. What we can do is change "variable names" for hosts in cucumber_testsuite, which should be addressed in SUSE/spacewalk#10730 Yes, as you know it would have been okay to do the renaming, and the arbitrary number of clients, in separate PRs, but why not. Yes, as you know it would have been okay to do the renaming, and the arbitrary number of clients, in separate PRs, but why not. We could do it in two separated PR, but that would imply changing twice all main.tf files in sumaform-test-runner that uses cucumber_testsuite, and I think this will not bring a benefit that justifies it. We could do it in two separated PR, but that would imply changing twice all main.tf files in sumaform-test-runner that uses cucumber_testsuite, Correct. and I think this will not bring a benefit that justifies it. I slightly disagree here :smile_cat: , but that's not important.
gharchive/pull-request
2019-12-04T10:25:37
2025-04-01T06:46:08.106830
{ "authors": [ "Bischoff", "moio", "rjmateus", "srbarrios" ], "repo": "uyuni-project/sumaform", "url": "https://github.com/uyuni-project/sumaform/pull/651", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
997060696
Increase RAM in dom0 on XEN Hypervisor What does this PR change? Increase RAM in dom0 on XEN Hypervisor. So we fix an issue caused by the OOM Killer killing a zypper refresh while bootstrapping through salt-ssh. Probably requires to increase the Xen VM memory amount to more than 2GB in the terraform modules.
gharchive/pull-request
2021-09-15T13:03:46
2025-04-01T06:46:08.108376
{ "authors": [ "cbosdo", "srbarrios" ], "repo": "uyuni-project/sumaform", "url": "https://github.com/uyuni-project/sumaform/pull/948", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
803774825
希望V2RAYN可以添加trojan go websocket功能。谢谢大神 希望V2RAYN可以添加trojan go websocket功能。谢谢大神 You went to the wrong place for that.
gharchive/issue
2021-02-08T17:31:10
2025-04-01T06:46:08.149397
{ "authors": [ "conray002", "database64128" ], "repo": "v2fly/v2ray-core", "url": "https://github.com/v2fly/v2ray-core/issues/661", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
208086334
V2Ray插件系统? 运行时动态加载dll/so插件喵? 具体可以是支持更多协议的插件喵 例如ICMP穿墙插件喵 例如DNS穿墙插件喵 这里还是正常人多一些,请尽量说得让别人能看懂。 Golang 在 1.8 中支持了插件模式,目前仅在 Linux 上可用。估计要等到 1.9 或之后的版本才可以全面使用。这此之前,技术上做不到动态加载第三方插件。 @v2ray 这只能说明go与go之间的FFI模式喵... https://golang.org/cmd/cgo/ https://github.com/sbinet/go-ffi 目前只能找到这些FFI的说喵 go的FFI状况看起来很糟糕喵... 没动力继续提Issue了喵
gharchive/issue
2017-02-16T10:50:59
2025-04-01T06:46:08.152567
{ "authors": [ "Robert-Tian", "mindcat", "v2ray" ], "repo": "v2ray/v2ray-core", "url": "https://github.com/v2ray/v2ray-core/issues/393", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
978847642
Use a lock to protect the mutable write buffer queue. What do these changes do? After introduce multi-thread socket server there seems a contention, that should be invalid for good clients, but the vineyardd shouldn't crash. Related issue number Resolves #463 Merged.
gharchive/pull-request
2021-08-25T08:08:05
2025-04-01T06:46:08.175742
{ "authors": [ "sighingnow" ], "repo": "v6d-io/v6d", "url": "https://github.com/v6d-io/v6d/pull/464", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
714681591
Solution to 0121_Best_Time_to_Buy_and_Sell_Stock solves issue #291 IMPORTANT: Please do not create a Pull Request without an issue. All Submissions: [x] I have read the CONTRIBUTING document. [x] My code is written in Python3 and is ending with .py [x] Have you checked to ensure there aren't other open Pull Requests for the same update/change? [x] I have checked that my submission does pass the test on LeetCode.com [x] Does your filename follow the naming Conventions? [x] Have you linked your PR to an Issue? Closing issues Put closes #XXXX in your comment to auto-close the issue that your PR fixes (if such). closes #291
gharchive/pull-request
2020-10-05T09:35:05
2025-04-01T06:46:08.191432
{ "authors": [ "amoghrajesh" ], "repo": "vJechsmayr/PythonAlgorithms", "url": "https://github.com/vJechsmayr/PythonAlgorithms/pull/317", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
176352878
Add context menu to master deploys Fixes #69 This change is Deployed : https://cdn.vaadin.com/vaadin-core-elements/master/vaadin-core-elements/ Reviewed 4 of 4 files at r1. Review status: all files reviewed at latest revision, all discussions resolved. Comments from Reviewable
gharchive/pull-request
2016-09-12T11:28:40
2025-04-01T06:46:08.276448
{ "authors": [ "manolo", "tehapo" ], "repo": "vaadin/vaadin-core-elements", "url": "https://github.com/vaadin/vaadin-core-elements/pull/70", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
174212486
Fixed suffix box-sizing to content box ...because their positioning relies on it For example including Bootstrap in the same page with vaadin-date-picker would make the icons border-box and thus misplaced This change is Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution. Reviewed 1 of 1 files at r1. Review status: all files reviewed at latest revision, all discussions resolved. Comments from Reviewable
gharchive/pull-request
2016-08-31T08:20:55
2025-04-01T06:46:08.280374
{ "authors": [ "CLAassistant", "tehapo", "tomivirkki" ], "repo": "vaadin/vaadin-date-picker", "url": "https://github.com/vaadin/vaadin-date-picker/pull/231", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
216410489
Alternative multiple columns syntax Fixes #5 This change is  Review status: 0 of 5 files reviewed at latest revision, 2 unresolved discussions, some commit checks broke. vaadin-form-layout.html, line 145 at r1 (raw file): _updateSize() { let columns = 6 * (1 + Math.floor(this.$.layout.clientWidth / 768)); I'm not smart enough to understand this formula :) Could you please add some explanatory comments? Why 1 +? What is 768? vaadin-form-layout.html, line 148 at r1 (raw file): this.updateStyles({'--vaadin-form-layout-columns': columns}); let columnBasis = this.$.layout.clientWidth / columns; use const by default; only use let if rebinding is needed Comments from Reviewable
gharchive/pull-request
2017-03-23T12:24:29
2025-04-01T06:46:08.285449
{ "authors": [ "limonte", "platosha" ], "repo": "vaadin/vaadin-form-layout", "url": "https://github.com/vaadin/vaadin-form-layout/pull/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
378730455
Prototype 1D moving content to body The idea is to move content of shadow root to light dom so that it's accessible by browser autocomplete and password managers. One option could be with vaadin-overlay. We can prototype this just using input element WIP can be found here: https://github.com/vaadin/vaadin-login/tree/prototype/%2326-move-content-body Findings: vaadin-overlay was used in order to make it add the elements in the root level no matter where the vaadin-login component is attached. Then, the current vaadin-login markup was moved into the vaadin-overlay shadow DOM by defining it within a template tag. Added a slot, so the username/password input (switched from vaadin-text-field/vaadin-password-field to plain HTML inputs) could go to the light DOM, hence making it be reachable by password managers. It was possible to make password managers (LastPass and 1Password) work. What needs to be fixed/improved: Possibility to add named slots which would make it easier to integrate with slotted vaadin-text-field 1.1. Named slot only worked if element is added to this.$.overlay.$.content which makes it go to content's shadow DOM Currently it doesn't work with ShadyDOM (inputs get lost) 2.1. Because of that, tests on the inputs are failing It wasn't able to fill the login form using LastPass popup, only through plugin's icon at toolbar (it looks like some issue with the backdrop) Investigate how to improve the usage of vaadin-overlay PS. Changed the demo just to make easier to see it working. Next step 1 D prototyping using https://github.com/vaadin/vaadin-text-field/tree/proto/slotted-input
gharchive/issue
2018-11-08T13:26:45
2025-04-01T06:46:08.307131
{ "authors": [ "DiegoCardoso", "alvarezguille" ], "repo": "vaadin/vaadin-login", "url": "https://github.com/vaadin/vaadin-login/issues/26", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2298068635
feat(service): add wildcard address resolver Summary This PR adds a service that expands wildcard addresses to their respective interface addresses equivalent. closes #1087 Context Users of nim-libp2p often listen to wildcard addresses like 0.0.0.0 (IPv4) and [::] (IPv6) to allow the OS to bind the transports to all available network interfaces. This is useful for applications that need to be accessible on multiple networks without specifying each interface. When listening on wildcard addresses, the OS binds to all available IP addresses, simplifying network configuration but requiring a mechanism to resolve these addresses to their specific network interfaces for communication with other peers. The listenAddrs field contains addresses the node listens on, which may include wildcard and private addresses (not directly reachable). The addrs field contains resolved addresses that other peers can use to connect, including public-facing NAT and port-forwarded addresses. Before this PR, nim-libp2p didn't offer a way to resolve wildcard addresses in the addrs field. This was reported on https://github.com/status-im/nimbus-eth2/issues/6060 and https://github.com/vacp2p/nim-libp2p/issues/1087. Changes Service created Documentation added Tests added Is it possible to add a withWilcardAddressResolverService() function in the switch builder? To have a simple way of creating it. This can work well in addition to withServices(@[ .... ]) Is it possible to add a withWilcardAddressResolverService() function in the switch builder? To have a simple way of creating it. This can work well in addition to withServices(@[ .... ]) Not a big fan tbh. Maybe we should always add it to the switch. Is there a reason to model the resolver as a service? Is there a situation in which this service would no be desired? Imo, the switch should offer this per default. The motivation was to make it optional. It might me a breaking change if we make it the default behavior tho. The motivation was to make it optional. It might be a breaking change if we make it the default behavior tho. With our plan to keep only the public marked API as stable in versions 1.x, we could break that. But I think keeping it as a Service is still nice cause the setup/stop is handled automatically by the Switch. The last commits that enabled the service by default were reverted as they caused strange errors when running the tests. Initially, it was nil access errors in the tor transport test, after those were fixed, random tests failed as the transports were not closed properly. This happened when running on my macOS m1. On CI the test would hang as it's possible to see in the commits. I added back the wildcard service enabled by default. There were some issues in the tests that I had to fix, but it seems that the weird errors happened cause of what was explained here https://github.com/vacp2p/nim-libp2p/pull/1105.
gharchive/pull-request
2024-05-15T14:15:34
2025-04-01T06:46:08.320452
{ "authors": [ "diegomrsantos", "lchenut" ], "repo": "vacp2p/nim-libp2p", "url": "https://github.com/vacp2p/nim-libp2p/pull/1099", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
168781289
SonarCore+Practical Logistics causing large CPU drain Hi Vaga! I am not sure how helpful this is to give to you, but I would like to provide it for reference's sake. The server's CPU while running version 1.9.3 of MageTech Gate was using over 75% for just Practical Logistics. The warmroast profile is below but I am afraid that is all I have. Please let me know if you need anything else I can provide. Image: http://imgur.com/a/6OdNe I can look into configs and see what I can change, but I will be sure to forward this issue onto the dev of SC and PL. sadly the 1.7.10 versions of these mods are not supported anymore. So either we deal with them eating the cpu or I pull them from the pack. I want this pack to be as server friendly as it can be so your input on the matter is welcome Auric. I have not seen the issue arise in a while, players appear to have disassembled their setups. If it becomes an issue again, we can disable it server side, no need to remove it from the pack. ok
gharchive/issue
2016-08-02T01:46:12
2025-04-01T06:46:08.336333
{ "authors": [ "AuricPolaris", "vagaprime" ], "repo": "vagaprime/MageTech-Gate", "url": "https://github.com/vagaprime/MageTech-Gate/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2152833576
about domain acc Thanks for your sharing. I try to apply this method (cdan_sdat.py) to UDA for an image regression task. But the domain acc keeps 100% with no decline. Can you give me some guidance? Thanks a lot. Thanks for your interest in our work. We suggest you start with domain adaptive image regression work and then apply SDAT to observe improvement. (see our README.MD on how to apply) The method cdan_sdat.py is meant for classification tasks and might not be generalized well for image regression.
gharchive/issue
2024-02-25T16:12:43
2025-04-01T06:46:08.361721
{ "authors": [ "dby629", "rangwani-harsh" ], "repo": "val-iisc/SDAT", "url": "https://github.com/val-iisc/SDAT/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
807408973
Feat/vid 1686 add did key All the tests are pointing to dev.vidchain.net where did:key didDocuments are resolved and the verification makes use of the latest version of @validatedid/did-jwt. You can update libraries, for husky you can check the vidchain-api root folder on how to adapt to new API Regarding dependencies updates, I've update them all in the last commit, including husky from v4 to v5. The vidchain api is still using v4, so I've followed the following guides to adapt husky configuration: Migrate v4 to v5 Helper Regarding dependencies updates, I've update them all in the last commit, including husky from v4 to v5. The vidchain api is still using v4, so I've followed the following guides to adapt husky configuration: Migrate v4 to v5 Helper I suggest you to give it a look to this new configuration I've pushed and test it yourself, please. I do not understand because I've updated it... as there is the new folder there. maybe I just forgot to do the update in the end... https://github.com/validatedid/vidchain-api/blob/development/.husky/commit-msg All changes required are done. Regarding the husky in vidchain api, I see the folder you're right. I just pulled last develop branch version and check husky version in package.json which is ^4.3.6.
gharchive/pull-request
2021-02-12T17:22:34
2025-04-01T06:46:08.387853
{ "authors": [ "iamtxena", "maurolucc" ], "repo": "validatedid/did-auth-oidc-siop", "url": "https://github.com/validatedid/did-auth-oidc-siop/pull/13", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1406725004
12K requests per minute to Wallbox API sign in endpoint Hello! I've noticed it's getting stuck in some cases doing up to 12K rpm to the sign-in endpoint. Not sure if it's related to the retry logic. Thank you Version 1.1.1 should have corrected issues with a request storm. The API response had the TTL reduced to just 15 min. The lates versions skips any refresh and just signs in when needed. Which should be when a HomeKit get happens, such as opening HomeKit or refreshing the homebridge accessories page. If you have some other app in the mix it might behave unexpectedly. I will need to see if there are any other changes in the API responses. awesome! thank you I assume you have not seen this again, the response for the token TTL has changed again to now 24hours.
gharchive/issue
2022-10-12T19:48:34
2025-04-01T06:46:08.442483
{ "authors": [ "AlbertMorenoDEV", "valiquette" ], "repo": "valiquette/homebridge-my-wallbox", "url": "https://github.com/valiquette/homebridge-my-wallbox/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
279061888
Asserts in tests I was dabbling around and noticed the Assert.Equal methods in the tests are having the wrong signature. For example: Assert.Equal(result.Succeeded, expected); should be: Assert.Equal(expected, result.Succeeded); Is it OK to change this (Assert.True and Assert.False) by using Shouldly? This way all our tests are written with Shouldly. I agree, we should switch fully to shouldly 👍
gharchive/issue
2017-12-04T16:36:45
2025-04-01T06:46:08.444767
{ "authors": [ "GooRiOn", "tdeschryver" ], "repo": "valit-stack/Valit", "url": "https://github.com/valit-stack/Valit/issues/135", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
211669737
How to style Typeahead options box Hi, Now i do realize it might be out of scope of this package but i don't know where else to post this question: How do i style the Typeahead options box? Everytime i click in the inspector to inspect its styling it closes and remove the DOM object. I want to make it full length (that of the input box) and probably allow it to open above the input field. Any pointers please? The way that works best for me to be able to inspect the DOM is using Chrome and the Chrome Dev Tools. As you have noticed, the minute you move from the page the menu goes away. Here is the trick: When the menu is open, hover over one of the menu items (not the input box) and then hit the [F8] key. This will pause the code preventing the menu from going away and allow you to inspect the DOM. @1-0-1 thanks man! its a perfect solution :)
gharchive/issue
2017-03-03T11:58:49
2025-04-01T06:46:08.450477
{ "authors": [ "1-0-1", "hassanasad" ], "repo": "valor-software/ng2-bootstrap", "url": "https://github.com/valor-software/ng2-bootstrap/issues/1705", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
167201478
Animation for the components Is it possible to activate animation for ng2-bootstrap components, for example Accordion? If not will this feature be implemented in the future? stable animation for angular2 should land with rc.5 and of course it will be added if you can any more question please use slack: https://www.hamsterpad.com/chat/ng2
gharchive/issue
2016-07-23T20:15:39
2025-04-01T06:46:08.452364
{ "authors": [ "urosjarc", "valorkin" ], "repo": "valor-software/ng2-bootstrap", "url": "https://github.com/valor-software/ng2-bootstrap/issues/774", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1532301936
Add mkdocs dependency The mono repo /docs has a dependency for a plugin; mkdocs-monorepo-plugin = "^1.0.4" Here we only want dependencies that are used in many of our repos. This is only used in one of them for now.
gharchive/issue
2023-01-13T13:23:30
2025-04-01T06:46:08.455283
{ "authors": [ "8ball030", "DavidMinarsch" ], "repo": "valory-xyz/tomte", "url": "https://github.com/valory-xyz/tomte/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
684240519
chore: use fast-glob use fast-glob instead of glob + pify. Much faster and lighter. Thanks for your contribution! I'd never heard of this library before, thank you for introducing me to it.
gharchive/pull-request
2020-08-23T19:17:21
2025-04-01T06:46:08.466893
{ "authors": [ "antfu", "brattonross" ], "repo": "vamplate/vite-plugin-voie", "url": "https://github.com/vamplate/vite-plugin-voie/pull/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
748081567
Null Safety Since Null Safety is in beta, it is time to update Kiwi library to support null safety. Are there any timelines on this? Thanks! Duplicate of #51
gharchive/issue
2020-11-21T19:00:19
2025-04-01T06:46:08.522790
{ "authors": [ "Yegorisa", "vanlooverenkoen" ], "repo": "vanlooverenkoen/kiwi", "url": "https://github.com/vanlooverenkoen/kiwi/issues/62", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2740801087
refactor: streamline OpenAI_Chat initialization and deprecate old parameters Ensure config is a dictionary and set default temperature. Raise exceptions for deprecated parameters: api_type, api_base, and api_version. Initialize OpenAI client with optional api_key and base_url from config. Update model selection logic to use 'gpt-4o-mini' for all cases if model is not provide. GPT-4o Mini is cheaper and more capable than GPT-3.5 Turbo. It also has the same context window as GPT-4o, so there’s no need to check if the prompt will fit on the smaller model.
gharchive/pull-request
2024-12-15T18:10:32
2025-04-01T06:46:08.524664
{ "authors": [ "PucaVaz" ], "repo": "vanna-ai/vanna", "url": "https://github.com/vanna-ai/vanna/pull/734", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
368365217
Alter Form Elements HTML If you are using a css framework for example bulma or bootstrap you have to use their html structure in forms. A nice feature would be to provide a way to change the default html given by vapid of all form elements. My first idea: Create an folder _form with files _form/_text.html or _form/_password.html and so on. @JohannesHoffmann Thanks for the suggestion. In cases like these, I think it might be better to bypass Vapid's form tag, and use Formspree directly (i.e., {{#form}} is just a convenience wrapper around Formpree). https://formspree.io/ Well that's right ;) - thank's for the work!
gharchive/issue
2018-10-09T19:39:31
2025-04-01T06:46:08.535087
{ "authors": [ "JohannesHoffmann", "srobbin" ], "repo": "vapid/vapid", "url": "https://github.com/vapid/vapid/issues/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
431216721
5-10 seconds for simple file change Hi there! from saving a template file till the first "change" is detected by ksync it takes like 5-10 seconds. Just then I see this in ksync watch: INFO[2266] updating pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2266] update complete pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2266] issuing reload pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2267] reloaded pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2270] updating pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2270] update complete pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2270] issuing reload pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama INFO[2270] reloaded pod=sf-deployment-9dfc49d8-skzp5 spec=prime-llama and same time this in kubectl logs ksync-mssnq -c syncthing -n kube-system: [ZBQPY] 21:53:31 VERBOSE: Device QMAOGBV-5PTTLL2-YIXNIJN-6CNFFGS-6AYJVZI-6CTIGVZ-2XUTA7R-6UMKDAT sent an index update for "prime-llama-sf-deployment-9dfc49d8-skzp5" with 1 items [ZBQPY] 21:53:31 VERBOSE: Folder "prime-llama-sf-deployment-9dfc49d8-skzp5" is now syncing [ZBQPY] 21:53:31 VERBOSE: Started syncing "prime-llama-sf-deployment-9dfc49d8-skzp5" / "templates/Default/main.html.twig" (update file) [ZBQPY] 21:53:31 VERBOSE: Finished syncing "prime-llama-sf-deployment-9dfc49d8-skzp5" / "templates/Default/main.html.twig" (update file): Success [ZBQPY] 21:53:31 VERBOSE: Remote change detected in folder "prime-llama-sf-deployment-9dfc49d8-skzp5": modified file templates/Default/main.html.twig [ZBQPY] 21:53:31 VERBOSE: Folder "prime-llama-sf-deployment-9dfc49d8-skzp5" is now idle [ZBQPY] 21:53:31 VERBOSE: Summary for folder "prime-llama-sf-deployment-9dfc49d8-skzp5" is map[errors:0 globalBytes:220047303 globalDeleted:0 globalDirectories:5636 globalFiles:27344 globalSymlinks:136 globalTotalItems:33116 inSyncBytes:220047303 inSyncFiles:27344 localBytes:220047303 localDeleted:0 localDirectories:5636 localFiles:27344 localSymlinks:136 localTotalItems:33116 needBytes:0 needDeletes:0 needDirectories:0 needFiles:0 needSymlinks:0 needTotalItems:0 pullErrors:0 sequence:125251 state:idle version:125251] [ZBQPY] 21:53:31 VERBOSE: Completion for folder "prime-llama-sf-deployment-9dfc49d8-skzp5" on device QMAOGBV-5PTTLL2-YIXNIJN-6CNFFGS-6AYJVZI-6CTIGVZ-2XUTA7R-6UMKDAT is 100% [ZBQPY] 21:53:32 VERBOSE: Folder "prime-llama-sf-deployment-9dfc49d8-skzp5" is now scan-waiting [ZBQPY] 21:53:32 VERBOSE: Folder "prime-llama-sf-deployment-9dfc49d8-skzp5" is now scanning [ZBQPY] 21:53:32 VERBOSE: Folder "prime-llama-sf-deployment-9dfc49d8-skzp5" is now idle I save the file using Webstorm. Maybe 1-2 seconds less if I safe with vim. ksync: Version: Release Go Version: go1.12.1 Git Commit: c27ce95 Git Tag: 0.3.6 Built: Mon Mar 18 17:06:42 +0000 2019 OS/Arch: darwin/amd64 service: Version: Release Go Version: go1.12.1 Git Commit: c27ce95 Git Tag: 0.3.6 Built: Mon Mar 18 17:02:16 +0000 2019 OSX Mojave Is this time delay normal? Thanks, Kim Hmm. It's definitely not, I typically see detection of local changes within a second. Remote will largely depend on the connection to the cluster. Ok, in my case the delay was caused because the initial sync was not completed yet and takes like 2 minutes. After its completed I'm down to like a second! You can check with kubectl logs ksync-mssnq -c syncthing -n kube-system -f --tail 20 the log output or just with ksync get to see if your configured sync has the status "watching".
gharchive/issue
2019-04-09T22:04:11
2025-04-01T06:46:08.539235
{ "authors": [ "timfallmk", "wuestkamp" ], "repo": "vapor-ware/ksync", "url": "https://github.com/vapor-ware/ksync/issues/281", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
672510043
The VBO or VAO is support? Hi, I found this control that can't write .SwapBuffers(); Have a method to draw the mesh from buffer? This is a general GL question; rendering/buffer swapping is controlled by WPF, rather than in the standard windowing-style approach. You don't need to call that, it's done for you by the platform.
gharchive/issue
2020-08-04T05:10:44
2025-04-01T06:46:08.559826
{ "authors": [ "ne8315ht6633", "varon" ], "repo": "varon/GLWpfControl", "url": "https://github.com/varon/GLWpfControl/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
148884677
13-Sokolov Флаг "--force-renderer-accessibility" не устанавливается, если были открыты другие окна Хрома. Перед запуском нужно закрыть другие окна. При первом запуске может появиться ошибка NonComVisibleBaseClass. Нужно в том же окне исключений убрать галку "Параметры исключений -> Остановить при возникновении этого типа" @vasily-v-ryabov Пожалуйста посмотрите В понедельник посмотрю детальнее. Уверен, есть какой-нибудь ключик типа --open-new-window. Про исключение пока не понял, буду запускать - посмотрю. Так и есть, ключик имеется, нагуглился с ходу. В течение дня сегодня попробую запустить без открытого хрома. Запустил без открытого Хрома. Сначала не смог найти "Логин", потому что у меня уже было залогинено 2 аккаунта (для этого можно заюзать incognito mode, кстати). Когда вышел из всех аккаунтов, получилось залогиниться, но на открытии explorer.exe всё застопорилось (по умолчанию открылась папка Libraries) и через несколько секунд упало с NullReferenceException. Проверял на Win7 x64. На разных осях по умолчанию разные папки открываются. Это должно быть учтено. Кроме того, Яндекс после логина выдаёт всякие назойливые советы, которые нужно закрывать, если они есть. Сделаю Incognito Искал по Automation ID. Попробую изменить стратегию поиска Уберу советы @vasily-v-ryabov Поправил, добавил закрытие для всплывающего предложения "Сохранить пароль?" У меня (инкогнито) не появляется в заголовке. Думаю, надо по регулярному выражению или по wildcard как-то искать. Убрал скобки из заголовка, однако это не помогло. Так и застыло на загруженной странице с логином. Ещё надо бы что-то с таймингами сделать, чтобы не так долго ждать, когда страница уже загрузилась, и пора вводить. @vasily-v-ryabov У меня (incognito) тоже не появляется, но при просмотре окна через inspect.exe всё находит. Перезапускал .exe из папки Release, с четвёртого раза начал находить окно. Возможно проблема в большом количестве открытых окон, которые приходится просматривать при поиске. Тайминги присутствуют из-за поиска элементов на странице. Как только нашёл - сразу вводит. В Inspect.exe действительно окно называется "Яндекс.Диск (Incognito)". Ладно, давай без инкогнито, чтобы не зависеть от языка. Но вот после открытия explorer.exe так ничего и не происходит (кроме падения через несколько секунд). Сразу открою explorer в нужной папке. Нашёл ключ: cmd> explorer.exe /select,C:\path\to\file\test.zip Да, так нормально. Только положите test.zip куда-нибудь в репо (что-то не вижу его) и путь относительно executable формируйте. В общем, цель - чтобы работало из коробки. В комментах - требования к начальным условиям (хром закрыт и т.п.). @vasily-v-ryabov Перед запуском: Убедиться, что в тестовом аккаунте на яндекс диске нет архива test.zip Разлогиниться из всех аккаунтов яндекс диска Закрыть все окна гугл хрома Но не из Яндекс-паспорта, кстати. Если честно, с этим Яндексом много мороки. С гуглом попроще было бы, я считаю. А ещё это должно быть в коде в комментах in English, а не здесь. @vasily-v-ryabov Добавил ОК, сейчас худо-бедно заработало. Осталось только автоматически позакрывать все открытые программой окна. Ещё у меня вопрос: откуда взят модуль MouseSimulator.cs? В нём никаких ссылок и копирайтов не видно. Надо указать хотя бы источник. А если требуется, то и копирайт. Не делал закрытие окон, потому что этого не было в задании. Доделать закрытие окон? MouseSimulator взят из комментария отсюда: http://stackoverflow.com/questions/23264529/webbrowser-contextmenu-run-action-of-menu-item И дописан. Каким образом это указать? Да, окна надо позакрывать. Раз взято со StackOverflow, достаточно в шапке модуля указать, что взято отсюда (ссылка на ответ) и доработано. @vasily-v-ryabov Поправил Кхм... Комменты в коде можно было бы и по-английски. Вопрос: а что, разве в Winium нет какой-нибудь Wait функциональности? Везде ставить Tread.Sleep() - это как-то не очень красиво. Сделал комментарии на английском. Не вижу ничего плохого в использовании Thread.sleep(). В самом Winium есть такие строчки: public void LeftButtonClick() { this.mouseSimulator.LeftButtonClick(); Thread.Sleep(250); } В репозитории Winium.Desktop есть только один такой вызов: Thread.Sleep(this.Automator.ActualCapabilities.LaunchDelay); И тот без magic numbers. Наверняка что-то есть без костыльных таймингов. Надо только хорошо поискать. В остальных случаях дёргается метод из Winium.Cruciatus, в котором и происходит Thread.Sleep(250); В Winium.Desktop нет такой функциональности: https://github.com/2gis/Winium.Desktop/wiki/Supported-Commands Ага, теперь понятно. Ну да, нигде нет совершенства. ОК. Вижу, в Winium.Cruciatus есть MouseSimulatorExt.cs. Скорее всего можно было использовать его вместо вставки со StackOverflow (всё-таки хоть и копипастный, а практически велосипед). Ладно. Думаю, этого достаточно. Основные проблемы и особенности высветились. Задание зачтено.
gharchive/pull-request
2016-04-16T19:54:00
2025-04-01T06:46:08.595284
{ "authors": [ "SokolovMS", "vasily-v-ryabov" ], "repo": "vasily-v-ryabov/ui-automation-course", "url": "https://github.com/vasily-v-ryabov/ui-automation-course/pull/1", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
543960419
Day406 Resolves #821. Pull Request Test Coverage Report for Build 1551 0 of 0 changed or added relevant lines in 0 files are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 100.0% Totals Change from base Build 1557: 0.0% Covered Lines: 5 Relevant Lines: 5 💛 - Coveralls
gharchive/pull-request
2019-12-30T17:53:36
2025-04-01T06:46:08.606285
{ "authors": [ "coveralls", "vaskoz" ], "repo": "vaskoz/dailycodingproblem-go", "url": "https://github.com/vaskoz/dailycodingproblem-go/pull/822", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
415617602
Look into anaconda + grakn installation failures for client-python This issue was originally posted by @flyingsilverfin on 2018-09-21 14:02. @flyingsilverfin special package need to be built for Anaconda. I assume the recommended way for installing PyPI packages into Anaconda environment is using pip (as with regular installations) Apparently it should be possible to use a standard pip package inside a conda environment, we just need to reproduce this and confirm it works.
gharchive/issue
2018-09-22T01:39:31
2025-04-01T06:46:08.666826
{ "authors": [ "flyingsilverfin", "grabl", "vmax" ], "repo": "vaticle/typedb-client-python", "url": "https://github.com/vaticle/typedb-client-python/issues/18", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1264231179
每次 updateSchema 就会重新重置表单的默认值 https://github.com/vbenjs/vue-vben-admin/blob/cfbd52bbe7e49dae7a6fd4a4a67c2823d86fd258/src/components/Form/src/hooks/useFormEvents.ts#L217 let updateData: Partial<FormSchema>[] = []; if (isObject(data)) { updateData.push(data as FormSchema); } if (isArray(data)) { updateData = [...data]; } const hasField = updateData.every( (item) => item.component === 'Divider' || (Reflect.has(item, 'field') && item.field), ); if (!hasField) { error( 'All children of the form Schema array that need to be updated must contain the `field` field', ); return; } const schema: FormSchema[] = []; const newSchema: FormSchema[] = [];//需要更新表单的集合 updateData.forEach((item) => { unref(getSchema).forEach((val) => { if (val.field === item.field) { // const newSchema = deepMerge(val, item); // schema.push(newSchema as FormSchema); newSchema.push(deepMerge(val, item)); schema.push(...newSchema); } else { schema.push(val); } }); }); // _setDefaultValue(schema); _setDefaultValue(newSchema);//需要更新表单的集合设置默认值 schemaRef.value = uniqBy(schema, 'field'); } 更改后的 updateSchema 函数 async function updateSchema(data: Partial<FormSchema> | Partial<FormSchema>[]) { let updateData: Partial<FormSchema>[] = []; if (isObject(data)) { updateData.push(data as FormSchema); } if (isArray(data)) { updateData = [...data]; } const hasField = updateData.every( (item) => item.component === 'Divider' || (Reflect.has(item, 'field') && item.field), ); if (!hasField) { error( 'All children of the form Schema array that need to be updated must contain the `field` field', ); return; } const schema: FormSchema[] = []; const newSchema: FormSchema[] = [];//需要更新表单的集合 updateData.forEach((item) => { unref(getSchema).forEach((val) => { if (val.field === item.field) { // const newSchema = deepMerge(val, item); // schema.push(newSchema as FormSchema); newSchema.push(deepMerge(val, item)); schema.push(...newSchema); } else { schema.push(val); } }); }); // _setDefaultValue(schema); _setDefaultValue(newSchema);//需要更新表单的集合设置默认值 schemaRef.value = uniqBy(schema, 'field'); }
gharchive/issue
2022-06-08T05:52:35
2025-04-01T06:46:08.705240
{ "authors": [ "gh852195168" ], "repo": "vbenjs/vue-vben-admin", "url": "https://github.com/vbenjs/vue-vben-admin/issues/1951", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2001539859
VFormDesign组件用到的过程中的报错 这个位置若找不到则idx为-1,此时判断if(idx)也为true进入到了判断里面就报错了 @datasre 是的,你可以提个pr修复他 bug已修复
gharchive/issue
2023-11-20T07:29:16
2025-04-01T06:46:08.707251
{ "authors": [ "datasre", "wangjue666" ], "repo": "vbenjs/vue-vben-admin", "url": "https://github.com/vbenjs/vue-vben-admin/issues/3304", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
750816042
Fix crash when query is empty Get jobs for all statuses by default. hey @robman87, we released 1.0 as a stable version, which means we would encourage users to bump the version and we will remove the latest-stable soon, can you please try reproducing the error on v1 and let us know if it sill happens in there? if so please feel free to send a PR as well 😄 thanks!
gharchive/pull-request
2020-11-25T13:25:29
2025-04-01T06:46:08.708651
{ "authors": [ "robman87", "vcapretz" ], "repo": "vcapretz/bull-board", "url": "https://github.com/vcapretz/bull-board/pull/165", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
322933588
Compositing a webview with an opengl context I am have a webview in one window and a 3d opengl view in another window. But I would like to composite them into the same window OR dock one window to the other. Wondering if SDL can help or if I should keep trying with glfw. I have not used SDL Hi @gedw99, I believe it isn't possible to do that if you're not using SDL for creating the windows. If you created the windows using SDL, you could use window.GetSurface() to get the pixels of the windows and blit one of the surfaces to another. Thank you the the thorough answer. Any examples of this that you know of. Its a bit specialised. Boy i wish i used SDL :) i see now that what i need to do is edit each of the programs i want to composite into SDL and give them a SDL "Veneer" on top so that the Main SDL windows can plug into them. Correct ? Here are the two systems i want to combine https://github.com/zserge/webview https://github.com/g3n/engine what about this ? http://lazyfoo.net/tutorials/SDL/09_the_viewport/index.php Its using SDL_RenderSetViewport. this looks like the way to get it working !!! https://github.com/golang-ui/nuklear/blob/master/cmd/nk-example-sdl2/main.go that gets me a opengl context with a SDL window managing it. I'm not familiar with how webview works. But you can either create separate windows in SDL, or do rendering inside one window. I couldn't get separate windows to follow each other as you drag them around since window movement event occurs only after the window was moved. If that's ok you can just use event to move the other window to the side of new windows position. Pressing 1 or 2 selects Renderer or OpenGL. Pressing R, G, B changes it's background color. Q exits. package main import ( "log" "runtime" "github.com/go-gl/gl/v2.1/gl" "github.com/veandco/go-sdl2/sdl" ) var winWidth int32 = 320 var winHeight int32 = 480 func main() { runtime.LockOSThread() err := sdl.Init(sdl.INIT_EVERYTHING) if err != nil { panic(err) } defer sdl.Quit() window1, err := sdl.CreateWindow("Renderer Window", sdl.WINDOWPOS_CENTERED, sdl.WINDOWPOS_CENTERED, winWidth, winHeight, 0) if err != nil { panic(err) } defer window1.Destroy() renderer, err := sdl.CreateRenderer(window1, -1, 0) if err != nil { panic(err) } defer renderer.Destroy() window2, err := sdl.CreateWindow("GL Window", sdl.WINDOWPOS_CENTERED, sdl.WINDOWPOS_CENTERED, winWidth, winHeight, sdl.WINDOW_OPENGL) if err != nil { panic(err) } defer window2.Destroy() x, y := window1.GetPosition() w, _ := window1.GetSize() window2.SetPosition(x+w, y) context, err := window2.GLCreateContext() if err != nil { panic(err) } defer sdl.GLDeleteContext(context) err = gl.Init() if err != nil { log.Fatalln(err) } w1ID, err := window1.GetID() if err != nil { panic(err) } w2ID, err := window2.GetID() if err != nil { panic(err) } isGL := false running := true for running { draw := false for event := sdl.PollEvent(); event != nil; event = sdl.PollEvent() { switch t := event.(type) { case *sdl.WindowEvent: if t.Event == sdl.WINDOWEVENT_MOVED { switch t.WindowID { case w1ID: window2.SetPosition(t.Data1+w, t.Data2) case w2ID: window1.SetPosition(t.Data1-w, t.Data2) } } case *sdl.KeyboardEvent: if t.Type == sdl.KEYDOWN { switch t.Keysym.Sym { case sdl.K_q: running = false break case sdl.K_1: isGL = false case sdl.K_2: isGL = true } if isGL { switch t.Keysym.Sym { case sdl.K_r: gl.ClearColor(1, 0, 0, 1) draw = true case sdl.K_g: gl.ClearColor(0, 1, 0, 1) draw = true case sdl.K_b: gl.ClearColor(0, 0, 1, 1) draw = true } if draw { gl.Clear(gl.COLOR_BUFFER_BIT) window2.GLSwap() } } if !isGL { switch t.Keysym.Sym { case sdl.K_r: renderer.SetDrawColor(255, 0, 0, 255) draw = true case sdl.K_g: renderer.SetDrawColor(0, 255, 0, 255) draw = true case sdl.K_b: renderer.SetDrawColor(0, 0, 255, 255) draw = true } if draw { renderer.Clear() renderer.Present() } } } case *sdl.QuitEvent: running = false } } } } The other option is to render into one window and manage where each item is rendered by yourself. This example will create SDL window with OpenGL rendering context that you can use with OpenGL and sdl.Renderer at the same time. Pressing R, G, B changes background color, Q exits. package main import ( "log" "github.com/go-gl/gl/v2.1/gl" "github.com/veandco/go-sdl2/sdl" ) var winWidth int32 = 320 var winHeight int32 = 240 var angle float32 var x float32 = 50.0 var y float32 = 50.0 func main() { runtime.LockOSThread() err := sdl.Init(sdl.INIT_EVERYTHING) if err != nil { panic(err) } defer sdl.Quit() window, err := sdl.CreateWindow("Renderer/GL Window", sdl.WINDOWPOS_CENTERED, sdl.WINDOWPOS_CENTERED, winWidth, winHeight, sdl.WINDOW_OPENGL) if err != nil { panic(err) } defer window.Destroy() context, err := window.GLCreateContext() if err != nil { panic(err) } defer sdl.GLDeleteContext(context) err = gl.Init() if err != nil { log.Fatalln(err) } glIdx := -1 nrd, err := sdl.GetNumRenderDrivers() if err != nil { panic(err) } rndInfo := sdl.RendererInfo{} for i := 0; i < nrd; i++ { _, err := sdl.GetRenderDriverInfo(i, &rndInfo) if err != nil { panic(err) } if rndInfo.Name == "opengl" { glIdx = i break } } renderer, err := sdl.CreateRenderer(window, glIdx, 0) if err != nil { panic(err) } defer renderer.Destroy() sdl.GLSetAttribute(sdl.GL_DOUBLEBUFFER, 1) gl.MatrixMode(gl.MODELVIEW) gl.LoadIdentity() running := true for running { draw := false for event := sdl.PollEvent(); event != nil; event = sdl.PollEvent() { switch t := event.(type) { case *sdl.KeyboardEvent: if t.Type == sdl.KEYDOWN { switch t.Keysym.Sym { case sdl.K_q: running = false break case sdl.K_r: renderer.SetDrawColor(255, 0, 0, 255) draw = true case sdl.K_g: renderer.SetDrawColor(0, 255, 0, 255) draw = true case sdl.K_b: renderer.SetDrawColor(0, 0, 255, 255) draw = true } if draw { renderer.Clear() renderer.Present() } } case *sdl.QuitEvent: running = false } } gl.Clear(gl.COLOR_BUFFER_BIT) gl.PushMatrix() gl.Translatef(float32(winWidth)/2, float32(winHeight)/2, 0) gl.Rotatef(angle, 0.0, 0.0, 1.0) gl.Begin(gl.QUADS) gl.Color3f(1.0, 0.0, 0.0) gl.Vertex2f(x, y) gl.Color3f(0.0, 1.0, 0.0) gl.Vertex2f(-x, y) gl.Color3f(0.0, 0.0, 1.0) gl.Vertex2f(-x, -y) gl.Color3f(1.0, 1.0, 1.0) gl.Vertex2f(x, -y) gl.End() gl.PopMatrix() renderer.Present() sdl.Delay(16) angle++ } } @malashin thank you soooo much for the sample of docking windows. that is as good as it gets. will give this a whirl now and see how it goes ! any progress @gedw99??? Hey @malashin your example with the docking windows is exactly what I was looking for thanks a lot. In the first window I create a WebView using : https://github.com/zserge/webview This also works so far, except for a few slight delays when moving a window. The other window will follow after a short time. I just wonder if it is possible to put these two into one main(parent) window?
gharchive/issue
2018-05-14T18:39:16
2025-04-01T06:46:08.777485
{ "authors": [ "dock-lab", "gedw99", "ghost", "malashin", "rucuriousyet", "veeableful" ], "repo": "veandco/go-sdl2", "url": "https://github.com/veandco/go-sdl2/issues/337", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
324634754
Add ability to mark messages unread This is essentially https://github.com/vector-im/riot-web/issues/4406 but for android: As in other messaging applications, it would be nice to mark a particular message in a room as unread, returning to the state before it was read. This is helpful for when you need a future reminder to do something, or you didn't have a chance to read everything up to the present, and would like to move the unread marker back. See the linked issue for additional discussion. I Copy the labels from https://github.com/vector-im/riot-web/issues/4406
gharchive/issue
2018-05-19T14:43:52
2025-04-01T06:46:08.802386
{ "authors": [ "bmarty", "matthijskooijman" ], "repo": "vector-im/riot-android", "url": "https://github.com/vector-im/riot-android/issues/2278", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1272503347
Change loader component to not use Math.random Loader component causes issues when its rendered server side. See here https://github.com/vegaprotocol/frontend-monorepo/pull/507#discussion_r896528340 Reopened, because there is still some things to improve: let's have one static version with no re-rendes in interval basics on css animation only
gharchive/issue
2022-06-15T16:57:23
2025-04-01T06:46:08.984264
{ "authors": [ "macqbat", "mattrussell36" ], "repo": "vegaprotocol/frontend-monorepo", "url": "https://github.com/vegaprotocol/frontend-monorepo/issues/570", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1313059367
Popup response for submission failed and successful is very similar Steps to reproduce: Put limit order IOC for price under the current market price so that it does not get hit Approved in wallet with enough assets to fulfill order At first glance, even though the submission failed, which is also in the text, the popup screen is very similar with colour and (including check V) with an order that is successful I have included two images, one that is successful and one that failed. this happened because the transaction status on IOC failure was default. it is a bug and I will address this in a new PR
gharchive/issue
2022-07-21T10:46:29
2025-04-01T06:46:08.986469
{ "authors": [ "MadalinaRaicu", "daunatv" ], "repo": "vegaprotocol/frontend-monorepo", "url": "https://github.com/vegaprotocol/frontend-monorepo/issues/837", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1826966527
chore: add rollup config Add rollup configuration to build cjs Accidentally closed, my apologies.
gharchive/pull-request
2023-07-28T19:12:03
2025-04-01T06:46:08.987311
{ "authors": [ "dexturr", "jeremyletang" ], "repo": "vegaprotocol/js-protos", "url": "https://github.com/vegaprotocol/js-protos/pull/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
954720950
Temporal Pose Smoothing and Instance Tracking Hey, I wanted to quickly comment on parts of your conclusion Also, one drawback of our approach is that it does not include tracking, the combination with a tracking algorithm remains future work. We (my colleague Emil and me) have also noticed this problem, especially when smoothing poses over time in scenarios with multiple people. When the person IDs are mixed, the algorithm tends towards their middle poses. That is, the person on one side is attracted to the other side and vice versa. This leads to hallucinations that look like artificial dances of the persons. I have recently implemented a global tracking solution based on the min-cost flow formulation https://github.com/cheind/py-globalflow that includes an application to track 2D human poses based on geometric joint features. When applied, the temporal smoothing improves dramatically as you can see from the following comparison video. https://youtu.be/aU3whnxvXFc Let me know what you think. Just an update: we've added appearance loss terms via Re-ID features to recover 'long-term' occluded persons. See https://www.youtube.com/watch?v=3pb1-teTw44 Docs updated https://github.com/cheind/py-globalflow Looks good! Do you have metrics on how well the tracking works? How does it compare to other methods on the PoseTrack benchmark? Hey! No we don't have any metrics yet. Global pose tracking was merely a proof-of-concept for us to see the results of pose smoothing on multi-person scenarios. Our goal is an real-time method that runs at interactive framerates. However, now that you mentioned it, I'm keen to find out how well the method actually performs on PoseTrack :) Btw., how did you compute the metrics for the multi-person datasets you mentioned, when your method is not multi-person capable? Best, Christoph The official evaluation script automatically allocates the detected poses to GT poses so there is no need for tracking. Got it!
gharchive/issue
2021-07-28T10:41:25
2025-04-01T06:46:08.992884
{ "authors": [ "cheind", "vegesm" ], "repo": "vegesm/pose_refinement", "url": "https://github.com/vegesm/pose_refinement/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
588607399
Error when using decompile-apk.sh this is the error message: jdgui-apk.sh: líne 4: source: util.sh: file not found is the same error with every option, the "util.sh" file exists i have the same problem did you solved ? try editor SH file line 4 source util.sh → ./source util.sh @skysumbra it must be source ./util.sh try sh => bash in every option bash jdgui-apk.sh $arg bash classyshark-apk.sh $arg bash jadx-apk.sh $arg bash bytecode-viewer.sh $arg @skysumbra it must be source ./util.sh This fixed the issue for me!
gharchive/issue
2020-03-26T17:46:44
2025-04-01T06:46:09.059683
{ "authors": [ "Aniskonig", "Frontesque", "Twopothead", "nicolasfritzges", "seedlord", "skysumbra" ], "repo": "venshine/decompile-apk", "url": "https://github.com/venshine/decompile-apk/issues/3", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1902070762
Typo in readme.md Prepatring INPUTS Hi Peter, Thank you for using CellphoneDB and for reporting it, this had been fixed for the next CellphoneDB release (if you see any more, feel free to check in https://github.com/ventolab/CellphoneDB/tree/scoring branch where we're preparing the new release). Many thanks again and good luck with your research. Best wishes, Robert.
gharchive/issue
2023-09-19T02:00:26
2025-04-01T06:46:09.061678
{ "authors": [ "datasome", "pvalle6" ], "repo": "ventolab/CellphoneDB", "url": "https://github.com/ventolab/CellphoneDB/issues/141", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
802992538
Question about raw_user_meta_data What's the purpose of copying the raw_user_meta_data column value to full_name and avatar_url? Couldn't these be set just to f.ex. an empty string? Is this just a leftover or does do you do this for a reason? https://github.com/vercel/nextjs-subscription-payments/blob/767eb938e541c635489703a5d8921b6e12aceb20/schema.sql#L26 This function is triggered when a new user is added to the auth schema (sign up or sign in via OAuth) and copies the name and avatar URL that we get from the OAuth provider (e.g. GitHub; Google; etc) to our public users table. You can find some docs about it here: https://supabase.io/docs/guides/auth#create-a-publicusers-table Gotya, thanks for the clarification. I think this is something that should be documented by Supabase.
gharchive/issue
2021-02-07T16:29:01
2025-04-01T06:46:09.851710
{ "authors": [ "sarukuku", "thorwebdev" ], "repo": "vercel/nextjs-subscription-payments", "url": "https://github.com/vercel/nextjs-subscription-payments/issues/49", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1152185195
stopped using bind in unnecessary areas Remove bind in unnecessary areas. Thanks for your review! I thought anonymous functions were more recognizable, but as you said, bundle size should be a priority. I have learned a great deal. So, I'll close this PR!
gharchive/pull-request
2022-02-26T19:31:33
2025-04-01T06:46:09.858506
{ "authors": [ "Cut0" ], "repo": "vercel/swr", "url": "https://github.com/vercel/swr/pull/1873", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1792014437
chore: suggest git upgrade Description We rely certain git flags to properly function and we can use the exit code 129 to identify when git doesn't recognize a flag we passed it. Currently suggesting 2.18 as --no-renames was added in that version and I believe that's the newest flag we rely on. Testing Instructions Eyes Because I was interested, ubuntu 16.04 LTS is on 2.7.4, but current 'in-life' LTS of both ubuntu and debian are on at least 2.20
gharchive/pull-request
2023-07-06T18:11:48
2025-04-01T06:46:09.860443
{ "authors": [ "arlyon", "chris-olszewski" ], "repo": "vercel/turbo", "url": "https://github.com/vercel/turbo/pull/5472", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
650147707
Error with "npm audit" Describe the bug When running "npm audit" in my dev-project, it works fine if the registry is set to "https://registry.npmjs.org" But I get the following error when running "npm audit" after running "npm set registry PRIVATE-VERDACCIO-REGISTRY" - npm ERR! code ENOAUDIT npm ERR! audit Your configured registry (PRIVATE-VERDACCIO-REGISTRY) does not support audit requests, or the audit endpoint is temporarily unavailable. npm ERR! A complete log of this run can be found in: npm ERR! /Users/** This error is new, and wasn't there before I upgraded verdaccio to the latest release, so not sure what I'm missing To Reproduce Steps to reproduce the behavior: npm set registry PRIVATE-VERDACCIO-REGISTRY npm install npm audit Expected behavior When i ran "npm install; npm audit", I had expected back list of vulnerabilities like - === npm audit security report === Run npm install express-fileupload@1.1.6 to resolve 1 vulnerability SEMVER WARNING: Recommended action is a potentially breaking change ┌───────────────┬──────────────────────────────────────────────────────────────┐ │ Low │ Denial of Service │ ├───────────────┼──────────────────────────────────────────────────────────────┤ │ Package │ express-fileupload │ ├───────────────┼──────────────────────────────────────────────────────────────┤ │ Dependency of │ express-fileupload │ ├───────────────┼──────────────────────────────────────────────────────────────┤ │ Path │ express-fileupload │ ├───────────────┼──────────────────────────────────────────────────────────────┤ │ More info │ https://npmjs.com/advisories/1216 │ └───────────────┴──────────────────────────────────────────────────────────────┘ found 1 low severity vulnerability in 323 scanned packages 1 vulnerability requires semver-major dependency updates. Docker || Kubernetes (please complete the following information): I'm running private registry Verdaccio 4.7.2 in a AWS ECS container created off of dockerhub image verdaccio/verdaccio Configuration File (cat ~/.config/verdaccio/config.yaml) /opt/verdaccio # cat /verdaccio/conf/config.yaml This is the config file used for the docker images. It allows all users to do anything, so don't use it on production systems. Do not configure host and port under listen in this file as it will be ignored when using docker. see https://github.com/verdaccio/verdaccio/blob/master/wiki/docker.md#docker-and-custom-port-configuration Look here for more config file examples: https://github.com/verdaccio/verdaccio/tree/master/conf path to a directory with all packages storage: /verdaccio/storage store: aws-s3-storage: bucket: private-npm-registry region: us-west-1 # US West (N. California) web: WebUI is enabled as default, if you want disable it, just uncomment this line title: NPM Registry logo: https://xxx.cloudfront.net/wp-content/themes/xxx/assets/img/logo.svg auth: htpasswd: file: /verdaccio/conf/htpasswd max_users: -1 a list of other known repositories we can talk to uplinks: npmjs: url: https://registry.npmjs.org/ packages: '@/': access: $authenticated publish: $authenticated proxy: npmjs '**': access: $authenticated publish: $authenticated proxy: npmjs middlewares: audit: enabled: true logs: {type: stdout, format: pretty, level: http} Debugging output $ NODE_DEBUG=request verdaccio display request calls (verdaccio <--> uplinks) $ DEBUG=express:* verdaccio enable extreme verdaccio debug mode (verdaccio api) $ npm -ddd prints: npm info it worked if it ends with ok npm verb cli [ npm verb cli '/usr/local/Cellar/node/12.5.0/bin/node', npm verb cli '/usr/local/bin/npm', npm verb cli '-ddd' npm verb cli ] npm info using npm@6.9.0 npm info using node@v12.5.0 Usage: npm <command> where <command> is one of: access, adduser, audit, bin, bugs, c, cache, ci, cit, clean-install, clean-install-test, completion, config, create, ddp, dedupe, deprecate, dist-tag, docs, doctor, edit, explore, get, help, help-search, hook, i, init, install, install-ci-test, install-test, it, link, list, ln, login, logout, ls, org, outdated, owner, pack, ping, prefix, profile, prune, publish, rb, rebuild, repo, restart, root, run, run-script, s, se, search, set, shrinkwrap, star, stars, start, stop, t, team, test, token, tst, un, uninstall, unpublish, unstar, up, update, v, version, view, whoami npm <command> -h quick help on <command> npm -l display full usage info npm help <term> search for help on <term> npm help npm involved overview Specify configs in the ini-formatted file: /path/.npmrc or on the command line via: npm <command> --key value Config info can be viewed via: npm help config $ npm config get registry prints: PRIVATE-VERDACCIO-REGISTRY Additional context The log file that was generated upon running of "npm audit" has the following content - cat .npm/_logs/2020-07-01T00_53_25_931Z-debug.log 0 info it worked if it ends with ok 1 verbose cli [ 1 verbose cli '/usr/local/Cellar/node/12.5.0/bin/node', 1 verbose cli '/usr/local/bin/npm', 1 verbose cli 'audit' 1 verbose cli ] 2 info using npm@6.9.0 3 info using node@v12.5.0 4 verbose npm-session 8821fc6732d58b82 5 http fetch POST 500 PRIVATE-VERDACCIO-REGISTRY/-/npm/v1/security/audits 15290ms 6 verbose stack Error: Your configured registry (PRIVATE-VERDACCIO-REGISTRY) does not support audit requests, or the audit endpoint is temporarily unavailable. 6 verbose stack at /usr/local/lib/node_modules/npm/lib/audit.js:201:18 6 verbose stack at tryCatcher (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/util.js:16:23) 6 verbose stack at Promise._settlePromiseFromHandler (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/promise.js:512:31) 6 verbose stack at Promise._settlePromise (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/promise.js:569:18) 6 verbose stack at Promise._settlePromise0 (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/promise.js:614:10) 6 verbose stack at Promise._settlePromises (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/promise.js:690:18) 6 verbose stack at _drainQueueStep (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/async.js:138:12) 6 verbose stack at _drainQueue (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/async.js:131:9) 6 verbose stack at Async._drainQueues (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/async.js:147:5) 6 verbose stack at Immediate.Async.drainQueues [as _onImmediate] (/usr/local/lib/node_modules/npm/node_modules/bluebird/js/release/async.js:17:14) 6 verbose stack at processImmediate (internal/timers.js:439:21) 7 verbose cwd /xxxxx 8 verbose Darwin 19.4.0 9 verbose argv "/usr/local/Cellar/node/12.5.0/bin/node" "/usr/local/bin/npm" "audit" 10 verbose node v12.5.0 11 verbose npm v6.9.0 12 error code ENOAUDIT 13 error audit Your configured registry (PRIVATE-VERDACCIO-REGISTRY) does not support audit requests, or the audit endpoint is temporarily unavailable. 14 verbose exit [ 1, true ] @toolsofraj please try with Verdaccio 4.6.2. It might be a regression in v4.7.0 This causes the error https://github.com/verdaccio/verdaccio/pull/1841 cc @hydra13 😢 As far I could check, as we move the app.use(bodyParser.json({ strict: false, limit: config.max_body_size || '10mb' })); of position. When the request gets to the audit plugin the stream seems to be consumed and the pipe stops to work, thus you get 500, mostly the server npmjs cannot proceed with the request and the plugin response is 500 by default even if the error is 400 like in this case, we are sending the body already parsed to npmjs. I've found a similar problem in the ExpressJS issues: You can only read a stream one time in Node.js. req was already read into req.body by this module, so your req.pipe is what's hanging I see, thanks for sharing @hydra13 this is the code we have in audit ... do you tihnk would be possible follow the same recommendation 🤔 ? const fetchAudit = (req: Request, res: Response & { report_error?: Function }): void => { const headers = req.headers; headers.host = 'https://registry.npmjs.org/'; const requestOptions = { url: 'https://registry.npmjs.org/-/npm/v1/security/audits', method: req.method, proxy: auth.config.https_proxy, req, strictSSL: this.strict_ssl, }; req .pipe(request(requestOptions)) .on('error', err => { if (typeof res.report_error === 'function') { return res.report_error(err); } this.logger.error(err); return res.status(500).end(); }) .pipe(res); }; yes, now I'm preparing PR with fixing this problem for audit plugin @juanpicado, let's check this PR: https://github.com/verdaccio/monorepo/pull/371 same error with, how it going? same error with it, how it going? No solution by far, please use Verdaccio 4.6.2 or anything lower than v4.7.0 until this gets fixed. @juanpicado yes, use 4.6.2 fixed it. Well, resuming the issue. At https://github.com/verdaccio/verdaccio/pull/1841 we just moved the body-parser before the middleware are executed to they can have access to the parsed body, metadata, etc. This causes an issue on the audit middleware (maybe others as well). The body-parser middleware consume the stream thus is not possible to pipe with request or any other fetch tool. The audit plugin catch the request and cannot process it, returning 500. So, I'm here asking for people with experience on streams to find a good solution. If someone has an idea how to solve it, please share your thoughts. But right now, audit cannot be used with v4.7.0 and ahead, and I don't plan to release a minor version until I find a way to solve this. @hydra13 I'm thinking to rollback that PR, I could not find other way. We can re-apply it later when a solution is being available. @juanpicado, I agree with you, because I haven't enough time for solving this problem right now. I will come back here later.
gharchive/issue
2020-07-02T00:16:33
2025-04-01T06:46:09.876556
{ "authors": [ "cdllqos", "hydra13", "juanpicado", "toolsofraj" ], "repo": "verdaccio/verdaccio", "url": "https://github.com/verdaccio/verdaccio/issues/1866", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
180974011
Translate to pt-BR Could I translate this guide to brazilian portuguese? :D That would be great! You can just fork it and I will link to your repo from the main page :) Cool. I will fork it. :D @cicerohen Let me know when it's ready. Closing that issue in the meantime : )
gharchive/issue
2016-10-04T18:50:54
2025-04-01T06:46:09.879005
{ "authors": [ "cicerohen", "verekia" ], "repo": "verekia/modern-js-stack-training", "url": "https://github.com/verekia/modern-js-stack-training/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
523850244
SHA256 hashes shown for 6.0.0 release downloads, not SHA512 as indicated Issue: For 6.0.0 Release, Hashes shown are SHA256, yet are referenced as SHA512. -----BEGIN PGP SIGNED MESSAGE----- Hash: SHA512 86513d8c8721ab868b49d53a9c05f29f52d1692e13276223d2036b803bc3e6dd verge-6.0.0-osx-unsigned.dmg 093720a14755f21737975bc8e3ac5ed901f03ac84737c499fd5956404e66c76f verge-6.0.0-osx64.tar.gz d10c48f5f01c1eb0c43c727417af61062f6033d59aa3ffb603ba06a7c8cb363d verge-6.0.0-win64-debug.zip b9e9bede576378e469e3a78071237ba0791ea90c102c1003e31d9931eaadcb5c verge-6.0.0-win64-setup-unsigned.exe 87506181a1a9d8823ccffdc51cc72199084a63c6fc1053eea6e8ab4e20849a8f verge-6.0.0-win64.zip d1f560355d301c249d0afb9c0c7ff57e29fd15bf6033e38778462fbddfdf2e43 verge-6.0.0-x86_64-linux-gnu-debug.tar.gz c9d3fbce999a88d8a484248fa67df9b0e566aa79e57c05a5a7dae4e4443544dd verge-6.0.0-x86_64-linux-gnu.tar.gz d85806ff3f89b436b6b2933a74fe64b2c6499364d5a68fa6811a81f850bbe056 verge-6.0.0.tar.gz -----BEGIN PGP SIGNATURE----- Oh sorry, replaced the hash function manually for now. I will try to update the hash procedure later on.
gharchive/issue
2019-11-16T14:14:53
2025-04-01T06:46:09.883148
{ "authors": [ "marpme", "todw1fd" ], "repo": "vergecurrency/VERGE", "url": "https://github.com/vergecurrency/VERGE/issues/999", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
859554313
旁路由下openclash无法更新订阅,提示网络失败 OpenClash 调试日志 生成时间: 2021-04-16 15:36:54 插件版本: v0.42.04-beta 隐私提示: 上传此日志前请注意检查、屏蔽公网IP、节点、密码等相关敏感信息 #===================== 系统信息 =====================# 主机型号: QEMU Virtual CPU version 2.5+ : 1 Core 1 Thread 固件版本: OpenWrt SNAPSHOT r0-5dcbd82 LuCI版本: git-21.096.35788-b200c7b-1 内核版本: 5.4.110 处理器架构: x86_64 #此项在使用Tun模式时应为ACCEPT 防火墙转发: REJECT #此项有值时建议到网络-接口-lan的设置中禁用IPV6的DHCP IPV6-DHCP: #此项结果应仅有配置文件的DNS监听地址 Dnsmasq转发设置: #===================== 依赖检查 =====================# dnsmasq-full: 已安装 coreutils: 已安装 coreutils-nohup: 已安装 bash: 已安装 curl: 已安装 jsonfilter: 已安装 ca-certificates: 已安装 ipset: 已安装 ip-full: 已安装 iptables-mod-tproxy: 已安装 kmod-ipt-tproxy: 已安装 iptables-mod-extra: 已安装 kmod-ipt-extra: 已安装 libcap: 未安装 libcap-bin: 未安装 ruby: 已安装 ruby-yaml: 已安装 ruby-psych: 已安装 ruby-pstore: 已安装 ruby-dbm: 已安装 kmod-tun(TUN模式): 已安装 luci-compat(Luci-19.07): 已安装 #===================== 内核检查 =====================# 运行状态: 未运行 已选择的架构: linux-amd64 #下方无法显示内核版本号时请确认您的内核版本是否正确或者有无权限 Tun内核版本: Tun内核文件: 不存在 Tun内核运行权限: 否 Game内核版本: Game内核文件: 不存在 Game内核运行权限: 否 Dev内核版本: Dev内核文件: 不存在 Dev内核运行权限: 否 #===================== 插件设置 =====================# 当前配置文件: 启动配置文件: /etc/openclash/ 运行模式: redir-host 默认代理模式: rule UDP流量转发(tproxy): 启用 DNS劫持: 启用 自定义DNS: 停用 IPV6-DNS解析: 停用 禁用Dnsmasq缓存: 停用 自定义规则: 停用 仅允许内网: 停用 仅代理命中规则流量: 停用 仅允许常用端口流量: 停用 绕过中国大陆IP: 停用 #启动异常时建议关闭此项后重试 混合节点: 停用 保留配置: 停用 #启动异常时建议关闭此项后重试 第三方规则: 停用 #===================== 配置文件 =====================# #===================== 防火墙设置 =====================# #NAT chain #Mangle chain #===================== IPSET状态 =====================# Name: music Name: music_http Name: music_https #===================== 路由表状态 =====================# #route -n Kernel IP routing table Destination Gateway Genmask Flags Metric Ref Use Iface 0.0.0.0 192.168.1.254 0.0.0.0 UG 0 0 0 br-lan 192.168.1.0 0.0.0.0 255.255.255.0 U 0 0 0 br-lan #ip route list default via 192.168.1.254 dev br-lan proto static 192.168.1.0/24 dev br-lan proto kernel scope link src 192.168.1.252 #ip rule show 0: from all lookup local 32766: from all lookup main 32767: from all lookup default #===================== 端口占用状态 =====================# #===================== 测试本机DNS查询 =====================# Server: 127.0.0.1 Address: 127.0.0.1#53 Name: www.baidu.com www.baidu.com canonical name = www.a.shifen.com Name: www.a.shifen.com Address 1: 14.215.177.39 Address 2: 14.215.177.38 www.baidu.com canonical name = www.a.shifen.com #===================== resolv.conf.d =====================# # Interface lan nameserver 114.114.114.114 nameserver 222.172.200.68 #===================== 测试本机网络连接 =====================# #===================== 测试本机网络下载 =====================# #===================== 最近运行日志 =====================# 2021-04-16 15:34:08 Warning: OpenClash Now Disabled, Need Start From Luci Page, Exit... 2021-04-16 15:34:31 Error: OpenClash 【Game】 Core Update Error 2021-04-16 15:34:31 Error: 【TUN】Core Version Check Error, Please Try Again After A few Seconds 2021-04-16 15:34:31 Error: OpenClash 【Dev】 Core Update Error 2021-04-16 15:34:39 Error: OpenClash 【Dev】 Core Update Error 2021-04-16 15:34:41 OpenClash Version Check Error, Please Try Again After A few seconds 2021-04-16 15:35:42 Error: Config Not Found 2021-04-16 15:36:02 Error: Config 【config】 Download Faild 2021-04-16 15:36:02 Error: Config 【config】Update Error 但是检测的时候,百度那些又都是显示正常的呢。 旁路由lan口的IP地址要设置下,路由自己要能上网 旁路由lan口的IP地址要设置下,路由自己要能上网 旁路由的IP,已经自己手动设置了一个IP.网关是主路由IP。DNS是当地运营商的。。 反正就是奇怪的很。就是更新不了任何文件呢 可能curl的问题,你的依赖没装好 折腾了2天,发现是因为主路由开启了IPV6呢,关闭一切IPV6就正常了。。。。这个坑踩的太深了。 折腾了2天,发现是因为主关闭开启了IPV6呢,一切IPV6就正常了。。。。这个坑踩的太深了。 能否说明一下如何关闭 IPv6? 说来也奇怪,我主路由本身是支持 IPv6 的。 这 OpenWrt 翻译有些实在没看懂。 lan口关闭就行了,编译时不选哪个dch6-only也行
gharchive/issue
2021-04-16T07:39:09
2025-04-01T06:46:09.920496
{ "authors": [ "hcym", "piaoyun", "vernesong", "wclebb" ], "repo": "vernesong/OpenClash", "url": "https://github.com/vernesong/OpenClash/issues/1359", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2380331276
访问部分网站证书不对应,页面指向到openwrt,关闭op后访问正常 Verify Steps [X] Tracker 我已经在 Issue Tracker 中找过我要提出的问题 [X] Branch 我知道 OpenClash 的 Dev 分支切换开关位于插件设置-版本更新中,或者我会手动下载并安装 Dev 分支的 OpenClash [X] Latest 我已经使用最新 Dev 版本测试过,问题依旧存在 [X] Relevant 我知道 OpenClash 与 内核(Core)、控制面板(Dashboard)、在线订阅转换(Subconverter)等项目之间无直接关系,仅相互调用 [X] Definite 这确实是 OpenClash 出现的问题 [ ] Contributors 我有能力协助 OpenClash 开发并解决此问题 [ ] Meaningless 我提交的是无意义的催促更新或修复请求 OpenClash Version v0.46.014-beta Bug on Environment Other OpenWrt Version OpenWrt 08.23.2023 by Kiddin' / LuCI Master git-24.234.21016-c65420d Bug on Platform Linux-amd64(x86-64) Describe the Bug 访问部分网站时,chrome提示如下: 您的连接不是私密连接 ,攻击者可能会试图从 www.uy5.net 窃取您的信息(例如:密码、通讯内容或信用卡信息)。了解详情 ,NET::ERR_CERT_COMMON_NAME_INVALID。www.uy5.net 通常会使用加密技术来保护您的信息。Chrome 此次尝试连接到 www.uy5.net 时,该网站发回了异常的错误凭据。这可能是因为有攻击者在试图冒充 www.uy5.net,或者 Wi-Fi 登录屏幕中断了此次连接。请放心,您的信息仍然是安全的,因为 Chrome 尚未进行任何数据交换便停止了连接。 您目前无法访问 www.uy5.net,因为此网站使用了 HSTS。网络错误和攻击通常是暂时的,因此,此网页稍后可能会恢复正常。 ping不通www.uy5.net,尝试使用nslookup,结果如下,DNS指向到op: 服务器: op Address: 10.0.0.2 名称: www.uy5.net Address: 198.18.1.205 尝试在op里调试日志测试连接,测试DNS,结果如下: 找不到任何连接日志! 可能是插件未在运行 可能是缓存导致浏览直接使用 IP 地址进行访问 可能是 DNS 未劫持成功,导致 Clash 无法正确反推出域名连接 可能是所填地址无法进行解析和连接 Status: 0 TC: false RD: true RA: true AD: false CD: false Question: Name: www.uy5.net. Qtype: 1 Qclass: 1 Answer: TTL: 1 data: 61.160.148.90 name: www.uy5.net. type: 1 Status: 0 TC: false RD: true RA: true AD: false CD: false Question: Name: www.uy5.net. Qtype: 28 Qclass: 1 Answer: TTL: 1 data: ::1 name: www.uy5.net. type: 28 To Reproduce 不清楚,只有部分网站一直出现该问题,如www.uy5.net和myavxx.xyz。当我关闭openclash后就可以正常访问了。 OpenClash Log OpenClash 调试日志 生成时间: 2024-06-28 20:13:04 插件版本: v0.46.014-beta 隐私提示: 上传此日志前请注意检查、屏蔽公网IP、节点、密码等相关敏感信息 #===================== 系统信息 =====================# 主机型号: QEMU Standard PC (i440FX + PIIX, 1996) 固件版本: OpenWrt 23.05.0-rc3 08.22.2023 LuCI版本: git-23.051.66410-a505bb1 内核版本: 5.15.127 处理器架构: x86_64 #此项有值时,如不使用IPv6,建议到网络-接口-lan的设置中禁用IPV6的DHCP IPV6-DHCP: DNS劫持: Dnsmasq 转发 #DNS劫持为Dnsmasq时,此项结果应仅有配置文件的DNS监听地址 Dnsmasq转发设置: 127.0.0.1#7874 #===================== 依赖检查 =====================# dnsmasq-full: 已安装 coreutils: 已安装 coreutils-nohup: 已安装 bash: 已安装 curl: 已安装 ca-certificates: 已安装 ipset: 已安装 ip-full: 已安装 libcap: 已安装 libcap-bin: 已安装 ruby: 已安装 ruby-yaml: 已安装 ruby-psych: 已安装 ruby-pstore: 已安装 kmod-tun(TUN模式): 已安装 luci-compat(Luci >= 19.07): 已安装 kmod-inet-diag(PROCESS-NAME): 已安装 unzip: 已安装 iptables-mod-tproxy: 已安装 kmod-ipt-tproxy: 已安装 iptables-mod-extra: 已安装 kmod-ipt-extra: 已安装 kmod-ipt-nat: 已安装 #===================== 内核检查 =====================# 运行状态: 运行中 运行内核:TUN 进程pid: 29370 运行权限: 29370: cap_dac_override,cap_net_bind_service,cap_net_admin,cap_net_raw,cap_sys_ptrace,cap_sys_resource=eip 运行用户: nobody 已选择的架构: linux-amd64 #下方无法显示内核版本号时请确认您的内核版本是否正确或者有无权限 Tun内核版本: 2023.08.17-13-gdcc8d87 Tun内核文件: 存在 Tun内核运行权限: 正常 Dev内核版本: v1.18.0-13-gd034a40 Dev内核文件: 存在 Dev内核运行权限: 正常 Meta内核版本: alpha-g0d4e57c Meta内核文件: 存在 Meta内核运行权限: 正常 #===================== 插件设置 =====================# 当前配置文件: /etc/openclash/config/老猫云.yaml 启动配置文件: /etc/openclash/老猫云.yaml 运行模式: fake-ip 默认代理模式: rule UDP流量转发(tproxy): 启用 自定义DNS: 启用 IPV6代理: 启用 IPV6-DNS解析: 启用 禁用Dnsmasq缓存: 启用 自定义规则: 启用 仅允许内网: 停用 仅代理命中规则流量: 启用 仅允许常用端口流量: 停用 绕过中国大陆IP: 启用 路由本机代理: 启用 #启动异常时建议关闭此项后重试 混合节点: 停用 保留配置: 启用 #启动异常时建议关闭此项后重试 第三方规则: 停用 #===================== 自定义规则 一 =====================# script: ## shortcuts: ## Notice: The core timezone is UTC ## CST 20:00-24:00 = time.now().hour > 12 and time.now().hour < 16 ## 内核时区为UTC,故以下time.now()函数的取值需要根据本地时区进行转换 ## 北京时间(CST) 20:00-24:00 = time.now().hour > 12 and time.now().hour < 16 ## quic: network == 'udp' and dst_port == 443 and (geoip(resolve_ip(host)) != 'CN' or geoip(dst_ip) != 'CN') ## time-limit: in_cidr(src_ip,'192.168.1.2/32') and time.now().hour < 20 or time.now().hour > 21 ## time-limit: src_ip == '192.168.1.2' and time.now().hour < 20 or time.now().hour > 21 ## code: | ## def main(ctx, metadata): ## directkeywordlist = ["baidu"] ## for directkeyword in directkeywordlist: ## if directkeyword in metadata["host"]: ## ctx.log('[Script] matched keyword %s use direct' % directkeyword) ## return "DIRECT" rules: ##- SCRIPT,quic,REJECT #shortcuts rule ##- SCRIPT,time-limit,REJECT #shortcuts rule ##- PROCESS-NAME,curl,DIRECT #匹配路由自身进程(curl直连) ##- DOMAIN-SUFFIX,google.com,Proxy #匹配域名后缀(交由Proxy代理服务器组) ##- DOMAIN-KEYWORD,google,Proxy #匹配域名关键字(交由Proxy代理服务器组) ##- DOMAIN,google.com,Proxy #匹配域名(交由Proxy代理服务器组) ##- DOMAIN-SUFFIX,ad.com,REJECT #匹配域名后缀(拒绝) ##- IP-CIDR,127.0.0.0/8,DIRECT #匹配数据目标IP(直连) ##- SRC-IP-CIDR,192.168.1.201/32,DIRECT #匹配数据发起IP(直连) ##- DST-PORT,80,DIRECT #匹配数据目标端口(直连) ##- SRC-PORT,7777,DIRECT #匹配数据源端口(直连) ##排序在上的规则优先生效,如添加(去除规则前的#号): ##IP段:192.168.1.2-192.168.1.200 直连 ##- SRC-IP-CIDR,192.168.1.2/31,DIRECT ##- SRC-IP-CIDR,192.168.1.4/30,DIRECT ##- SRC-IP-CIDR,192.168.1.8/29,DIRECT ##- SRC-IP-CIDR,192.168.1.16/28,DIRECT ##- SRC-IP-CIDR,192.168.1.32/27,DIRECT ##- SRC-IP-CIDR,192.168.1.64/26,DIRECT ##- SRC-IP-CIDR,192.168.1.128/26,DIRECT ##- SRC-IP-CIDR,192.168.1.192/29,DIRECT ##- SRC-IP-CIDR,192.168.1.200/32,DIRECT ##IP段:192.168.1.202-192.168.1.255 直连 ##- SRC-IP-CIDR,192.168.1.202/31,DIRECT ##- SRC-IP-CIDR,192.168.1.204/30,DIRECT ##- SRC-IP-CIDR,192.168.1.208/28,DIRECT ##- SRC-IP-CIDR,192.168.1.224/27,DIRECT ##此时IP为192.168.1.1和192.168.1.201的客户端流量走代理(策略),其余客户端不走代理 ##因为Fake-IP模式下,IP地址为192.168.1.1的路由器自身流量可走代理(策略),所以需要排除 ##仅设置路由器自身直连: ##- SRC-IP-CIDR,192.168.1.1/32,DIRECT ##- SRC-IP-CIDR,198.18.0.1/32,DIRECT ##DDNS ##- DOMAIN-SUFFIX,checkip.dyndns.org,DIRECT ##- DOMAIN-SUFFIX,checkipv6.dyndns.org,DIRECT ##- DOMAIN-SUFFIX,checkip.synology.com,DIRECT ##- DOMAIN-SUFFIX,ifconfig.co,DIRECT ##- DOMAIN-SUFFIX,api.myip.com,DIRECT ##- DOMAIN-SUFFIX,ip-api.com,DIRECT ##- DOMAIN-SUFFIX,ipapi.co,DIRECT ##- DOMAIN-SUFFIX,ip6.seeip.org,DIRECT ##- DOMAIN-SUFFIX,members.3322.org,DIRECT ##在线IP段转CIDR地址:http://ip2cidr.com ##############################################################---PT站点 - DOMAIN-SUFFIX,themoviedb.org,Proxies #刮削 - DOMAIN-SUFFIX,m-team.cc,DIRECT #馒头 - DOMAIN-SUFFIX,m-team.io,DIRECT #馒头 - DOMAIN-SUFFIX,hddolby.com,DIRECT #高清 - DOMAIN-SUFFIX,hdatmos.club,DIRECT #阿童木 - DOMAIN-SUFFIX,btschool.club,DIRECT #学校 ##############################################################---VPN ##- DOMAIN-SUFFIX,dler.io,DIRECT #订阅模板 ##- DOMAIN-SUFFIX,xn--mest5a943ag8x.xyz,DIRECT #翻墙 - DOMAIN-SUFFIX,一元机场.com,DIRECT #一元机场 - DOMAIN-SUFFIX,patriot.ninja,DIRECT #BTW ##############################################################---小说 - DOMAIN-SUFFIX,360cdnjiasu.com,DIRECT - DOMAIN-SUFFIX,zwwx.org,DIRECT - DOMAIN-SUFFIX,yuzhaiwu.uk,DIRECT - DOMAIN-SUFFIX,xsyq.cc,DIRECT ##############################################################---漫画 - DOMAIN-SUFFIX,18comic.vip,Netflix #禁漫天堂 - DOMAIN-SUFFIX,manhuagui.com,Netflix #漫画柜 - DOMAIN-SUFFIX,hanime1.me,Netflix # ##############################################################---影视剧 - DOMAIN-SUFFIX,zjtu.tv,DIRECT #追剧兔 - DOMAIN-SUFFIX,91porn.com,Proxies #91 ##############################################################---游戏 - DOMAIN-SUFFIX,5eplay.com,DIRECT #5E ##############################################################---其他 - DOMAIN-SUFFIX,supes.top,DIRECT #OP固件 - DOMAIN-SUFFIX,openwrt.ai,DIRECT #OP固件 ##- DOMAIN-SUFFIX,speedtest.net,Proxies #测速 - DOMAIN-SUFFIX,xtatcha.com,DIRECT #群晖&openclash ##- DOMAIN-SUFFIX,github.com,DIRECT #github - DOMAIN-SUFFIX,githubusercontent.com,Proxies #github仓库 - DOMAIN-SUFFIX,ghproxy.com,Proxies #等待验证??? - DOMAIN-SUFFIX,debian.org,Proxies #debian - DOMAIN-SUFFIX,snipaste.com,DIRECT #截图snipaste - DOMAIN-SUFFIX,syncthing.net,DIRECT #syncthing同步 - DOMAIN-SUFFIX,pi-hole.net,DIRECT #pi-hole ##- DOMAIN-SUFFIX,axutongxue.net,DIRECT #阿虚同学的储物柜 ##- DOMAIN-SUFFIX,uy5.net,DIRECT #克隆窝 - DOMAIN-SUFFIX,jnoljinugtfc12.buzz,Proxies #Bemoter跨境电商 - DOMAIN-SUFFIX,chatgpt.com,YouTube #ChatGPT #===================== 自定义覆写设置 =====================# #!/bin/sh . /usr/share/openclash/ruby.sh . /usr/share/openclash/log.sh . /lib/functions.sh # This script is called by /etc/init.d/openclash # Add your custom overwrite scripts here, they will be take effict after the OpenClash own srcipts LOG_OUT "Tip: Start Running Custom Overwrite Scripts..." LOGTIME=$(echo $(date "+%Y-%m-%d %H:%M:%S")) LOG_FILE="/tmp/openclash.log" CONFIG_FILE="$1" #config path #Simple Demo: #General Demo #1--config path #2--key name #3--value #ruby_edit "$CONFIG_FILE" "['redir-port']" "7892" #ruby_edit "$CONFIG_FILE" "['secret']" "123456" #ruby_edit "$CONFIG_FILE" "['dns']['enable']" "true" #Hash Demo #1--config path #2--key name #3--hash type value #ruby_edit "$CONFIG_FILE" "['experimental']" "{'sniff-tls-sni'=>true}" #ruby_edit "$CONFIG_FILE" "['sniffer']" "{'sniffing'=>['tls','http']}" #Array Demo: #1--config path #2--key name #3--position(start from 0, end with -1) #4--value #ruby_arr_insert "$CONFIG_FILE" "['dns']['nameserver']" "0" "114.114.114.114" #Array Add From Yaml File Demo: #1--config path #2--key name #3--position(start from 0, end with -1) #4--value file path #5--value key name in #4 file #ruby_arr_add_file "$CONFIG_FILE" "['dns']['fallback-filter']['ipcidr']" "0" "/etc/openclash/custom/openclash_custom_fallback_filter.yaml" "['fallback-filter']['ipcidr']" #Ruby Script Demo: #ruby -ryaml -rYAML -I "/usr/share/openclash" -E UTF-8 -e " # begin # Value = YAML.load_file('$CONFIG_FILE'); # rescue Exception => e # puts '${LOGTIME} Error: Load File Failed,【' + e.message + '】'; # end; #General # begin # Thread.new{ # Value['redir-port']=7892; # Value['tproxy-port']=7895; # Value['port']=7890; # Value['socks-port']=7891; # Value['mixed-port']=7893; # }.join; # rescue Exception => e # puts '${LOGTIME} Error: Set General Failed,【' + e.message + '】'; # ensure # File.open('$CONFIG_FILE','w') {|f| YAML.dump(Value, f)}; # end" 2>/dev/null >> $LOG_FILE exit 0 #===================== 自定义防火墙设置 =====================# #!/bin/sh . /usr/share/openclash/log.sh . /lib/functions.sh # This script is called by /etc/init.d/openclash # Add your custom firewall rules here, they will be added after the end of the OpenClash iptables rules LOG_OUT "Tip: Start Add Custom Firewall Rules..." exit 0 #===================== IPTABLES 防火墙设置 =====================# #IPv4 NAT chain # Generated by iptables-save v1.8.7 on Fri Jun 28 20:13:07 2024 *nat :PREROUTING ACCEPT [10777:1239672] :INPUT ACCEPT [8587:941418] :OUTPUT ACCEPT [10205:639673] :POSTROUTING ACCEPT [1273:84156] :MINIUPNPD - [0:0] :MINIUPNPD-POSTROUTING - [0:0] :openclash - [0:0] :openclash_output - [0:0] :postrouting_lan_rule - [0:0] :postrouting_rule - [0:0] :prerouting_lan_rule - [0:0] :prerouting_rule - [0:0] :zone_lan_postrouting - [0:0] :zone_lan_prerouting - [0:0] -A PREROUTING -d 8.8.4.4/32 -p tcp -m comment --comment "OpenClash Google DNS Hijack" -m tcp --dport 53 -j REDIRECT --to-ports 7892 -A PREROUTING -d 8.8.8.8/32 -p tcp -m comment --comment "OpenClash Google DNS Hijack" -m tcp --dport 53 -j REDIRECT --to-ports 7892 -A PREROUTING -p tcp -m tcp --dport 53 -m comment --comment "OpenClash DNS Hijack" -j REDIRECT --to-ports 53 -A PREROUTING -p udp -m udp --dport 53 -m comment --comment "OpenClash DNS Hijack" -j REDIRECT --to-ports 53 -A PREROUTING -m comment --comment "!fw3: Custom prerouting rule chain" -j prerouting_rule -A PREROUTING -i br-lan -m comment --comment "!fw3" -j zone_lan_prerouting -A PREROUTING -p tcp -j openclash -A OUTPUT -j openclash_output -A POSTROUTING -o eth0 -j MASQUERADE -A POSTROUTING -m comment --comment "!fw3: Custom postrouting rule chain" -j postrouting_rule -A POSTROUTING -o br-lan -m comment --comment "!fw3" -j zone_lan_postrouting -A MINIUPNPD -p udp -m udp --dport 8568 -j DNAT --to-destination 10.0.0.64:8568 -A MINIUPNPD -p udp -m udp --dport 8629 -j DNAT --to-destination 10.0.0.11:8629 -A MINIUPNPD -p udp -m udp --dport 8567 -j DNAT --to-destination 10.0.0.11:8567 -A MINIUPNPD -p udp -m udp --dport 8587 -j DNAT --to-destination 10.0.0.63:8568 -A MINIUPNPD -p udp -m udp --dport 8657 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD -p udp -m udp --dport 8656 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD -p udp -m udp --dport 8592 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD -p udp -m udp --dport 8573 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD -p udp -m udp --dport 8579 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD -p udp -m udp --dport 8665 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD -p tcp -m tcp --dport 41573 -j DNAT --to-destination 10.0.0.15:22000 -A MINIUPNPD -p tcp -m tcp --dport 26066 -j DNAT --to-destination 10.0.0.15:22000 -A MINIUPNPD -p udp -m udp --dport 8614 -j DNAT --to-destination 10.0.0.63:8567 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8568 -j MASQUERADE --to-ports 8587 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8657 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8656 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8592 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8573 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8579 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8665 -A MINIUPNPD-POSTROUTING -s 10.0.0.15/32 -p tcp -m tcp --sport 22000 -j MASQUERADE --to-ports 41573 -A MINIUPNPD-POSTROUTING -s 10.0.0.15/32 -p tcp -m tcp --sport 22000 -j MASQUERADE --to-ports 26066 -A MINIUPNPD-POSTROUTING -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j MASQUERADE --to-ports 8614 -A openclash -m set --match-set localnetwork dst -j RETURN -A openclash -d 198.18.0.0/16 -p tcp -j REDIRECT --to-ports 7892 -A openclash -m set --match-set china_ip_route dst -m set ! --match-set china_ip_route_pass dst -j RETURN -A openclash -p tcp -j REDIRECT --to-ports 7892 -A openclash_output -d 198.18.0.0/16 -p tcp -m owner ! --uid-owner 65534 -j REDIRECT --to-ports 7892 -A openclash_output -m set --match-set localnetwork dst -j RETURN -A openclash_output -m owner ! --uid-owner 65534 -m set --match-set china_ip_route dst -m set ! --match-set china_ip_route_pass dst -j RETURN -A openclash_output -p tcp -m owner ! --uid-owner 65534 -j REDIRECT --to-ports 7892 -A zone_lan_postrouting -j MINIUPNPD-POSTROUTING -A zone_lan_postrouting -j MINIUPNPD-POSTROUTING -A zone_lan_postrouting -m comment --comment "!fw3: Custom lan postrouting rule chain" -j postrouting_lan_rule -A zone_lan_postrouting -m comment --comment "!fw3" -j FULLCONENAT -A zone_lan_prerouting -j MINIUPNPD -A zone_lan_prerouting -j MINIUPNPD -A zone_lan_prerouting -m comment --comment "!fw3: Custom lan prerouting rule chain" -j prerouting_lan_rule -A zone_lan_prerouting -m comment --comment "!fw3" -j FULLCONENAT COMMIT # Completed on Fri Jun 28 20:13:07 2024 #IPv4 Mangle chain # Generated by iptables-save v1.8.7 on Fri Jun 28 20:13:07 2024 *mangle :PREROUTING ACCEPT [1967458:1390469759] :INPUT ACCEPT [656613:741235110] :FORWARD ACCEPT [1328395:651042481] :OUTPUT ACCEPT [606113:728530006] :POSTROUTING ACCEPT [1877109:1372924368] :openclash - [0:0] :openclash_output - [0:0] :openclash_upnp - [0:0] -A PREROUTING -p udp -j openclash -A OUTPUT -p udp -j openclash_output -A openclash -i lo -j RETURN -A openclash -m set --match-set localnetwork dst -j RETURN -A openclash -p udp -m udp --dport 53 -j RETURN -A openclash -d 198.18.0.0/16 -p udp -j TPROXY --on-port 7895 --on-ip 0.0.0.0 --tproxy-mark 0x162/0xffffffff -A openclash -m set --match-set china_ip_route dst -m set ! --match-set china_ip_route_pass dst -j RETURN -A openclash -p udp -j openclash_upnp -A openclash -p udp -j TPROXY --on-port 7895 --on-ip 0.0.0.0 --tproxy-mark 0x162/0xffffffff -A openclash_output -d 198.18.0.0/16 -p udp -m owner ! --uid-owner 65534 -j MARK --set-xmark 0x162/0xffffffff -A openclash_upnp -s 10.0.0.64/32 -p udp -m udp --sport 8568 -j RETURN -A openclash_upnp -s 10.0.0.11/32 -p udp -m udp --sport 8629 -j RETURN -A openclash_upnp -s 10.0.0.11/32 -p udp -m udp --sport 8567 -j RETURN -A openclash_upnp -s 10.0.0.63/32 -p udp -m udp --sport 8568 -j RETURN -A openclash_upnp -s 10.0.0.63/32 -p udp -m udp --sport 8567 -j RETURN -A openclash_upnp -s 10.0.0.15/32 -p tcp -m tcp --sport 22000 -j RETURN COMMIT # Completed on Fri Jun 28 20:13:07 2024 #IPv4 Filter chain # Generated by iptables-save v1.8.7 on Fri Jun 28 20:13:07 2024 *filter :INPUT ACCEPT [0:0] :FORWARD ACCEPT [0:0] :OUTPUT ACCEPT [0:0] :LUCKY - [0:0] :MINIUPNPD - [0:0] :forwarding_lan_rule - [0:0] :forwarding_rule - [0:0] :input_lan_rule - [0:0] :input_rule - [0:0] :output_lan_rule - [0:0] :output_rule - [0:0] :reject - [0:0] :syn_flood - [0:0] :zone_lan_dest_ACCEPT - [0:0] :zone_lan_forward - [0:0] :zone_lan_input - [0:0] :zone_lan_output - [0:0] :zone_lan_src_ACCEPT - [0:0] -A INPUT -p udp -m udp --dport 443 -m comment --comment "OpenClash QUIC REJECT" -m set ! --match-set china_ip_route dst -j REJECT --reject-with icmp-port-unreachable -A INPUT -j LUCKY -A INPUT -i lo -m comment --comment "!fw3" -j ACCEPT -A INPUT -m comment --comment "!fw3: Custom input rule chain" -j input_rule -A INPUT -m conntrack --ctstate RELATED,ESTABLISHED -m comment --comment "!fw3" -j ACCEPT -A INPUT -p tcp -m tcp --tcp-flags FIN,SYN,RST,ACK SYN -m comment --comment "!fw3" -j syn_flood -A INPUT -i br-lan -m comment --comment "!fw3" -j zone_lan_input -A FORWARD -m comment --comment "!fw3: Custom forwarding rule chain" -j forwarding_rule -A FORWARD -m conntrack --ctstate RELATED,ESTABLISHED -m comment --comment "!fw3" -j ACCEPT -A FORWARD -i br-lan -m comment --comment "!fw3" -j zone_lan_forward -A OUTPUT -o lo -m comment --comment "!fw3" -j ACCEPT -A OUTPUT -m comment --comment "!fw3: Custom output rule chain" -j output_rule -A OUTPUT -m conntrack --ctstate RELATED,ESTABLISHED -m comment --comment "!fw3" -j ACCEPT -A OUTPUT -o br-lan -m comment --comment "!fw3" -j zone_lan_output -A LUCKY -p tcp -m tcp --dport 9951 -j ACCEPT -A LUCKY -p tcp -m tcp --dport 19951 -j ACCEPT -A MINIUPNPD -d 10.0.0.64/32 -p udp -m udp --dport 8568 -j ACCEPT -A MINIUPNPD -d 10.0.0.11/32 -p udp -m udp --dport 8629 -j ACCEPT -A MINIUPNPD -d 10.0.0.11/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8568 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A MINIUPNPD -d 10.0.0.15/32 -p tcp -m tcp --dport 22000 -j ACCEPT -A MINIUPNPD -d 10.0.0.15/32 -p tcp -m tcp --dport 22000 -j ACCEPT -A MINIUPNPD -d 10.0.0.63/32 -p udp -m udp --dport 8567 -j ACCEPT -A reject -p tcp -m comment --comment "!fw3" -j REJECT --reject-with tcp-reset -A reject -m comment --comment "!fw3" -j REJECT --reject-with icmp-port-unreachable -A syn_flood -m limit --limit 25/sec --limit-burst 50 -m comment --comment "!fw3" -j RETURN -A syn_flood -m comment --comment "!fw3" -j DROP -A zone_lan_dest_ACCEPT -o br-lan -m conntrack --ctstate INVALID -m comment --comment "!fw3: Prevent NAT leakage" -j DROP -A zone_lan_dest_ACCEPT -o br-lan -m comment --comment "!fw3" -j ACCEPT -A zone_lan_forward -j MINIUPNPD -A zone_lan_forward -j MINIUPNPD -A zone_lan_forward -m comment --comment "!fw3: Custom lan forwarding rule chain" -j forwarding_lan_rule -A zone_lan_forward -p tcp -m comment --comment "!fw3: 旁路由" -j zone_lan_dest_ACCEPT -A zone_lan_forward -p udp -m comment --comment "!fw3: 旁路由" -j zone_lan_dest_ACCEPT -A zone_lan_forward -m conntrack --ctstate DNAT -m comment --comment "!fw3: Accept port forwards" -j ACCEPT -A zone_lan_forward -m comment --comment "!fw3" -j zone_lan_dest_ACCEPT -A zone_lan_input -m comment --comment "!fw3: Custom lan input rule chain" -j input_lan_rule -A zone_lan_input -m conntrack --ctstate DNAT -m comment --comment "!fw3: Accept port redirections" -j ACCEPT -A zone_lan_input -m comment --comment "!fw3" -j zone_lan_src_ACCEPT -A zone_lan_output -m comment --comment "!fw3: Custom lan output rule chain" -j output_lan_rule -A zone_lan_output -m comment --comment "!fw3" -j zone_lan_dest_ACCEPT -A zone_lan_src_ACCEPT -i br-lan -m conntrack --ctstate NEW,UNTRACKED -m comment --comment "!fw3" -j ACCEPT COMMIT # Completed on Fri Jun 28 20:13:07 2024 #IPv6 NAT chain # Generated by ip6tables-save v1.8.7 on Fri Jun 28 20:13:07 2024 *nat :PREROUTING ACCEPT [2786:245423] :INPUT ACCEPT [2756:241313] :OUTPUT ACCEPT [4783:430353] :POSTROUTING ACCEPT [4783:430353] :openclash_output - [0:0] -A PREROUTING -d 2001:4860:4860::8844/128 -p tcp -m comment --comment "OpenClash Google DNS Hijack" -m tcp --dport 53 -j ACCEPT -A PREROUTING -d 2001:4860:4860::8888/128 -p tcp -m comment --comment "OpenClash Google DNS Hijack" -m tcp --dport 53 -j ACCEPT -A PREROUTING -p tcp -m tcp --dport 53 -m comment --comment "OpenClash DNS Hijack" -j REDIRECT --to-ports 53 -A PREROUTING -p udp -m udp --dport 53 -m comment --comment "OpenClash DNS Hijack" -j REDIRECT --to-ports 53 -A OUTPUT -j openclash_output -A openclash_output -m set --match-set localnetwork6 dst -j RETURN -A openclash_output -m owner ! --uid-owner 65534 -m set --match-set china_ip6_route dst -m set ! --match-set china_ip6_route_pass dst -j RETURN -A openclash_output -p tcp -m owner ! --uid-owner 65534 -j REDIRECT --to-ports 7892 COMMIT # Completed on Fri Jun 28 20:13:07 2024 #IPv6 Mangle chain # Generated by ip6tables-save v1.8.7 on Fri Jun 28 20:13:07 2024 *mangle :PREROUTING ACCEPT [65463:92607214] :INPUT ACCEPT [64611:92536204] :FORWARD ACCEPT [0:0] :OUTPUT ACCEPT [46194:6514554] :POSTROUTING ACCEPT [46238:6520230] :openclash - [0:0] -A PREROUTING -j openclash -A openclash -i lo -j RETURN -A openclash -m set --match-set localnetwork6 dst -j RETURN -A openclash -p udp -m udp --dport 53 -j RETURN -A openclash -m set --match-set china_ip6_route dst -m set ! --match-set china_ip6_route_pass dst -j RETURN -A openclash -p tcp -m comment --comment "OpenClash TCP Tproxy" -j TPROXY --on-port 7895 --on-ip :: --tproxy-mark 0x162/0xffffffff -A openclash -p udp -m comment --comment "OpenClash UDP Tproxy" -j TPROXY --on-port 7895 --on-ip :: --tproxy-mark 0x162/0xffffffff COMMIT # Completed on Fri Jun 28 20:13:07 2024 #IPv6 Filter chain # Generated by ip6tables-save v1.8.7 on Fri Jun 28 20:13:07 2024 *filter :INPUT ACCEPT [0:0] :FORWARD ACCEPT [0:0] :OUTPUT ACCEPT [0:0] :LUCKY - [0:0] :MINIUPNPD - [0:0] :forwarding_lan_rule - [0:0] :forwarding_rule - [0:0] :input_lan_rule - [0:0] :input_rule - [0:0] :output_lan_rule - [0:0] :output_rule - [0:0] :reject - [0:0] :syn_flood - [0:0] :zone_lan_dest_ACCEPT - [0:0] :zone_lan_forward - [0:0] :zone_lan_input - [0:0] :zone_lan_output - [0:0] :zone_lan_src_ACCEPT - [0:0] -A INPUT -p udp -m udp --dport 443 -m comment --comment "OpenClash QUIC REJECT" -m set ! --match-set china_ip6_route dst -j REJECT --reject-with icmp6-port-unreachable -A INPUT -j LUCKY -A INPUT -i lo -m comment --comment "!fw3" -j ACCEPT -A INPUT -m comment --comment "!fw3: Custom input rule chain" -j input_rule -A INPUT -m conntrack --ctstate RELATED,ESTABLISHED -m comment --comment "!fw3" -j ACCEPT -A INPUT -p tcp -m tcp --tcp-flags FIN,SYN,RST,ACK SYN -m comment --comment "!fw3" -j syn_flood -A INPUT -i br-lan -m comment --comment "!fw3" -j zone_lan_input -A FORWARD -m comment --comment "!fw3: Custom forwarding rule chain" -j forwarding_rule -A FORWARD -m conntrack --ctstate RELATED,ESTABLISHED -m comment --comment "!fw3" -j ACCEPT -A FORWARD -i br-lan -m comment --comment "!fw3" -j zone_lan_forward -A OUTPUT -o lo -m comment --comment "!fw3" -j ACCEPT -A OUTPUT -m comment --comment "!fw3: Custom output rule chain" -j output_rule -A OUTPUT -m conntrack --ctstate RELATED,ESTABLISHED -m comment --comment "!fw3" -j ACCEPT -A OUTPUT -o br-lan -m comment --comment "!fw3" -j zone_lan_output -A LUCKY -p tcp -m tcp --dport 19950 -j ACCEPT -A LUCKY -p tcp -m tcp --dport 9951 -j ACCEPT -A LUCKY -p tcp -m tcp --dport 19951 -j ACCEPT -A reject -p tcp -m comment --comment "!fw3" -j REJECT --reject-with tcp-reset -A reject -m comment --comment "!fw3" -j REJECT --reject-with icmp6-port-unreachable -A syn_flood -m limit --limit 25/sec --limit-burst 50 -m comment --comment "!fw3" -j RETURN -A syn_flood -m comment --comment "!fw3" -j DROP -A zone_lan_dest_ACCEPT -o br-lan -m conntrack --ctstate INVALID -m comment --comment "!fw3: Prevent NAT leakage" -j DROP -A zone_lan_dest_ACCEPT -o br-lan -m comment --comment "!fw3" -j ACCEPT -A zone_lan_forward -j MINIUPNPD -A zone_lan_forward -j MINIUPNPD -A zone_lan_forward -m comment --comment "!fw3: Custom lan forwarding rule chain" -j forwarding_lan_rule -A zone_lan_forward -p tcp -m comment --comment "!fw3: 旁路由" -j zone_lan_dest_ACCEPT -A zone_lan_forward -p udp -m comment --comment "!fw3: 旁路由" -j zone_lan_dest_ACCEPT -A zone_lan_forward -m comment --comment "!fw3" -j zone_lan_dest_ACCEPT -A zone_lan_input -m comment --comment "!fw3: Custom lan input rule chain" -j input_lan_rule -A zone_lan_input -m comment --comment "!fw3" -j zone_lan_src_ACCEPT -A zone_lan_output -m comment --comment "!fw3: Custom lan output rule chain" -j output_lan_rule -A zone_lan_output -m comment --comment "!fw3" -j zone_lan_dest_ACCEPT -A zone_lan_src_ACCEPT -i br-lan -m conntrack --ctstate NEW,UNTRACKED -m comment --comment "!fw3" -j ACCEPT COMMIT # Completed on Fri Jun 28 20:13:07 2024 #===================== IPSET状态 =====================# Name: localnetwork Type: hash:net Revision: 7 Header: family inet hashsize 1024 maxelem 65536 bucketsize 12 initval 0xea875668 Size in memory: 944 References: 3 Number of entries: 10 Name: china_ip_route Type: hash:net Revision: 7 Header: family inet hashsize 2048 maxelem 1000000 bucketsize 12 initval 0x838e5ca9 Size in memory: 195512 References: 4 Number of entries: 7088 Name: china_ip_route_pass Type: hash:net Revision: 7 Header: family inet hashsize 1024 maxelem 1000000 bucketsize 12 initval 0x8918102a Size in memory: 464 References: 3 Number of entries: 0 Name: china_ip6_route Type: hash:net Revision: 7 Header: family inet6 hashsize 1024 maxelem 1000000 bucketsize 12 initval 0x1a6572ec Size in memory: 92544 References: 3 Number of entries: 2016 Name: china_ip6_route_pass Type: hash:net Revision: 7 Header: family inet6 hashsize 1024 maxelem 1000000 bucketsize 12 initval 0xfbddc031 Size in memory: 1248 References: 2 Number of entries: 0 Name: localnetwork6 Type: hash:net Revision: 7 Header: family inet6 hashsize 1024 maxelem 65536 bucketsize 12 initval 0x39ecb137 Size in memory: 2400 References: 2 Number of entries: 16 #===================== 路由表状态 =====================# #IPv4 #route -n Kernel IP routing table Destination Gateway Genmask Flags Metric Ref Use Iface 0.0.0.0 10.0.0.1 0.0.0.0 UG 0 0 0 br-lan 10.0.0.0 0.0.0.0 255.255.255.0 U 0 0 0 br-lan #ip route list default via 10.0.0.1 dev br-lan proto static 10.0.0.0/24 dev br-lan proto kernel scope link src 10.0.0.2 #ip rule show 0: from all lookup local 32765: from all fwmark 0x162 lookup 354 32766: from all lookup main 32767: from all lookup default #IPv6 #route -A inet6 Kernel IPv6 routing table Destination Next Hop Flags Metric Ref Use Iface ::/0 :: U 1024 1 0 lo ::/0 fe80::c877:6ff:fefb:bc95 UG 512 5 0 br-lan ::/0 fe80::c877:6ff:fefb:bc95 UG 512 1 0 br-lan 240e:3a1:8438:e190::/64 :: !n 2147483647 6 0 lo 240e:3a1:8439:ef00::/64 :: U 256 5 0 br-lan 240e:3a1:8439:ef00::/64 :: !n 2147483647 1 0 lo fd56:b34c:1a0e::/64 :: U 256 5 0 br-lan fd56:b34c:1a0e::/64 :: !n 2147483647 1 0 lo fe80::/64 :: U 256 6 0 br-lan ::/0 :: !n -1 2 0 lo ::1/128 :: Un 0 7 0 lo 240e:3a1:8439:ef00::/128 :: Un 0 3 0 br-lan 240e:3a1:8439:ef00:2c4a:99ff:feaa:2233/128 :: Un 0 7 0 br-lan fd56:b34c:1a0e::/128 :: Un 0 3 0 br-lan *WAN IP*:2233/128 :: Un 0 4 0 br-lan fe80::/128 :: Un 0 3 0 br-lan fe80::2c4a:99ff:feaa:2233/128 :: Un 0 7 0 br-lan ff00::/8 :: U 256 7 0 br-lan ::/0 :: !n -1 2 0 lo #ip -6 route list default from 240e:3a1:8439:ef00::/64 via fe80::c877:6ff:fefb:bc95 dev br-lan proto static metric 512 pref medium default from fd56:b34c:1a0e::/64 via fe80::c877:6ff:fefb:bc95 dev br-lan proto static metric 512 pref medium unreachable 240e:3a1:8438:e190::/64 dev lo proto static metric 2147483647 pref medium 240e:3a1:8439:ef00::/64 dev br-lan proto static metric 256 pref medium unreachable 240e:3a1:8439:ef00::/64 dev lo proto static metric 2147483647 pref medium fd56:b34c:1a0e::/64 dev br-lan proto static metric 256 pref medium unreachable fd56:b34c:1a0e::/64 dev lo proto static metric 2147483647 pref medium fe80::/64 dev br-lan proto kernel metric 256 pref medium #ip -6 rule show 0: from all lookup local 32765: from all fwmark 0x162 lookup 354 32766: from all lookup main #===================== 端口占用状态 =====================# tcp 0 0 :::7891 :::* LISTEN 29370/clash tcp 0 0 :::7890 :::* LISTEN 29370/clash tcp 0 0 :::7893 :::* LISTEN 29370/clash tcp 0 0 :::7892 :::* LISTEN 29370/clash tcp 0 0 :::7895 :::* LISTEN 29370/clash tcp 0 0 :::9090 :::* LISTEN 29370/clash udp 0 0 :::37435 :::* 29370/clash udp 0 0 :::43649 :::* 29370/clash udp 0 0 :::47745 :::* 29370/clash udp 0 0 :::58048 :::* 29370/clash udp 0 0 :::7874 :::* 29370/clash udp 0 0 :::7891 :::* 29370/clash udp 0 0 :::7892 :::* 29370/clash udp 0 0 :::7893 :::* 29370/clash udp 0 0 :::7895 :::* 29370/clash udp 0 0 :::55037 :::* 29370/clash udp 0 0 :::48968 :::* 29370/clash udp 0 0 :::49163 :::* 29370/clash udp 0 0 :::32834 :::* 29370/clash udp 0 0 :::41100 :::* 29370/clash udp 0 0 :::57521 :::* 29370/clash udp 0 0 :::41140 :::* 29370/clash udp 0 0 :::41145 :::* 29370/clash udp 0 0 :::37080 :::* 29370/clash udp 0 0 :::53501 :::* 29370/clash udp 0 0 :::51504 :::* 29370/clash udp 0 0 :::35172 :::* 29370/clash udp 0 0 :::55770 :::* 29370/clash udp 0 0 :::54816 :::* 29370/clash udp 0 0 :::38436 :::* 29370/clash udp 0 0 :::58920 :::* 29370/clash #===================== 测试本机DNS查询(www.baidu.com) =====================# Server: 127.0.0.1 Address: 127.0.0.1:53 Non-authoritative answer: www.baidu.com canonical name = www.a.shifen.com Name: www.a.shifen.com Address: 180.101.50.242 Name: www.a.shifen.com Address: 180.101.50.188 Non-authoritative answer: www.baidu.com canonical name = www.a.shifen.com Name: www.a.shifen.com Address: 240e:e9:6002:15c:0:ff:b015:146f Name: www.a.shifen.com Address: 240e:e9:6002:15a:0:ff:b05c:1278 #===================== 测试内核DNS查询(www.instagram.com) =====================# Status: 0 TC: false RD: true RA: true AD: false CD: false Question: Name: www.instagram.com. Qtype: 1 Qclass: 1 Answer: TTL: 1 data: 31.13.87.34 name: www.instagram.com. type: 1 Status: 0 TC: false RD: true RA: true AD: false CD: false Question: Name: www.instagram.com. Qtype: 28 Qclass: 1 Answer: TTL: 1 data: 2a03:2880:f112:83:face:b00c:0:25de name: www.instagram.com. type: 28 Dnsmasq 当前默认 resolv 文件:/tmp/resolv.conf.d/resolv.conf.auto #===================== /tmp/resolv.conf.d/resolv.conf.auto =====================# # Interface IPV6 nameserver fd56:b34c:1a0e::1 # Interface lan nameserver 223.5.5.5 nameserver 119.29.29.29 #===================== 测试本机网络连接(www.baidu.com) =====================# HTTP/1.1 200 OK Bdpagetype: 1 Bdqid: 0xac03f7bd002353e5 Connection: keep-alive Content-Length: 409479 Content-Type: text/html; charset=utf-8 Date: Fri, 28 Jun 2024 12:13:07 GMT Server: BWS/1.1 Set-Cookie: BIDUPSID=A3A2BB9C132015C319103F629EB041E8; expires=Thu, 31-Dec-37 23:55:55 GMT; max-age=2147483647; path=/; domain=.baidu.com Set-Cookie: PSTM=1719576787; expires=Thu, 31-Dec-37 23:55:55 GMT; max-age=2147483647; path=/; domain=.baidu.com Set-Cookie: BDSVRTM=0; path=/ Set-Cookie: BD_HOME=1; path=/ Set-Cookie: BAIDUID=A3A2BB9C132015C319103F629EB041E8:FG=1; Path=/; Domain=baidu.com; Max-Age=31536000 Set-Cookie: BAIDUID_BFESS=A3A2BB9C132015C319103F629EB041E8:FG=1; Path=/; Domain=baidu.com; Max-Age=31536000; Secure; SameSite=None Traceid: 1719576787061319988212395022990576931813 Vary: Accept-Encoding X-Ua-Compatible: IE=Edge,chrome=1 X-Xss-Protection: 1;mode=block #===================== 测试本机网络下载(raw.githubusercontent.com) =====================# HTTP/2 404 content-security-policy: default-src 'none'; style-src 'unsafe-inline'; sandbox strict-transport-security: max-age=31536000 x-content-type-options: nosniff x-frame-options: deny x-xss-protection: 1; mode=block content-type: text/plain; charset=utf-8 x-github-request-id: F7EC:BBEB7:2B4973:32FFE0:667EA8B2 accept-ranges: bytes date: Fri, 28 Jun 2024 12:13:08 GMT via: 1.1 varnish x-served-by: cache-nrt-rjtf7700077-NRT x-cache: HIT x-cache-hits: 1 x-timer: S1719576788.267217,VS0,VE1 vary: Authorization,Accept-Encoding,Origin access-control-allow-origin: * cross-origin-resource-policy: cross-origin x-fastly-request-id: 4587b7a64bd18e3caeafc0fc570d131c715c11a4 expires: Fri, 28 Jun 2024 12:18:08 GMT source-age: 33 content-length: 14 #===================== 最近运行日志(自动切换为Debug模式) =====================# 12:13:12 DBG [Matcher] find process failed error=process not found addr=173.24.72.216 12:13:12 INF [TCP] connected lAddr=10.0.0.4:40732 rAddr=45.9.62.29:23333 mode=rule rule=Match() proxy=DIRECT 12:13:12 WRN [TCP] dial failed error=dial tcp4 74.48.66.71:5675: connect: connection refused proxy=DIRECT lAddr=10.0.0.4:58778 rAddr=74.48.66.71:5675 rule=Match rulePayload= 12:13:12 DBG [UDP] accept session lAddr=10.0.0.15:52088 rAddr=5.5.5.5:55555 inbound=TProxy 12:13:12 DBG [Matcher] find process failed error=process not found addr=5.5.5.5 12:13:12 INF [UDP] connected lAddr=10.0.0.15:52088 rAddr=5.5.5.5:55555 mode=rule rule=Match() proxy=DIRECT 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:46799 rAddr=129.154.201.4:8999 inbound=Redir 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:49485 rAddr=20.212.33.239:51413 inbound=Redir 12:13:13 DBG [Matcher] find process failed error=process not found addr=129.154.201.4 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:33373 rAddr=149.88.26.162:42076 inbound=Redir 12:13:13 DBG [Matcher] find process failed error=process not found addr=20.212.33.239 12:13:13 DBG [Matcher] find process failed error=process not found addr=149.88.26.162 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:58967 rAddr=131.186.43.131:32836 inbound=Redir 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:34734 rAddr=114.24.98.134:9833 inbound=Redir 12:13:13 DBG [Matcher] find process failed error=process not found addr=114.24.98.134 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:56270 rAddr=172.247.123.11:13780 inbound=Redir 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:40040 rAddr=178.238.229.54:40888 inbound=Redir 12:13:13 DBG [Matcher] find process failed error=process not found addr=172.247.123.11 12:13:13 DBG [TCP] accept connection lAddr=10.0.0.4:58443 rAddr=107.172.79.141:60012 inbound=Redir 12:13:13 DBG [Matcher] find process failed error=process not found addr=178.238.229.54 12:13:13 DBG [Matcher] find process failed error=process not found addr=107.172.79.141 12:13:13 DBG [Matcher] find process failed error=process not found addr=131.186.43.131 12:13:13 INF [TCP] connected lAddr=10.0.0.4:49485 rAddr=20.212.33.239:51413 mode=rule rule=Match() proxy=DIRECT 12:13:14 WRN [TCP] dial failed error=dial tcp4 168.70.70.144:38708: i/o timeout proxy=DIRECT lAddr=10.0.0.4:51460 rAddr=168.70.70.144:38708 rule=Match rulePayload= 12:13:14 WRN [TCP] dial failed error=dial tcp4 147.122.43.76:56459: i/o timeout proxy=DIRECT lAddr=10.0.0.4:60130 rAddr=147.122.43.76:56459 rule=Match rulePayload= 12:13:14 DBG [TCP] accept connection lAddr=10.0.0.4:40114 rAddr=107.182.30.5:33913 inbound=Redir 12:13:14 DBG [TCP] accept connection lAddr=10.0.0.4:59886 rAddr=132.226.231.217:51413 inbound=Redir 12:13:14 DBG [Matcher] find process failed error=process not found addr=132.226.231.217 12:13:14 DBG [Matcher] find process failed error=process not found addr=107.182.30.5 12:13:14 DBG [TCP] accept connection lAddr=10.0.0.4:47166 rAddr=192.9.241.40:58799 inbound=Redir 12:13:14 DBG [Matcher] find process failed error=process not found addr=192.9.241.40 12:13:14 WRN [TCP] dial failed error=dial tcp4 192.9.241.40:58799: connect: connection refused proxy=DIRECT lAddr=10.0.0.4:47166 rAddr=192.9.241.40:58799 rule=Match rulePayload= 12:13:14 INF [TCP] connected lAddr=10.0.0.4:58443 rAddr=107.172.79.141:60012 mode=rule rule=Match() proxy=DIRECT 12:13:15 WRN [TCP] dial failed error=dial tcp4 142.171.65.158:37420: i/o timeout proxy=DIRECT lAddr=10.0.0.4:42686 rAddr=142.171.65.158:37420 rule=Match rulePayload= 12:13:15 DBG [TCP] accept connection lAddr=10.0.0.4:38827 rAddr=221.124.194.171:35583 inbound=Redir 12:13:15 DBG [TCP] accept connection lAddr=10.0.0.4:57471 rAddr=42.98.167.113:49293 inbound=Redir 12:13:15 DBG [Matcher] find process failed error=process not found addr=221.124.194.171 12:13:15 DBG [TCP] accept connection lAddr=10.0.0.4:53375 rAddr=20.212.33.239:51413 inbound=Redir 12:13:15 DBG [Matcher] find process failed error=process not found addr=42.98.167.113 12:13:15 DBG [Matcher] find process failed error=process not found addr=20.212.33.239 12:13:15 DBG [TCP] accept connection lAddr=10.0.0.4:44427 rAddr=221.124.194.171:35583 inbound=Redir 12:13:15 DBG [TCP] accept connection lAddr=10.0.0.4:33057 rAddr=142.171.46.31:38621 inbound=Redir 12:13:15 DBG [Matcher] find process failed error=process not found addr=221.124.194.171 12:13:15 DBG [TCP] accept connection lAddr=10.0.0.4:60642 rAddr=213.35.127.250:51413 inbound=Redir 12:13:15 DBG [Matcher] find process failed error=process not found addr=142.171.46.31 12:13:15 DBG [Matcher] find process failed error=process not found addr=213.35.127.250 12:13:15 WRN [TCP] dial failed error=dial tcp4 42.98.167.113:49293: connect: connection refused proxy=DIRECT lAddr=10.0.0.4:57471 rAddr=42.98.167.113:49293 rule=Match rulePayload= 12:13:15 INF [TCP] connected lAddr=10.0.0.4:53375 rAddr=20.212.33.239:51413 mode=rule rule=Match() proxy=DIRECT 12:13:15 INF [TCP] connected lAddr=10.0.0.4:33057 rAddr=142.171.46.31:38621 mode=rule rule=Match() proxy=DIRECT 12:13:15 INF [TCP] connected lAddr=10.0.0.4:60642 rAddr=213.35.127.250:51413 mode=rule rule=Match() proxy=DIRECT 12:13:15 INF [TCP] connected lAddr=10.0.0.4:38827 rAddr=221.124.194.171:35583 mode=rule rule=Match() proxy=DIRECT 12:13:15 INF [TCP] connected lAddr=10.0.0.4:44427 rAddr=221.124.194.171:35583 mode=rule rule=Match() proxy=DIRECT 12:13:15 DBG [UDP] accept session lAddr=10.0.0.15:55489 rAddr=5.5.5.5:55555 inbound=TProxy 12:13:15 DBG [Matcher] find process failed error=process not found addr=5.5.5.5 12:13:15 INF [UDP] connected lAddr=10.0.0.15:55489 rAddr=5.5.5.5:55555 mode=rule rule=Match() proxy=DIRECT 12:13:16 DBG [TCP] accept connection lAddr=10.0.0.4:51898 rAddr=107.172.79.141:60012 inbound=Redir 12:13:16 DBG [TCP] accept connection lAddr=10.0.0.4:48340 rAddr=142.171.46.31:38621 inbound=Redir 12:13:16 DBG [TCP] accept connection lAddr=10.0.0.4:38723 rAddr=125.199.241.115:63219 inbound=Redir 12:13:16 DBG [TCP] accept connection lAddr=10.0.0.4:54764 rAddr=118.161.143.211:58919 inbound=Redir 12:13:16 DBG [TCP] accept connection lAddr=10.0.0.4:42681 rAddr=143.198.63.21:45705 inbound=Redir 12:13:16 DBG [Matcher] find process failed error=process not found addr=125.199.241.115 12:13:16 DBG [Matcher] find process failed error=process not found addr=107.172.79.141 12:13:16 DBG [Matcher] find process failed error=process not found addr=142.171.46.31 12:13:16 DBG [Matcher] find process failed error=process not found addr=143.198.63.21 12:13:16 DBG [Matcher] find process failed error=process not found addr=118.161.143.211 12:13:16 INF [TCP] connected lAddr=10.0.0.4:54764 rAddr=118.161.143.211:58919 mode=rule rule=Match() proxy=DIRECT 12:13:16 INF [TCP] connected lAddr=10.0.0.4:51898 rAddr=107.172.79.141:60012 mode=rule rule=Match() proxy=DIRECT 12:13:16 WRN [TCP] dial failed error=dial tcp4 143.198.63.21:45705: connect: connection refused proxy=DIRECT lAddr=10.0.0.4:42681 rAddr=143.198.63.21:45705 rule=Match rulePayload= 12:13:16 INF [TCP] connected lAddr=10.0.0.4:48340 rAddr=142.171.46.31:38621 mode=rule rule=Match() proxy=DIRECT 12:13:17 WRN [TCP] dial failed error=dial tcp4 24.4.59.106:36922: i/o timeout proxy=DIRECT lAddr=10.0.0.4:56128 rAddr=24.4.59.106:36922 rule=Match rulePayload= 12:13:17 WRN [TCP] dial failed error=dial tcp4 13.231.128.231:27985: i/o timeout proxy=DIRECT lAddr=10.0.0.4:34827 rAddr=13.231.128.231:27985 rule=Match rulePayload= 12:13:17 WRN [TCP] dial failed error=dial tcp4 173.24.72.216:51413: i/o timeout proxy=DIRECT lAddr=10.0.0.4:56247 rAddr=173.24.72.216:51413 rule=Match rulePayload= 12:13:17 DBG [TCP] accept connection lAddr=10.0.0.4:52679 rAddr=123.194.32.40:51413 inbound=Redir 12:13:17 DBG [TCP] accept connection lAddr=10.0.0.4:51357 rAddr=171.239.139.2:1010 inbound=Redir 12:13:17 DBG [TCP] accept connection lAddr=10.0.0.4:45273 rAddr=51.89.151.110:9001 inbound=Redir 12:13:17 DBG [TCP] accept connection lAddr=10.0.0.4:55891 rAddr=103.120.10.160:63754 inbound=Redir 12:13:17 DBG [Matcher] find process failed error=process not found addr=123.194.32.40 12:13:17 DBG [Matcher] find process failed error=process not found addr=171.239.139.2 12:13:17 DBG [Matcher] find process failed error=process not found addr=103.120.10.160 12:13:17 DBG [Matcher] find process failed error=process not found addr=51.89.151.110 12:13:18 WRN [TCP] dial failed error=dial tcp4 129.154.201.4:8999: i/o timeout proxy=DIRECT lAddr=10.0.0.4:46799 rAddr=129.154.201.4:8999 rule=Match rulePayload= 12:13:18 WRN [TCP] dial failed error=dial tcp4 149.88.26.162:42076: i/o timeout proxy=DIRECT lAddr=10.0.0.4:33373 rAddr=149.88.26.162:42076 rule=Match rulePayload= 12:13:18 WRN [TCP] dial failed error=dial tcp4 114.24.98.134:9833: i/o timeout proxy=DIRECT lAddr=10.0.0.4:34734 rAddr=114.24.98.134:9833 rule=Match rulePayload= 12:13:18 DBG [TCP] accept connection lAddr=10.0.0.4:36913 rAddr=94.75.73.210:48500 inbound=Redir 12:13:18 DBG [TCP] accept connection lAddr=10.0.0.4:44940 rAddr=123.194.32.40:51413 inbound=Redir 12:13:18 DBG [TCP] accept connection lAddr=10.0.0.4:53933 rAddr=213.35.127.250:51413 inbound=Redir 12:13:18 DBG [TCP] accept connection lAddr=10.0.0.4:47206 rAddr=152.67.210.31:41122 inbound=Redir 12:13:18 DBG [Matcher] find process failed error=process not found addr=213.35.127.250 12:13:18 DBG [Matcher] find process failed error=process not found addr=123.194.32.40 12:13:18 DBG [Matcher] find process failed error=process not found addr=94.75.73.210 12:13:18 WRN [TCP] dial failed error=dial tcp4 172.247.123.11:13780: i/o timeout proxy=DIRECT lAddr=10.0.0.4:56270 rAddr=172.247.123.11:13780 rule=Match rulePayload= 12:13:18 DBG [Matcher] find process failed error=process not found addr=152.67.210.31 12:13:18 WRN [TCP] dial failed error=dial tcp4 178.238.229.54:40888: i/o timeout proxy=DIRECT lAddr=10.0.0.4:40040 rAddr=178.238.229.54:40888 rule=Match rulePayload= 12:13:18 WRN [TCP] dial failed error=dial tcp4 131.186.43.131:32836: i/o timeout proxy=DIRECT lAddr=10.0.0.4:58967 rAddr=131.186.43.131:32836 rule=Match rulePayload= 12:13:18 DBG [TCP] accept connection lAddr=10.0.0.15:3495 rAddr=a.nel.cloudflare.com:443 inbound=Redir 12:13:18 DBG [Matcher] find process failed error=process not found addr=a.nel.cloudflare.com 12:13:18 DBG [Matcher] resolve success host=a.nel.cloudflare.com ip=35.190.80.1 12:13:18 DBG [DNS] dns response source=10.0.0.10:53 qType=A name=33668.laomao-53875.xyz. answer=["120.233.27.193"] 12:13:18 WRN [TCP] dial failed error=dial tcp4 152.67.210.31:41122: connect: connection refused proxy=DIRECT lAddr=10.0.0.4:47206 rAddr=152.67.210.31:41122 rule=Match rulePayload= 12:13:18 INF [TCP] connected lAddr=10.0.0.15:3495 rAddr=a.nel.cloudflare.com:443 mode=rule rule=DomainSuffix(cloudflare.com) proxy=Proxies[台湾10] #===================== 最近运行日志获取完成(自动切换为silent模式) =====================# #===================== 活动连接信息 =====================# 1. SourceIP:【10.0.0.15】 - Host:【qqwry.api.skk.moe】 - DestinationIP:【172.67.148.227】 - Network:【tcp】 - RulePayload:【443】 - Lastchain:【DIRECT】 2. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 3. SourceIP:【10.0.0.15】 - Host:【activity.windows.com】 - DestinationIP:【20.69.137.228】 - Network:【tcp】 - RulePayload:【windows.com】 - Lastchain:【台湾10】 4. SourceIP:【10.0.0.15】 - Host:【raw.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 5. SourceIP:【10.0.0.15】 - Host:【alive.github.com】 - DestinationIP:【140.82.113.26】 - Network:【tcp】 - RulePayload:【github】 - Lastchain:【台湾10】 6. SourceIP:【10.0.0.16】 - Host:【mtalk.google.com】 - DestinationIP:【74.125.23.188】 - Network:【tcp】 - RulePayload:【mtalk.google.com】 - Lastchain:【台湾10】 7. SourceIP:【10.0.0.15】 - Host:【github.githubassets.com】 - DestinationIP:【185.199.110.154】 - Network:【tcp】 - RulePayload:【github】 - Lastchain:【台湾10】 8. SourceIP:【10.0.0.15】 - Host:【avatars2.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 9. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 10. SourceIP:【10.0.0.4】 - Host:【Empty】 - DestinationIP:【221.124.194.171】 - Network:【tcp】 - RulePayload:【】 - Lastchain:【DIRECT】 11. SourceIP:【10.0.0.15】 - Host:【a.nel.cloudflare.com】 - DestinationIP:【35.190.80.1】 - Network:【tcp】 - RulePayload:【cloudflare.com】 - Lastchain:【台湾10】 12. SourceIP:【10.0.0.2】 - Host:【op.supes.top】 - DestinationIP:【172.67.213.212】 - Network:【tcp】 - RulePayload:【supes.top】 - Lastchain:【DIRECT】 13. SourceIP:【10.0.0.4】 - Host:【Empty】 - DestinationIP:【211.51.7.97】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 14. SourceIP:【10.0.0.15】 - Host:【client.wns.windows.com】 - DestinationIP:【20.198.162.76】 - Network:【tcp】 - RulePayload:【windows.com】 - Lastchain:【台湾10】 15. SourceIP:【10.0.0.15】 - Host:【ext2-tyo3.steamserver.net】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【steamserver.net】 - Lastchain:【DIRECT】 16. SourceIP:【10.0.0.15】 - Host:【avatars1.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 17. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 18. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 19. SourceIP:【10.0.0.15】 - Host:【content-autofill.googleapis.com】 - DestinationIP:【172.217.160.74】 - Network:【tcp】 - RulePayload:【google】 - Lastchain:【台湾10】 20. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 21. SourceIP:【10.0.0.15】 - Host:【avatars0.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 22. SourceIP:【10.0.0.15】 - Host:【avatars.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 23. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 24. SourceIP:【10.0.0.15】 - Host:【avatars3.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 25. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 26. SourceIP:【10.0.0.15】 - Host:【stun.syncthing.net】 - DestinationIP:【139.59.84.212】 - Network:【udp】 - RulePayload:【syncthing.net】 - Lastchain:【DIRECT】 27. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 28. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 29. SourceIP:【10.0.0.15】 - Host:【avatars.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 30. SourceIP:【10.0.0.15】 - Host:【api.ipify.org】 - DestinationIP:【172.67.74.152】 - Network:【tcp】 - RulePayload:【443】 - Lastchain:【DIRECT】 31. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 32. SourceIP:【10.0.0.15】 - Host:【private-user-images.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 33. SourceIP:【10.0.0.15】 - Host:【clients4.google.com】 - DestinationIP:【172.217.163.46】 - Network:【tcp】 - RulePayload:【google】 - Lastchain:【台湾10】 34. SourceIP:【10.0.0.15】 - Host:【www.youtube.com】 - DestinationIP:【199.59.148.229】 - Network:【tcp】 - RulePayload:【youtube.com】 - Lastchain:【新加坡03】 35. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 36. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 37. SourceIP:【10.0.0.4】 - Host:【Empty】 - DestinationIP:【221.124.194.171】 - Network:【tcp】 - RulePayload:【】 - Lastchain:【DIRECT】 38. SourceIP:【10.0.0.15】 - Host:【api-ipv4.ip.sb】 - DestinationIP:【104.26.12.31】 - Network:【tcp】 - RulePayload:【ip.sb】 - Lastchain:【台湾10】 39. SourceIP:【10.0.0.15】 - Host:【mtalk.google.com】 - DestinationIP:【74.125.203.188】 - Network:【tcp】 - RulePayload:【mtalk.google.com】 - Lastchain:【台湾10】 40. SourceIP:【10.0.0.4】 - Host:【db.xtatcha.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【xtatcha.com】 - Lastchain:【DIRECT】 41. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 42. SourceIP:【10.0.0.16】 - Host:【cn.pool.ntp.org】 - DestinationIP:【162.159.200.1】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 43. SourceIP:【10.0.0.15】 - Host:【a.nel.cloudflare.com】 - DestinationIP:【35.190.80.1】 - Network:【tcp】 - RulePayload:【cloudflare.com】 - Lastchain:【台湾10】 44. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 45. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 46. SourceIP:【10.0.0.2】 - Host:【dl.openwrt.ai】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【openwrt.ai】 - Lastchain:【DIRECT】 47. SourceIP:【10.0.0.15】 - Host:【chromewebstore.google.com】 - DestinationIP:【142.251.42.238】 - Network:【tcp】 - RulePayload:【google】 - Lastchain:【台湾10】 48. SourceIP:【10.0.0.15】 - Host:【raw.githubusercontent.com】 - DestinationIP:【】 - Network:【tcp】 - RulePayload:【githubusercontent.com】 - Lastchain:【台湾10】 49. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 50. SourceIP:【10.0.0.15】 - Host:【Empty】 - DestinationIP:【5.5.5.5】 - Network:【udp】 - RulePayload:【】 - Lastchain:【DIRECT】 ### OpenClash Config _No response_ ### Expected Behavior 求助,请问该问题能否解决,或者能否通过设置缓解 ### Additional Context _No response_ 我是用LEDE 5.15 固件,open clash的DNS指向adguardhome,访问快手和京东 APP 会有证书校验无效提示,其他设置不变,改用阿里或者其他没有广告过滤的DNS提示就会消失。 还有用immortalwrt k54 编译的固件,相同设置,有广告过滤的dns也没有错误提示 另外用了 hagezi 过滤规则提示会增加,不用这个规则只会提示一次,用来会变成六七次 我是用LEDE 5.15 固件,open clash的DNS指向adguardhome,访问快手和京东 APP 会有证书校验无效提示,其他设置不变,改用阿里或者其他没有广告过滤的DNS提示就会消失。 还有用immortalwrt k54 编译的固件,相同设置,有广告过滤的dns也没有错误提示 另外用了 hagezi 过滤规则提示会增加,不用这个规则只会提示一次,用来会变成六七次 这个DNS问题可以解决吗? 我是从adguardhome指向openclash,openclash指向pihole。家里部分设备不经过op时不触发这个问题,确认了不是pihole的问题,倒是没试过不经过adh会不会触发,后面我会再试试,谢谢 我是用LEDE 5.15 固件,open clash的DNS指向adguardhome,访问快手和京东 APP 会有证书校验无效提示,其他设置不变,改用阿里或者其他没有广告过滤的DNS提示就会消失。 还有用immortalwrt k54 编译的固件,相同设置,有广告过滤的dns也没有错误提示 另外用了 hagezi 过滤规则提示会增加,不用这个规则只会提示一次,用来会变成六七次 这个DNS问题可以解决吗? 我是从adguardhome指向openclash,openclash指向pihole。家里部分设备不经过op时不触发这个问题,确认了不是pihole的问题,倒是没试过不经过adh会不会触发,后面我会再试试,谢谢 又试了一下,发现把 adguardhome - 设置 - 拦截模式 改成 REFUSED 证书的提示就消失了。希望能帮到你 我遇到类似问题,只不过我用的passwall,试了openclash也是这个问题。 我遇到类似问题,只不过我用的passwall,试了openclash也是这个问题。 我把openclash的ipv6流量关闭,只解析ipv6的DNS后就没有该问题了。每次出问题ping的时候都是DNS指向了openclash的fakeip池子,所以做了该尝试。
gharchive/issue
2024-06-28T12:31:28
2025-04-01T06:46:09.954079
{ "authors": [ "Tatchaxzw", "lztxi", "shaw21003" ], "repo": "vernesong/OpenClash", "url": "https://github.com/vernesong/OpenClash/issues/3940", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
92630897
Minor improvements + tests simplifications this is a PR that improves minor stuff in codegen itself and simplify tests by removing boiler plate and making them more type safe. @purplefox @cescoffier @pmlopes please review Looks good to me.
gharchive/pull-request
2015-07-02T12:58:32
2025-04-01T06:46:09.972988
{ "authors": [ "cescoffier", "vietj" ], "repo": "vert-x3/vertx-codegen", "url": "https://github.com/vert-x3/vertx-codegen/pull/36", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
754557304
A single checkpoint should not supersede a succeedingThenComplete Describe the feature Given code like this: @Test void testProc(VertxTestContext testContext) { var checkpoint = testContext.checkpoint(); returnsFuture() .onSuccess(v -> checkpoint.flag()) .compose(v -> anotherReturnsFuture()) .onComplete(testContext.succeedingThenComplete()); } the user's intent is almost always to flag the checkpoint but to keep running until testContext.succeedingThenComplete() is called with a succeeded future. Currently, creating a checkpoint like in the above statement will actually end the test as soon as the test thread notices checkpoint.flag() (i.e. the injected testContext.awaitCompletion will be satisfied). This is very clunky. Use cases A common use case is to run some side-by-side check independent the test code and use checkpoints to flag that your assertions were actually run. For example you might mock a class and checkpoint.flag() inside a mocked method that runs assertions on its arguments. Contribution I can implement this feature. When there are checkpoints TestContext completes when they have all completed, so onComplete shall also flag a checkpoint. Also how would you spot that the test context succeedingThenComplete method could eventually be called? succeedingThenComplete is really a shortcut for completing immediately when one asynchronous operation completes. Yes, but what happens is the test completes before succeedingThenComplete 's returned handler is called, because of something with checkpoint.flag().
gharchive/issue
2020-12-01T16:48:26
2025-04-01T06:46:09.983773
{ "authors": [ "doctorpangloss", "jponge" ], "repo": "vert-x3/vertx-junit5", "url": "https://github.com/vert-x3/vertx-junit5/issues/90", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1819561558
Remove redundant type variables in the guide. I am reading https://verus-lang.github.io/verus/guide/spec_lib.html and find a TODO is left: TODO: lemma_len_intersect::<A>(...) should not need the ::<A>. I removed those type variables and verus can also work. It seems that verus has supported type inference. Hi! Thank you for fix!
gharchive/pull-request
2023-07-25T05:07:05
2025-04-01T06:46:10.006790
{ "authors": [ "HaoYang670", "utaal" ], "repo": "verus-lang/verus", "url": "https://github.com/verus-lang/verus/pull/706", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1720172161
Integrate via new Tana Inbox API Avoid the clipboard route by pushing clippings directly to Tana via the new Inbox API. The problem with this route is that the Tana API does not provide Tana-paste capability. So tagging, fields, etc. are much more complex to create and since tags need the underlying node ID of the supertag, and there's currently no API for discovering those, the user experience for configuring the tag would be cumbersome. Hence, putting this on the backlog for the future when Tana offers either an API for tag discovery or Tana-paste support in the Inbox API. Current beta build includes Tana Inbox support with a configuration UI to set up all the nodeid's etc.
gharchive/issue
2023-05-22T18:02:49
2025-04-01T06:46:10.008385
{ "authors": [ "verveguy" ], "repo": "verveguy/clip2tana", "url": "https://github.com/verveguy/clip2tana/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1401818429
Made a Flink demo tutorial As title. I made a video Flink demo tutorial by using the Flink OceanBase CDC connector as an example. Reference doc: https://github.com/ververica/flink-cdc-connectors/blob/master/docs/content/connectors/oceanbase-cdc.md https://github.com/ververica/flink-cdc-connectors/blob/master/docs/content/快速上手/oceanbase-tutorial-zh.md Video: Bilibili YouTube @Amber1990Zhang So cool!! Would you like to open a PR to add the video links to the document?
gharchive/issue
2022-10-08T04:32:27
2025-04-01T06:46:10.010937
{ "authors": [ "Amber1990Zhang", "leonardBang" ], "repo": "ververica/flink-cdc-connectors", "url": "https://github.com/ververica/flink-cdc-connectors/issues/1601", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
981997653
[mysql] MysqlParallelSource Stability Improvement [x] 1. Skip binlog reading, if low == high [x] 2. If concurrency=1, there is no need to wait with checkpoint. (done) #365 [x] 3. Improve the log and clean up debezium logs [x] 4. Investigate rds master/standby switching scenarios can works or not [ ] 5. Bind the DebeziumReader life circle with split [ ] 6. Investigate MySqlParallelSource could use single serverId or not Closing this issue because it was created before version 2.3.0 (2022-11-10). Please try the latest version of Flink CDC to see if the issue has been resolved. If the issue is still valid, kindly report it on Apache Jira under project Flink with component tag Flink CDC. Thank you!
gharchive/issue
2021-08-29T05:16:25
2025-04-01T06:46:10.013852
{ "authors": [ "PatrickRen", "leonardBang" ], "repo": "ververica/flink-cdc-connectors", "url": "https://github.com/ververica/flink-cdc-connectors/issues/373", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1007660010
java.lang.NumberFormatException: Value out of range. Value:"139" Radix:10 Environment : Flink version : 1.13.1 Flink CDC version: 2.0.2 Database and version: MySQL 5.7.33 To Reproduce Steps to reproduce the behavior: Thes test data : The test code : The error : java.lang.NumberFormatException: Value out of range. Value:"139" Radix:10 at java.lang.Byte.parseByte(Byte.java:151) at java.lang.Byte.parseByte(Byte.java:175) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.lambda$createNotNullConverter$739d85a5$2(RowDataDebeziumDeserializeSchema.java:167) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.lambda$wrapIntoNullableConverter$6d4c36ad$1(RowDataDebeziumDeserializeSchema.java:394) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.convertField(RowDataDebeziumDeserializeSchema.java:384) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.lambda$createRowConverter$47aa93ac$1(RowDataDebeziumDeserializeSchema.java:371) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.lambda$wrapIntoNullableConverter$6d4c36ad$1(RowDataDebeziumDeserializeSchema.java:394) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.extractAfterRow(RowDataDebeziumDeserializeSchema.java:127) at com.ververica.cdc.debezium.table.RowDataDebeziumDeserializeSchema.deserialize(RowDataDebeziumDeserializeSchema.java:102) at com.ververica.cdc.connectors.mysql.source.reader.MySqlRecordEmitter.emitRecord(MySqlRecordEmitter.java:92) at com.ververica.cdc.connectors.mysql.source.reader.MySqlRecordEmitter.emitRecord(MySqlRecordEmitter.java:53) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:128) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:294) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:423) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:681) at org.apache.flink.streaming.runtime.tasks.StreamTask.executeInvoke(StreamTask.java:636) at org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:647) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:620) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:779) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at java.lang.Thread.run(Thread.java:748) Closing this issue because it was created before version 2.3.0 (2022-11-10). Please try the latest version of Flink CDC to see if the issue has been resolved. If the issue is still valid, kindly report it on Apache Jira under project Flink with component tag Flink CDC. Thank you!
gharchive/issue
2021-09-27T03:52:32
2025-04-01T06:46:10.023136
{ "authors": [ "PatrickRen", "dczeeee" ], "repo": "ververica/flink-cdc-connectors", "url": "https://github.com/ververica/flink-cdc-connectors/issues/466", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1076386436
java.lang.RuntimeException: One or more fetchers have encountered exception Describe the bug A clear and concise description of what the bug is. Environment : Flink version : 1.13.1 Flink CDC version: 2.0.2 Database and version: 8.0.13 To Reproduce Steps to reproduce the behavior: Thes test data : The test code :'scan.startup.mode' = 'initial' The error : 2021-12-09 20:40:16 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199) at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:294) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:423) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:681) at org.apache.flink.streaming.runtime.tasks.StreamTask.executeInvoke(StreamTask.java:636) at org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:647) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:620) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:779) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.RuntimeException: SplitFetcher thread 539 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:146) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:101) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:42) at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1181) at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ... 1 more Caused by: io.debezium.DebeziumException: A slave with the same server_uuid/server_id as this slave has connected to the master; the first event '' at 4, the last event read from '../log/mysql-bin.498949' at 347093718, the last byte read from '../log/mysql-bin.498949' at 347093718. Error code: 1236; SQLSTATE: HY000. at io.debezium.connector.mysql.MySqlStreamingChangeEventSource.wrap(MySqlStreamingChangeEventSource.java:1142) ... 5 more Caused by: com.github.shyiko.mysql.binlog.network.ServerException: A slave with the same server_uuid/server_id as this slave has connected to the master; the first event '' at 4, the last event read from '../log/mysql-bin.498949' at 347093718, the last byte read from '../log/mysql-bin.498949' at 347093718. at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:937) ... 3 more Additional Description If applicable, add screenshots to help explain your problem. 我也遇到了这个问题,Flink CDC2.1, 已经设置了server-id,但还是会出现这个问题 您好,我也遇到了类似的问题,cdc2.1.1,flink1.13.5.请问您解决了吗? 同样遇到这个问题,目前有解法了吗? cdc2.1.1,flink1.13.5 我也报了这个错,我是使用代码连接mysql binlog,通过配置参数多个库(全表),在代码中循环配置从而使得每个库都是一个独立的source,但host都是同一个,猜测是 serverid失效。 经测试发现只要一个host一个source就正常,多个host还没测试。 同样遇到这个问题 Flink 1.13.0 Cdc 2.1.1 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199) ~[flink-table-blink_2.11-1.13.0.jar:1.13.0] at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154) ~[flink-table-blink_2.11-1.13.0.jar:1.13.0] at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116) ~[flink-table-blink_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:294) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:419) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:661) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:623) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:776) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at org.apache.flink.runtime.taskmanager.Task.run(Task.java:563) ~[flink-dist_2.11-1.13.0.jar:1.13.0] at java.lang.Thread.run(Thread.java:745) ~[?:1.8.0_121] Caused by: java.lang.RuntimeException: SplitFetcher thread 12 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:146) ~[flink-table-blink_2.11-1.13.0.jar:1.13.0] at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:101) ~[flink-table-blink_2.11-1.13.0.jar:1.13.0] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[?:1.8.0_121] at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_121] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) ~[?:1.8.0_121] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ~[?:1.8.0_121] ... 1 more Caused by: com.ververica.cdc.connectors.shaded.org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:42) ~[flink-sql-connector-mongodb-cdc-2.1.1.jar:2.1.1] at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1185) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] ... 1 more Caused by: io.debezium.DebeziumException at io.debezium.connector.mysql.MySqlStreamingChangeEventSource.wrap(MySqlStreamingChangeEventSource.java:1146) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1185) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] ... 1 more Caused by: java.io.EOFException at com.github.shyiko.mysql.binlog.io.ByteArrayInputStream.read(ByteArrayInputStream.java:209) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.io.ByteArrayInputStream.readInteger(ByteArrayInputStream.java:51) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.event.deserialization.EventHeaderV4Deserializer.deserialize(EventHeaderV4Deserializer.java:35) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.event.deserialization.EventHeaderV4Deserializer.deserialize(EventHeaderV4Deserializer.java:27) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.event.deserialization.EventDeserializer.nextEvent(EventDeserializer.java:221) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$1.nextEvent(MySqlStreamingChangeEventSource.java:233) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:945) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ~[flink-sql-connector-mysql-cdc-2.1.1.jar:2.1.1] ... 1 more 同样遇到了这个问题, flink 1.13.6 CDC 2.2.1。 同样遇到了这个问题, flink 1.13.6 CDC 2.2.1。在阿里云上通过flink-cdc 将PoladrDB数据库的数据同步到holo上,启动多个任务 , 每个任务同步一张表,多个表的数据源是相同的。 原来只有一个任务在跑的时候,没发生报错的情况 环境 flink 1.13.3,flink cdc 2.1.1,scala 2.12,mysql 8.0.18 日志信息: 2022-07-25 14:37:51 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:223) at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:294) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:423) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:684) at org.apache.flink.streaming.runtime.tasks.StreamTask.executeInvoke(StreamTask.java:639) at org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:650) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:623) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:779) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at java.lang.Thread.run(Thread.java:750) Caused by: java.lang.RuntimeException: SplitFetcher thread 1 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:148) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:103) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: com.ververica.cdc.connectors.shaded.org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:42) at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1185) at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ... 1 more Caused by: io.debezium.DebeziumException: A slave with the same server_uuid/server_id as this slave has connected to the master; the first event 'mysql-bin.000047' at 64977309, the last event read from './mysql-bin.000047' at 65068948, the last byte read from './mysql-bin.000047' at 65068948. Error code: 1236; SQLSTATE: HY000. at io.debezium.connector.mysql.MySqlStreamingChangeEventSource.wrap(MySqlStreamingChangeEventSource.java:1146) ... 5 more Caused by: com.github.shyiko.mysql.binlog.network.ServerException: A slave with the same server_uuid/server_id as this slave has connected to the master; the first event 'mysql-bin.000047' at 64977309, the last event read from './mysql-bin.000047' at 65068948, the last byte read from './mysql-bin.000047' at 65068948. at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:937) ... 3 more 同样遇到此问题: flink : 13.5 cdc : 2.2.1 mysql : 5.7 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:223) at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:305) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:423) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:684) at org.apache.flink.streaming.runtime.tasks.StreamTask.executeInvoke(StreamTask.java:639) at org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:650) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:623) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:779) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.RuntimeException: SplitFetcher thread 10 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:148) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:103) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: io.debezium.DebeziumException: Unexpected error while connecting to MySQL and looking at gtid_purged variable: at io.debezium.connector.mysql.MySqlConnection.purgedGtidSet(MySqlConnection.java:290) at com.ververica.cdc.connectors.mysql.debezium.task.context.StatefulTaskContext.checkGtidSet(StatefulTaskContext.java:222) at com.ververica.cdc.connectors.mysql.debezium.task.context.StatefulTaskContext.isBinlogAvailable(StatefulTaskContext.java:190) at com.ververica.cdc.connectors.mysql.debezium.task.context.StatefulTaskContext.loadStartingOffsetState(StatefulTaskContext.java:177) at com.ververica.cdc.connectors.mysql.debezium.task.context.StatefulTaskContext.configure(StatefulTaskContext.java:113) at com.ververica.cdc.connectors.mysql.debezium.reader.BinlogSplitReader.submitSplit(BinlogSplitReader.java:93) at com.ververica.cdc.connectors.mysql.debezium.reader.BinlogSplitReader.submitSplit(BinlogSplitReader.java:65) at com.ververica.cdc.connectors.mysql.source.reader.MySqlSplitReader.checkSplitOrStartNext(MySqlSplitReader.java:163) at com.ververica.cdc.connectors.mysql.source.reader.MySqlSplitReader.fetch(MySqlSplitReader.java:73) at org.apache.flink.connector.base.source.reader.fetcher.FetchTask.run(FetchTask.java:56) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:140) ... 6 more try to set server_id value, 同遇到类似问题,有解决方案了吗 flink cdc mysql 2.2.1 mysql 5.7 scala2.11 flink 1.13.1 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199) ~[flink-connector-base-1.13.1.jar:1.13.1] at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154) ~[flink-connector-base-1.13.1.jar:1.13.1] at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116) ~[flink-connector-base-1.13.1.jar:1.13.1] at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:294) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:423) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:681) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.tasks.StreamTask.executeInvoke(StreamTask.java:636) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:647) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:620) ~[flink-streaming-java_2.11-1.13.1.jar:1.13.1] at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:779) ~[flink-runtime_2.11-1.13.1.jar:1.13.1] at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) ~[flink-runtime_2.11-1.13.1.jar:1.13.1] at java.lang.Thread.run(Thread.java:748) ~[na:1.8.0_211] 求解 【Caused by: io.debezium.DebeziumException: A slave with the same server_uuid/server_id as this slave has connected to the master; the first event 'mysql-bin.000047' at 64977309】报错指出了same server_uuid。 保证整个所有的serverid 都不要重复,有种情况要注意,source的并行度为2,假设A任务指定了 serverid 为5601, 其他任务的serverid 就不能制定为5602了,因为A任务的并行度为2 serverid 分配成 5601 5602. flink 1.13.16 flink-cdc 2.2.1 mysql:5.7.25 mysql-driver:5.1.49 增量同步阶段报错 请问有人遇到过吗 16:17:56,961 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Source: MySQL Source Demo -> Sink: Print to Std. Out (1/1) (ba10f3a939779c1435626570e396e950) switched from RUNNING to FAILED on e9414b4e-a65b-4398-95e4-9de240dc1fe8 @ localhost (dataPort=-1). java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:223) at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:305) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:69) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:66) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:423) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:204) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:684) at org.apache.flink.streaming.runtime.tasks.StreamTask.executeInvoke(StreamTask.java:639) at org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:650) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:623) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:779) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:148) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:103) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266) at java.util.concurrent.FutureTask.run(FutureTask.java) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:42) at com.ververica.cdc.connectors.mysql.debezium.task.context.MySqlErrorHandler.setProducerThrowable(MySqlErrorHandler.java:72) at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1185) at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ... 1 more Caused by: io.debezium.DebeziumException: command 30 not supported now,sql=[J06�͔k��'*�}O] Error code: 1105; SQLSTATE: HY000. at io.debezium.connector.mysql.MySqlStreamingChangeEventSource.wrap(MySqlStreamingChangeEventSource.java:1146) ... 5 more Caused by: com.github.shyiko.mysql.binlog.network.ServerException: command 30 not supported now,sql=[J06�͔k��'*�}O] at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:937) ... 3 more 2023-05-31 01:29:54 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:225) at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:169) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:130) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:354) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:68) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:496) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:203) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:809) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:761) at org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:937) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) at java.lang.Thread.run(Thread.java:750) Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:150) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: com.ververica.cdc.connectors.shaded.org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:42) at com.ververica.cdc.connectors.mysql.debezium.task.context.MySqlErrorHandler.setProducerThrowable(MySqlErrorHandler.java:72) at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1185) at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ... 1 more Caused by: io.debezium.DebeziumException: dispatch command 30 not supported now, connection id[565580] user[advc] addr[10.130.0.98:47164] db[INFORMATION_SCHEMA] Error code: 1105; SQLSTATE: HY000. at io.debezium.connector.mysql.MySqlStreamingChangeEventSource.wrap(MySqlStreamingChangeEventSource.java:1146) ... 5 more Caused by: com.github.shyiko.mysql.binlog.network.ServerException: dispatch command 30 not supported now, connection id[565580] user[advc] addr[10.130.0.98:47164] db[INFORMATION_SCHEMA] at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:937) ... 3 more flink 1.4.5 flink-sql-mysql-cdc 2.2.1 jar mysql5.7.99 java.lang.RuntimeException: One or more fetchers have encountered exception at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:225) at org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:169) at org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:130) at org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:354) at org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:68) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:496) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:203) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:809) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:761) at org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:937) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) at java.lang.Thread.run(Thread.java:750) Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:150) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: com.ververica.cdc.connectors.shaded.org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:42) at com.ververica.cdc.connectors.mysql.debezium.task.context.MySqlErrorHandler.setProducerThrowable(MySqlErrorHandler.java:72) at io.debezium.connector.mysql.MySqlStreamingChangeEventSource$ReaderThreadLifecycleListener.onCommunicationFailure(MySqlStreamingChangeEventSource.java:1185) at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:973) at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:606) at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:850) ... 1 more Caused by: io.debezium.DebeziumException: dispatch command 30 not supported now, connection id[565580] user[advc] addr[xxx.xxx.xxx.xxx:47164] db[INFORMATION_SCHEMA] Error code: 1105; SQLSTATE: HY000. at io.debezium.connector.mysql.MySqlStreamingChangeEventSource.wrap(MySqlStreamingChangeEventSource.java:1146) ... 5 more Caused by: com.github.shyiko.mysql.binlog.network.ServerException: dispatch command 30 not supported now, connection id[565580] user[advc] addr[xxx.xxx.xxx.xxx:47164] db[INFORMATION_SCHEMA] at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:937) ... 3 more 我也报错。但我用的 hologres-connector-flink-1.15 ,没有可配置 server-id 的参数,怎么解。。
gharchive/issue
2021-12-10T03:28:34
2025-04-01T06:46:10.080439
{ "authors": [ "1747166759", "Aaronzk", "DongTL", "FQKang", "cmzz", "deepthinkin", "dongpengfei2", "frozenheartboy", "libra612", "maben996", "peterjqy", "slankka", "winskin", "yongzhao-qcc", "zhangyukun230", "zjp123456" ], "repo": "ververica/flink-cdc-connectors", "url": "https://github.com/ververica/flink-cdc-connectors/issues/709", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }