id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
271211603
Issues with sub options Issue by wadewegner Thursday Oct 01, 2015 at 18:07 GMT Originally opened as https://github.com/gsscoder/commandline/issues/249 I'm having an issue with the parser not figuring out my sub-options. Here's the code in Main: var invokedVerb = ""; var invokedVerbInstance = new object(); var options = new Options(); if (!CommandLine.Parser.Default.ParseArguments(args, options, (verb, subOptions) => { // if parsing succeeds the verb name and correct instance // will be passed to onVerbCommand delegate (string,object) invokedVerb = verb; if (!string.IsNullOrEmpty(invokedVerb)) invokedVerbInstance = subOptions; })) { Environment.Exit(CommandLine.Parser.DefaultExitCodeFail); } This is pretty much right from the wiki. My options are defined as this: class Options { public Options() { CpVerb = new CPSubOptions { }; } [VerbOption("cp", HelpText = "Interact with cp.")] public CPSubOptions CpVerb { get; set; } [HelpVerbOption] public string GetUsage(string verb) { return HelpText.AutoBuild(this, verb); } } class CPSubOptions { [Option('n', "name", HelpText = "Get details name.")] public string Name { get; set; } } When I run with cp as the args, it works and the parser determines that CPSubOptions is the subOptions and invokedVerbInstance is set appropriately. However, when I run with cp --name blah or cp -n blah, the subOptions is null, so it's not parsed properly. Any help? Comment by nemec Sunday Oct 04, 2015 at 06:50 GMT What version are you using? Based on the stable branch, your sample code works perfectly (for me). Based on Master, it fails to compile since the library has gone through some major changes for 2.0. Comment by Mizipzor Tuesday Feb 02, 2016 at 20:25 GMT @wadewegner did you figure it out? I also tried your code and as @nemec I get the expected result; subOptions is not null and Name has a value. Comment by gsscoder Monday Oct 05, 2015 at 22:43 GMT :paperclip: Please read: #250. Thanks for collaboration and patience! :sweat_smile: Giacomo P.S.: valid only for me, if a main contrib. is available he's obviously free to support devs/users of the project. Comment by Jamby93 Monday Apr 18, 2016 at 19:59 GMT Same issue here with an almost identical class configuration. subOptions passed to callback is null, although options.subOption is valid with correct values (options.CpVerb in your case). Comment by nemec Friday Apr 29, 2016 at 00:08 GMT @Jamby93 @brandorf latest stable or prerelease? I just tested again with 1.9.71 from NuGet, copy and pasting the above, and it works perfectly. VS 2015, .Net 4.5.2 in case you're running something different. The exact code: https://gist.github.com/nemec/bc9d78e47d7a7d8a72a9711488b3bfef Comment by brandorf Wednesday Apr 27, 2016 at 19:16 GMT I'm seeing this as well with latest version from nuget.
gharchive/issue
2017-11-04T18:02:09
2025-04-01T04:33:51.603706
{ "authors": [ "ericnewton76" ], "repo": "commandlineparser/commandline", "url": "https://github.com/commandlineparser/commandline/issues/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2661558124
Custom picker: selection check overlay icon sometimes not visible https://github.com/user-attachments/assets/8d1bbfe0-54f1-4640-8d4a-838334232140 Hey @nicolas-raoul i want work on this issue .
gharchive/issue
2024-11-15T10:39:26
2025-04-01T04:33:51.631587
{ "authors": [ "nicolas-raoul", "sonalyadav1" ], "repo": "commons-app/apps-android-commons", "url": "https://github.com/commons-app/apps-android-commons/issues/5931", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
259974596
Tech-improvement: Use 'newInstance' pattern for fragments For fragments, use the newInstance pattern instead of creating new objects directly. Fragment newInstance() pattern prevents memory leaks and makes it obvious what arguments are used by the fragment. May I take this up? @Ishti: Yes go ahead 🙂 On Oct 7, 2017 19:50, "Ishti Gupta" notifications@github.com wrote: May I take this up? — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/commons-app/apps-android-commons/issues/884#issuecomment-334926614, or mute the thread https://github.com/notifications/unsubscribe-auth/AAGFBh9hQ8aAOKFqaiAzgAq_rvXevrkPks5sp1d4gaJpZM4PhZ3n . Are we using the "newInstance() pattern", or did we decide against it? Chris.
gharchive/issue
2017-09-23T01:50:27
2025-04-01T04:33:51.635231
{ "authors": [ "chrisdebian", "ishtig", "nicolas-raoul", "psh" ], "repo": "commons-app/apps-android-commons", "url": "https://github.com/commons-app/apps-android-commons/issues/884", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2616009686
Fix issue #5814 - Rotated pictures shown as failed to upload, actually uploaded successfully area:EditActivity.kt, UploadWorker.kt Fixes #5814 Issue page https://github.com/commons-app/apps-android-commons/issues/5814 Added permission handling to allow the program to properly access the rotated image. Additionally fixed a bug: the database was being accessed on the wrong thread, causing the program to crash. I cannot confirm that this issue is fully resolved, as I was unable to reproduce the issue's description on my test devices. I still need the original reporter of the issue to test it again. (Currently, none of the problems described in the issue occur on my test devices, so I believe all possible fixes have been made.) Personally, I think the issue described was caused by not obtaining read-write permissions, which led to a specific situation where the program may have skipped the permission check during the testing environment at the time. This could have caused the upload to succeed but still show as failed. I was able to partially reproduce the bug: At the point where the upload should be show as failure, the program crashed. I believe that in the testing environment when the issue was raised, the program forcefully bypassed the error for some reason, which led to the second bug described in the issue. This error should have caused the crash. Would you mind solving the conflict? :-) The code from Main solved the same problem and used a better solution, so I adopted the code from Main Change it to final var Actually I somehow do not observe the issue on the main branch anymore. Are you still able to reproduce the issue on main? Does this issue exist? I was unable to reproduce this while uploading a rotated image. Can you @QUIETTTTTTTTTT share steps to reproduce if its still occurring on main?
gharchive/pull-request
2024-10-26T17:10:53
2025-04-01T04:33:51.639792
{ "authors": [ "QUIETTTTTTTTTT", "neeldoshii", "nicolas-raoul" ], "repo": "commons-app/apps-android-commons", "url": "https://github.com/commons-app/apps-android-commons/pull/5899", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2739422859
Remove temporary String.strip() implementation What changes did you make and why? Context: #5141 this PR fixed the problem about white space characters not being trimmed properly when using java.lang.String#trim() by implementing String#strip() internal implementation because it was introduced in JDK 11. Since, we're at 17 we can remove this one with the actual one. Thanks :)
gharchive/pull-request
2024-12-14T01:05:44
2025-04-01T04:33:51.641138
{ "authors": [ "parneet-guraya" ], "repo": "commons-app/apps-android-commons", "url": "https://github.com/commons-app/apps-android-commons/pull/6028", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
667671891
Typo in BaseDataTypes.xsd and Reported-Data.xsd Line 223 and 224 of BaseDataTypes.xsd has ACRatedOuput and should be ACRatedOutput (missing 't') in Type name The type is then referred to in Reported-Data.xsd on lines 520 & 689 so needs fixing there too. The above line numbers refer to the CEPC-8.0.0 files. Thanks for this @rogerdawkins - good spot. There is a problem with just fixing this right now: this error has been carried over from the legacy schema that we've inherited, and there is already a bunch of software out there that produces output to match the schema with the misspelled element (and probably has done for 10 years or more). We'll have to get in touch with Air Conditioning report software providers to co-ordinate the change in element name. But thanks for raising the issue - we will definitely get things moving to put it right... it just won't be immediate. @TheDoubleK No probs. The typo doesn't appear in the XML output, it is only an element and complexType definition in the XSD files so may be easier to implement. I'll leave it with you.
gharchive/issue
2020-07-29T08:56:15
2025-04-01T04:33:51.648759
{ "authors": [ "TheDoubleK", "rogerdawkins" ], "repo": "communitiesuk/epb-register-api", "url": "https://github.com/communitiesuk/epb-register-api/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
867360486
Create an API for invalidation of ICLA from PCC & send email to the concerned contributor Create a new API for the invalidation of ICLA from Project Manager Console. And also send the following email for it. Hello {{.RecipientName}}, This is a notification email from EasyCLA regarding the CLA Group {{.CLAGroupName}}. You had previously signed an Individual CLA (ICLA) to contribute to the project on your own behalf. However, the Project Manager has marked your ICLA as invalidated. This might be because the ICLA may have been signed in error, if your contributions should have been on behalf of your employer rather than on your own behalf. As a result, you will no longer be able to contribute until you are again authorized under another signed CLA. You may need to arrange for your employer to sign a Corporate CLA (CCLA) for the project, or to add you to their authorization list if they have already signed one. Please contact the Project Manager for this project if you have questions about why you were removed. The Project Manager is {{PROJECT MANAGER - name and email address}}. {normal CLA signoff} Deployed to DEV & Staging through https://github.com/communitybridge/easycla/releases/tag/v2.0.39 Deployed to PROD through https://github.com/communitybridge/easycla/releases/tag/v2.0.44
gharchive/issue
2021-04-26T07:24:54
2025-04-01T04:33:51.652062
{ "authors": [ "vinod-kadam" ], "repo": "communitybridge/easycla", "url": "https://github.com/communitybridge/easycla/issues/2897", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1624854175
Production : Different cla group project activity logs are displayed at "O-RAN Software Community" project. Steps : Login company dashboard Navigate to url : https://organization.lfx.linuxfoundation.org/foundation/a092M00001IkOVVQA3/project/a092M00001IUjZPQA1/cla Actual results : Nephio Project activity log displayed at "O-RAN Software Community" project easycla page Refer screenshot : @thakurveerendras kindly verify Now properly O-RAN Software Community project activity logs are display at production, So making issue status QA-Done
gharchive/issue
2023-03-15T06:39:33
2025-04-01T04:33:51.655306
{ "authors": [ "thakurveerendras" ], "repo": "communitybridge/easycla", "url": "https://github.com/communitybridge/easycla/issues/3859", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2327339698
Move costs endpoint to api key This pr moves the cost endpoint to use the internal api key instead of OAuth2. This is because the payment api uses an api key. CI: No security warnings found
gharchive/pull-request
2024-05-31T09:14:19
2025-04-01T04:33:51.656175
{ "authors": [ "amartin7663", "ch-code-analysis" ], "repo": "companieshouse/accounts-filing-api", "url": "https://github.com/companieshouse/accounts-filing-api/pull/89", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2243210116
bug: modify CHANGED_RESOURCE_URI to include private JU-247 CI: Security warnings found!
gharchive/pull-request
2024-04-15T09:47:05
2025-04-01T04:33:51.656967
{ "authors": [ "ch-code-analysis", "hwatkins-ch" ], "repo": "companieshouse/company-exemptions-data-api", "url": "https://github.com/companieshouse/company-exemptions-data-api/pull/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
327914476
Incorrect matrix type coersion in rowsum.y / rowsum.z These functions use class() to cast reduced matrices to integer, when they should use storage.mode(). Under certain conditions (such as when running simulateCells()), this can cause crashes. @joshua-d-campbell for general reference: I realized we were writing things like class(mat) = "integer" to try to coerce matrices to the right type. I then realized that class() refers more to the data structure than the type of the actual contained variables. We should use storage.mode(mat) = "integer" instead.
gharchive/issue
2018-05-30T21:59:15
2025-04-01T04:33:51.671552
{ "authors": [ "definitelysean" ], "repo": "compbiomed/celda", "url": "https://github.com/compbiomed/celda/issues/245", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
49632640
Could I override the carret position to get custom offsets? I'd like to calculate pixel offsets of other "carrets" than the selected one. Could I provide my own character offset for which I want the pixel offset? API example? What you you mean by API example? @Darker: please reopen if you're still looking for this feature and can provide an example of some sort. Thanks!
gharchive/issue
2014-11-20T23:27:01
2025-04-01T04:33:51.678274
{ "authors": [ "Darker", "dandv", "jonathanong" ], "repo": "component/textarea-caret-position", "url": "https://github.com/component/textarea-caret-position/issues/18", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
964925755
Building image with scratch and amzn2 and publish to AWS ECR Public Proposed changes Adds AWS Codebuild settings, AWS CFN template for AWS CodeBuild. This builds the prometheus exporter in two architectures and publishes a manifest for version 0.9.0 Version of images on AWS have not yet the latest codebuild images version
gharchive/pull-request
2021-08-10T12:22:56
2025-04-01T04:33:51.679799
{ "authors": [ "JohnPreston" ], "repo": "compose-x/nginx-prometheus-exporter", "url": "https://github.com/compose-x/nginx-prometheus-exporter/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
158002973
BitrixInstaller: bitrix-theme или bitrix-template? В описании указано bitrix-template, а на самом деле оно bitrix-theme. Как так? @niksamokhvalov Да, косяк, спасибо.
gharchive/issue
2016-06-01T20:45:37
2025-04-01T04:33:51.690260
{ "authors": [ "niksamokhvalov", "rschweppes" ], "repo": "composer/installers", "url": "https://github.com/composer/installers/issues/305", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1734400492
Need a tool to stop a task once in progress. we need a tool which can stop a long running worker task one in progress. As a first step we really need to signal the workers (via the taskManager) to stop when they user types stop at the rcf read prompt
gharchive/issue
2023-05-31T14:46:09
2025-04-01T04:33:51.698032
{ "authors": [ "stephengaito" ], "repo": "computeFarm/pyComputeFarm", "url": "https://github.com/computeFarm/pyComputeFarm/issues/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2560040149
[SETUP]: Dockerfile missing Need a Dockerfile and compose file so that any contributor can easily run the project without bearing the trouble of installing the ruby and jekyll. Things to make sure: The container is build through the Dockerfile The volume should bind between the host and the container I will work on this issue, assign it to me. Sure go ahead, i will assign this to you.
gharchive/issue
2024-10-01T19:34:27
2025-04-01T04:33:51.699625
{ "authors": [ "aayush105", "gaurovgiri" ], "repo": "computerclubkec/computerclubkec.github.io", "url": "https://github.com/computerclubkec/computerclubkec.github.io/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1267761080
Some great changesgit add wd.py I have some great idea!! Here is the code... Interesting idea...
gharchive/pull-request
2022-06-10T16:18:39
2025-04-01T04:33:51.700716
{ "authors": [ "jlphillipsphd", "thejoshuacox" ], "repo": "coms-reu/testing", "url": "https://github.com/coms-reu/testing/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1832637130
Search packages by tags As of now, packages have some tags assigned to them, but those have just informative purpose, and thus not very useful. I would like to have ability to search packages with specific tag(s). For instance, see intx package below: It has tag #biginteger - I would like to click on it and see other packages that implement extended precision arithmetic (now it is just inactive element on page). More even, if I now enter "biginteger" in search box, I won't find even intx iself. Being about to select "all" would be super useful because the tags are super inconsistent and have no structure. Taking testing libraries, I want to generically search for them to get an idea of what exists... however not all of them have "test" in the name and there are 2 dozen + tags that need to be Manually selected Another example where there is more then one page It would also be great to sort the tags like dee2db1cbaa6c9e56086e9b758c1bf93da50fb30 :)
gharchive/issue
2023-08-02T07:50:26
2025-04-01T04:33:51.886123
{ "authors": [ "VL-Impact", "prince-chrismc" ], "repo": "conan-io/web", "url": "https://github.com/conan-io/web/issues/69", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
235335262
Page Attribute - Content included in search index / Pagelist Create a new page attribute without Content included in search index. This field is not able to be filtered by in Pagelist. Edit the attribute and mark "Content included in search index" and clear the c5 cache. This attribute still cannot be filtered in \Concrete\Core\Page\PageList unless you completely recreate the attribute. I think you mean "Field available in advanced search." instead of "Content included in search index." But yes, this appears to be true.
gharchive/issue
2017-06-12T19:07:02
2025-04-01T04:33:51.934320
{ "authors": [ "aembler", "jeverd01" ], "repo": "concrete5/concrete5", "url": "https://github.com/concrete5/concrete5/issues/5552", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
616467085
[cli] c5 cli doesn't grab the correct database config file if env variable CONCRETE5_ENV is set. If the concrete5 environment is set with the environment variable "CONCRETE5_ENV" to a custom value like "develop", the cli commands fail with the following error: dockeruser@6905ed170aa1://srv/app# concrete/bin/concrete5 list Doctrine\DBAL\Driver\PDOException: SQLSTATE[HY000] [2002] No such file or directory in file /srv/app/concrete/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php on line 31 Following situation: Environment is set in a nginx docker container. # nginx config fastcgi_param CONCRETE5_ENV develop; folder structure app +-- application | +-- config | +-- database.php | +-- develop.database.php Issue: The c5 installation works without any problem. But if I run any of the available cli commands within the container, the cli can't grab the correct database connection. Inserting the credentials from develop.database.php into the database.php file solves the problem. So the cli clearly grabs the wrong database config. Solutions: Either should the cli grab the correct database connection by the environment provided ... ... or there should be a cli argument --env=develop for each command. What's the concrete5 version? I tried to fix this issue here 👉 https://github.com/concrete5/concrete5/pull/8447 Should work on the latest develop branch. @biplobice Thanks. I'll test it with your pull request. Forgot to search in the pull request for solutions. The c5 version is: 8.5.2 :facepalm: Solution is too obvious. Since in my development environment CONCRETE5_ENV gets set by fastcgi_param in nginx , there is no way this variable gets set, if php is accessed directly within the container. (The request/call newer passes nginx, so no variabe is set.) By adding the following to my docker compose setup, every thing works fine. #docker-compose.yml php: ... environment: - CONCRETE5_ENV=${CONCRETE5_ENV} env_file: .env #.env CONCRETE5_ENV=develop
gharchive/issue
2020-05-12T08:33:05
2025-04-01T04:33:51.939970
{ "authors": [ "Kaapiii", "biplobice" ], "repo": "concrete5/concrete5", "url": "https://github.com/concrete5/concrete5/issues/8638", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
647061927
Delete Database Query Log Single Page A Japanese user claimed that database query log single page still lives on upgraded 8.5.3 & 8.5.4. I guess it will be nice to discuss whether to delete or keep database query log single page during next upgrade if exits. https://concrete5-japan.org/community/forums/translate/post-19121/ IMO, I don't mind of keeping it some user may add additional customization to it. If upgrade deletes the Single Page, it may delete some user's customization by accident. I just want to start low priority discussion what to do. (whether we have time to add or not) It shouldn't be. The code is fully gone, it's removed in Version20181212000000
gharchive/issue
2020-06-29T02:12:57
2025-04-01T04:33:51.942566
{ "authors": [ "aembler", "katzueno" ], "repo": "concrete5/concrete5", "url": "https://github.com/concrete5/concrete5/issues/8811", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2663557175
Update to belt 1.2 Checklist [x] Used a personal fork of the feedstock to propose changes [x] Bumped the build number (if the version is unchanged) [x] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [x] Ensured the license file is being packaged. Bumping version from 1.1 to 1.2. Hi! This is the friendly automated conda-forge-linting service. I wanted to let you know that I linted all conda-recipes in your PR (recipe/meta.yaml) and found some lint. Here's what I've got... For recipe/meta.yaml: ❌ This recipe is using a compiler, which now requires adding a build dependence on {{ stdlib("c") }} as well. Note that this rule applies to each output of the recipe using a compiler. For further details, please see https://github.com/conda-forge/conda-forge.github.io/issues/2102. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe/meta.yaml) and found it was in an excellent condition. @conda-forge-admin, please rerender
gharchive/pull-request
2024-11-16T00:41:12
2025-04-01T04:33:51.953200
{ "authors": [ "MichaelEhrlichman", "conda-forge-admin" ], "repo": "conda-forge/belt-feedstock", "url": "https://github.com/conda-forge/belt-feedstock/pull/1", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
243856791
Split clangdev libraries/headers to separate feedstock? Hey all, I have some applications which MUST link against clang headers from 4.0.0, but I would still like to use the system installation of clang (older) as it is required by nvcc. The current way clangdev's feedstock is setup makes it difficult to do this, as I always get both the libraries/headers as well as the clang binaries as a package deal. Is there some way we can split them up so they are installable separately? (For backwards compatibility, I would assume that clangdev continues to ship both, and there would be a separate libraries only package added.) I'm willing to put in the work if this is something that you would support. If you are building a simple C application, there is no need to have clangdev as a run time dependency. That said, if you are using C++ or OpenMP, there are some known issues due to clangdev being a run time dependency. At least with the C++ case, static linking is a viable workaround/option. Though we are discussing how to resolve dependencies on dynamic libraries for C++ and OpenMP. It may be best resolved once conda-build 3 is generally available in conda-forge. Could you please share some more about your use case? For example what features are you using? What is needed at runtime? And so on? Hi @jakirkham, thanks for responding. So if I understand correctly, the suggestion is to list clangdev as a compile-time dependency ONLY, and make sure that we statically link against clangdev. (Actually, in general, I have always found it a bit murky how to determine the runtime versus compile dependencies of libraries I package.) As PET is not my project, I do not actually know if this will work but it seems like something to try first. @isuruf, this has already happened right or are there things that still need to be done here? Yes
gharchive/issue
2017-07-18T21:27:23
2025-04-01T04:33:51.959086
{ "authors": [ "ezyang", "isuruf", "jakirkham" ], "repo": "conda-forge/clangdev-feedstock", "url": "https://github.com/conda-forge/clangdev-feedstock/issues/15", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1110582429
fix run constraint for cudatoolkit>=11.0 This attempts to fix #67 and #71. According to https://docs.nvidia.com/deploy/cuda-compatibility/, cudatoolkit 11.X should run on all systems with __cuda 11.0-11.6. Similar to https://github.com/conda-forge/nvcc-feedstock/pull/71/files, the run_constraint is fixed. Checklist [x] Used a personal fork of the feedstock to propose changes [x] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition. I tested the suggested changes and ran into problems. Apparently the PTX instructions were incompatible. That PTX incompatibility is mentioned here: https://docs.nvidia.com/deploy/cuda-compatibility/#application-considerations. As I understand, this is a requirement introduced by the nvcc part (and therefore not relevant here), but I am not a 100% sure. In any case: once the compiled program contains PTX instructions, the run_constraint for that application MUST be as it was before __cuda => major.minor, even for versions >=11.X. I am not sure the proposed change is safe In any case: once the compiled program contains PTX instructions, the run_constraint for that application MUST be as it was before __cuda => major.minor, even for versions >=11.X. I am not sure the proposed change is safe In theory it's not safe, but I believe what the community previously landed on was that if a package is reliant on PTX it's that package's responsibility to pin the __cuda package appropriately to handle that. Otherwise, for packages without PTX there isn't really a path forward in allowing using older driver versions that should work fine. cc @jakirkham @leofang @isuruf to confirm Would it be too much if the bot provided this information if compiler('cuda') is detected during linting? What information? That you should override the default run_exports if your project uses PTX. That you should override the default run_exports if your project uses PTX. Using compiler('cuda') doesn't necessarily mean you're using PTX. You can do a compilation without PTX. That you should override the default run_exports if your project uses PTX. Using compiler('cuda') doesn't necessarily mean you're using PTX. You can do a compilation without PTX. True, but if I understand @jaimergp's point correctly we wanna give the PR reviewer a heads-up if potentially PTX could be generated/compiled in a feedstock. I think it's reasonable, though I wonder if it'd cause too much noise. We can move the discussion to conda-smithy, perhaps? Yes, that was my question, thanks @leofang It is about whether it would be too noisy to print this "hey pay attention if you are using PTX because this might bite you". I'll create an issue on conda-smithy for CUDA hints. I think we had more ideas around that too and we can brainstorm there. I adjusted the PR with the requested changes. Also thanks for the clarification about the PTX instructions Will let @jakirkham review before merging. Thanks @gabm! Thanks @gabm for the PR and everyone for the reviews! 😄
gharchive/pull-request
2022-01-21T15:17:17
2025-04-01T04:33:51.977617
{ "authors": [ "conda-forge-linter", "gabm", "jaimergp", "jakirkham", "kkraus14", "leofang" ], "repo": "conda-forge/cudatoolkit-feedstock", "url": "https://github.com/conda-forge/cudatoolkit-feedstock/pull/76", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1875656671
custatevec for aarch64/ppc64le requires glibc 2.27+ See https://github.com/conda-forge/cuquantum-feedstock/pull/47#issuecomment-1700751650. Once the conda-forge infrastructure is ready, we should restore these two lines https://github.com/conda-forge/cuquantum-feedstock/blob/ce8944d672469f8aef72cc7f6952f07df4be9909/recipe/meta.yaml#L99-L102 Right now they are commented out because we need to consume them from within the CI (to build cuquantum-python). Note: When the infra is ready, we should switch this line https://github.com/conda-forge/cuquantum-feedstock/blob/d21a50c51af96049b43d82971bda690b979a7af8/conda-forge.yml#L20 to use native_and_emulated to turn on QEMU at the test stage.
gharchive/issue
2023-08-31T14:30:08
2025-04-01T04:33:51.980970
{ "authors": [ "leofang" ], "repo": "conda-forge/cuquantum-feedstock", "url": "https://github.com/conda-forge/cuquantum-feedstock/issues/48", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
661348790
Specify rx version and match other dependencies in setup.py Checklist [x] Used a personal fork of the feedstock to propose changes [x] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [x] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition. I do have some suggestions for making it better though... For recipe: License is not an SPDX identifier (or a custom LicenseRef) nor an SPDX license expression. Documentation on acceptable licenses can be found here. The change in version ranges is due to warnings such as WARNING conda.models.version:get_matcher(528): Using .* with relational operator is superfluous and deprecated and will be removed in a future version of conda. Your spec was 3.*, but conda is ignoring the .* and treating it as 3 WARNING:conda.models.version:Using .* with relational operator is superfluous and deprecated and will be removed in a future version of conda. Your spec was 2.*, but conda is ignoring the .* and treating it as 2 WARNING conda.models.version:get_matcher(528): Using .* with relational operator is superfluous and deprecated and will be removed in a future version of conda. Your spec was 2.*, but conda is ignoring the .* and treating it as 2 See comment in cylc-flow feedstock PR for more. @conda-forge-admin, please rerender
gharchive/pull-request
2020-07-20T04:20:45
2025-04-01T04:33:51.986230
{ "authors": [ "conda-forge-linter", "kinow" ], "repo": "conda-forge/cylc-uiserver-feedstock", "url": "https://github.com/conda-forge/cylc-uiserver-feedstock/pull/6", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
644317098
@conda-forge-admin, please re-render @conda-forge-admin, please re-render. Hi! This is the friendly automated conda-forge-webservice. I just wanted to let you know that I rerendered the recipe in conda-forge/freexl-feedstock#15.
gharchive/issue
2020-06-24T04:54:35
2025-04-01T04:33:51.996496
{ "authors": [ "conda-forge-linter", "jakirkham" ], "repo": "conda-forge/freexl-feedstock", "url": "https://github.com/conda-forge/freexl-feedstock/issues/14", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2253968007
Rebuild with minor cleanups and to test libintl errors on Windows Checklist [x] Used a personal fork of the feedstock to propose changes [x] Bumped the build number (if the version is unchanged) ~[ ] Reset the build number to 0 (if the version changed)~ [x] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [x] Ensured the license file is being packaged. Seeing a failure to import Gtk with pygobject over at https://github.com/conda-forge/gnuradio-feedstock/pull/262 which seems to be related to recent glib changes and libintl. So in addition to cleaning up the recipe by removing the freetype dependency (which logs show we don't link) and adding the new stdlib jinja, this will test to see if that problem is solved by a rebuild. Looks like the atk rebuild has fixed the libintl errors, so this is ready.
gharchive/pull-request
2024-04-19T20:46:17
2025-04-01T04:33:52.015393
{ "authors": [ "ryanvolz" ], "repo": "conda-forge/gtk3-feedstock", "url": "https://github.com/conda-forge/gtk3-feedstock/pull/68", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1181099437
Migrate to Python 3.10 Checklist [ ] Used a personal fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Changes taken from https://github.com/conda-forge/kwant-feedstock/pull/93 Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition. @conda-forge-admin, please rerender @akhmerov, you pinned both Matplotlib and Sympy in https://github.com/conda-forge/kwant-feedstock/pull/89. It is not clear to me why exactly. I removed the innings here to try to see whether the builds will pass. @akhmerov, thanks for fixing 14.3! I have updated this PR. @conda-forge-admin, please rerender OK, this new failure is an absolute mystery :sweat_smile: Hopefully, it goes away when restarting the CI 🥲
gharchive/pull-request
2022-03-25T18:20:34
2025-04-01T04:33:52.038973
{ "authors": [ "akhmerov", "basnijholt", "conda-forge-linter" ], "repo": "conda-forge/kwant-feedstock", "url": "https://github.com/conda-forge/kwant-feedstock/pull/98", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
892724605
MNT: rerender Hi! This is the friendly automated conda-forge-webservice. I've rerendered the recipe as instructed in #6. Here's a checklist to do before merging. [ ] Bump the build number if needed. Fixes #6 Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
gharchive/pull-request
2021-05-16T18:01:22
2025-04-01T04:33:52.049738
{ "authors": [ "conda-forge-linter" ], "repo": "conda-forge/mypy-boto3-s3-feedstock", "url": "https://github.com/conda-forge/mypy-boto3-s3-feedstock/pull/7", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
788414806
Add jdbc support Checklist [ ] Used a personal fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
gharchive/pull-request
2021-01-18T16:47:43
2025-04-01T04:33:52.053236
{ "authors": [ "conda-forge-linter", "nehaljwani" ], "repo": "conda-forge/mysql-connector-cpp-feedstock", "url": "https://github.com/conda-forge/mysql-connector-cpp-feedstock/pull/7", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1429315732
Rebuild for Python 3.11 Checklist [X] Used a personal fork of the feedstock to propose changes [X] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [X] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [X] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition. @conda-forge-admin, please rerender Closing in favor of #4.
gharchive/pull-request
2022-10-31T06:15:33
2025-04-01T04:33:52.057176
{ "authors": [ "XuehaiPan", "conda-forge-linter" ], "repo": "conda-forge/nvitop-feedstock", "url": "https://github.com/conda-forge/nvitop-feedstock/pull/3", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2681688123
Revert SQLAlchemy upgrade Checklist [ ] Used a personal fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe/meta.yaml) and found it was in an excellent condition. I do have some suggestions for making it better though... For recipe/meta.yaml: ℹ️ noarch: python recipes should usually follow the syntax in our documentation for specifying the Python version. For the host section of the recipe, you should usually use python {{ python_min }} for the python entry. For the run section of the recipe, you should usually use python >={{ python_min }} for the python entry. For the test.requires section of the recipe, you should usually use python {{ python_min }} for the python entry. If the package requires a newer Python version than the currently supported minimum version on conda-forge, you can override the python_min variable by adding a Jinja2 set statement at the top of your recipe (or using an equivalent context variable for v1 recipes). ℹ️ The recipe is not parsable by parser conda-recipe-manager. Your recipe may not receive automatic updates and/or may not be compatible with conda-forge's infrastructure. Please check the logs for more information and ensure your recipe can be parsed. This message was generated by GitHub Actions workflow run https://github.com/conda-forge/conda-forge-webservices/actions/runs/11966206728. Examine the logs at this URL for more detail. appears correct the way it is ; reverting would be a regression / mismatch against source https://github.com/apache/iceberg-python/blob/main/pyproject.toml#L79C1-L79C54 sqlalchemy = { version = "^2.0.18", optional = true } Thanks @rxm7706 !!
gharchive/pull-request
2024-11-22T03:59:38
2025-04-01T04:33:52.075448
{ "authors": [ "conda-forge-admin", "rxm7706", "sumanth-manchala" ], "repo": "conda-forge/pyiceberg-feedstock", "url": "https://github.com/conda-forge/pyiceberg-feedstock/pull/27", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
458200167
Update to version 0.9.500.2.0 Checklist [x] Used a fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
gharchive/pull-request
2019-06-19T20:00:03
2025-04-01T04:33:52.078768
{ "authors": [ "conda-forge-linter", "dbast" ], "repo": "conda-forge/r-rcpparmadillo-feedstock", "url": "https://github.com/conda-forge/r-rcpparmadillo-feedstock/pull/8", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
416383998
Update to version 0.12, update recipe Checklist [x] Used a fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
gharchive/pull-request
2019-03-02T10:36:33
2025-04-01T04:33:52.085107
{ "authors": [ "conda-forge-linter", "dbast" ], "repo": "conda-forge/r-servr-feedstock", "url": "https://github.com/conda-forge/r-servr-feedstock/pull/2", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2448467929
[ci skip] adding user @rxm7706 Hi! This is the friendly automated conda-forge-webservice. I've added user @rxm7706 as instructed in #25. Merge this PR to add the user. Please do not rerender this PR or change it in any way. It has [ci skip] in the commit message to avoid pushing a new build and so the build configuration in the feedstock should not be changed. Please contact conda-forge/core to have this PR merged, if the maintainer is unresponsive. Fixes #25 @rxm7706 , please give your 👍🏼 for this.
gharchive/pull-request
2024-08-05T12:44:43
2025-04-01T04:33:52.087055
{ "authors": [ "conda-forge-admin", "igortg" ], "repo": "conda-forge/radon-feedstock", "url": "https://github.com/conda-forge/radon-feedstock/pull/26", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1769916737
build: :arrow_up: Update recipe to v0.19.0 Checklist [ ] Used a personal fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. @conda-forge-admin, please rerender
gharchive/pull-request
2023-06-22T15:20:48
2025-04-01T04:33:52.103067
{ "authors": [ "Anselmoo" ], "repo": "conda-forge/spectrafit-feedstock", "url": "https://github.com/conda-forge/spectrafit-feedstock/pull/54", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
161497254
Update to 1.4.0 Added multi-threaded support. thanks @lukecampbell Better date parsing. thanks @akorosov Travis cleanup Py 3.5 compat Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
gharchive/pull-request
2016-06-21T17:38:31
2025-04-01T04:33:52.146316
{ "authors": [ "conda-forge-linter", "ocefpaf" ], "repo": "conda-forge/thredds_crawler-feedstock", "url": "https://github.com/conda-forge/thredds_crawler-feedstock/pull/5", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1176188741
Delete conda_build_config.yaml @conda-forge-admin please rerender half way through the pytorch migration I found a patch that made gcc 10 work with cuda. I want to submit this bofore I forget about it. Checklist [ ] Used a personal fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition. hmm, the rerender didn't happen. :/
gharchive/pull-request
2022-03-22T02:35:11
2025-04-01T04:33:52.150424
{ "authors": [ "conda-forge-linter", "hmaarrfk" ], "repo": "conda-forge/torchsparse-feedstock", "url": "https://github.com/conda-forge/torchsparse-feedstock/pull/17", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1996916718
Remove python constraints Checklist [x] Used a personal fork of the feedstock to propose changes [x] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. @conda-forge-admin, please rerender @conda-forge-admin, please rerender @h-vetinari, hi, could I get your help here please? Seems you figured things out? In general, if you want to build per python version, don't specify a lower bound, but leave it as plain - python, like that conda smithy will correctly generate separate jobs @h-vetinari, yes, we figured out the root cause. I don't know if it is really an issue with conda or not, but it looks frustrating. Maintainers of other conda-forge feedstocks may see this issue in their feedstock at some point. It would be great to fix the issue or document it somewhere. Thanks for your response, btw.
gharchive/pull-request
2023-11-16T13:56:06
2025-04-01T04:33:52.154956
{ "authors": [ "YarShev", "h-vetinari" ], "repo": "conda-forge/unidist-packages-feedstock", "url": "https://github.com/conda-forge/unidist-packages-feedstock/pull/17", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1664965963
v0.23.0 Checklist [ ] Used a personal fork of the feedstock to propose changes [ ] Bumped the build number (if the version is unchanged) [ ] Reset the build number to 0 (if the version changed) [ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering) [ ] Ensured the license file is being packaged. @conda-forge-admin, please rerender
gharchive/pull-request
2023-04-12T17:31:01
2025-04-01T04:33:52.164279
{ "authors": [ "jeff-hernandez" ], "repo": "conda-forge/woodwork-feedstock", "url": "https://github.com/conda-forge/woodwork-feedstock/pull/65", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1234473998
Fix extras require issue Closes #325 I rebased it on the main branch.
gharchive/pull-request
2022-05-12T20:56:41
2025-04-01T04:33:52.165365
{ "authors": [ "woutdenolf" ], "repo": "conda-incubator/grayskull", "url": "https://github.com/conda-incubator/grayskull/pull/326", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1387938723
Export PollingMode Describe the purpose of your pull request Export PollingMode enum so it can be used in consuming apps [ ] I have covered the applied changes with automated tests. [ ] I have executed the full automated test set against my changes. [ ] I have validated my changes against all supported platform versions. [x] I have read and accepted the contribution agreement. v1.0.1 is released, I hope it will work for you. Saw that, thanks for the quick reply! Its works fine 👍
gharchive/pull-request
2022-09-27T15:15:00
2025-04-01T04:33:52.266594
{ "authors": [ "henninghall", "laliconfigcat" ], "repo": "configcat/react-sdk", "url": "https://github.com/configcat/react-sdk/pull/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1060834107
Add SmtSpec for SMT proofs This will be needed for implementing ADR-040 in Cosmos-SDK; specifically this PR depends on it: https://github.com/cosmos/cosmos-sdk/pull/10015 This will need a Go version bump, should it go to 0.6.8 or would 0.7.0 be more semver-appropriate? @roysc bumping I am happy to merge as soon as there are a couple test specs. There is a new JSON format for them an all now (thanks to the work on solidity), which might be easier to use cross-language. But even adding a positive test in Go would be nice. Sorry, I haven't had time to get to creating a proper serialization tool like ics23-iavl for the test data. I could look at just copying over some data as Go code in the meantime, but it's not particularly urgent yet to merge the spec into this repo, so I figured this could wait for the proper solution. After testing more thoroughly, I found that the proof verification algorithm needs to be updated to check for empty child nodes in non-existence proof paths to support SMT proofs - see PR at https://github.com/confio/ics23/pull/61. I've also written a ics23-smt library to generate and check proof data (that depends on the above PR). I will rebase this branch on that fix and add the tests and data to complete it. After testing more thoroughly, I found that the proof verification algorithm needs to be updated to check for empty child nodes in non-existence proof paths to support SMT proofs 😄 See I'm not just being annoying when asking for test cases... they do find issues. I will review #61 now Haha yes, the tests in the SDK were a bit too trivial it seems Bumping this one to check the status. Are you waiting for my review or am I waiting for more tests? This is also ready for review, sorry for the delay. The JS part is not complete, but I will leave it in for now. I've rebased and updated this now that the placeholder check is merged. I've also had to add a fix to the placeholder check logic, since it was being called with the wrong branch index. The fix is to detect the index from the padding within the empty check.
gharchive/pull-request
2021-11-23T04:42:08
2025-04-01T04:33:52.272833
{ "authors": [ "ethanfrey", "roysc" ], "repo": "confio/ics23", "url": "https://github.com/confio/ics23/pull/57", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
435221578
Add support for ethereum patricia trie proofs One of the most used Merkle stores is the patricia trie from ethereum, and particularly with the thought of ethereum interoperability and peg zones, we should also add support for translating go-ethereum/trie proofs into this format. It will likely force some changes as well between trees and tries. And will help complete emperical interoperability with all major Merkle stores in the foreseeable cosmos ecosystem. In particular, we support proofs from SecureTrie You can start with Trie, and adapt to SecureTrie later, as it is clearer Some links: Trie.Prove() Write all nodes to proofsDB VerifyProof - this algorithm must be in line with confio/proofs - all we care about replicating decodeNode Somehow decodeNode should validate the hashes line up Made an attempt to implement this in https://github.com/confio/proofs-ethereum Turns out to be impossible to securely extract the key from the patricia tree merkle proofs without a large amount of custom code. I mentioned this in the README #14 and will close this issue. If there is a viable approach to implement, I will attempt to do so.
gharchive/issue
2019-04-19T15:51:46
2025-04-01T04:33:52.277527
{ "authors": [ "ethanfrey" ], "repo": "confio/proofs", "url": "https://github.com/confio/proofs/issues/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
590965767
AvroGen: Uncompilable code for records having a field with the same name as the record Description When generating code for a record having a field with the same name as the record, AvroGen produces a class containing a property having the same name as the class, which C# does not allow. Confluent.Apache.Avro package version used is 1.7.7.7 This is problematic when using the Avro.CodeGen class to generate Avro types at run-time. How to reproduce Invoke AvroGen on the schema below. It generates a class named "cars" containing a property named "cars". { "type": "record", "name": "person", "namespace": "Kafka.Tests.AvroTypes", "fields": [ { "name": "garage", "type": { "type": "record", "name": "cars", "fields": [ { "name": "cars", "type": { "type": "array", "items": { "type": "record", "name": "car", "fields": [ { "name": "model", "type": "string" }, { "name": "make", "type": "string" } ] } } } ] } } ] } In v1.4.0, we've migrated to using the official Apache Avro packages. Please these versions of the tools instead and report any issues to the Avro repo: https://github.com/apache/avro We've also introduced support for protobuf and json - you might want to consider these serialization formats instead which have better implementations in .net.
gharchive/issue
2020-03-31T09:59:53
2025-04-01T04:33:52.281275
{ "authors": [ "bhaeussermann", "mhowlett" ], "repo": "confluentinc/confluent-kafka-dotnet", "url": "https://github.com/confluentinc/confluent-kafka-dotnet/issues/1230", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
997334364
[Question] Why is the JsonSerializer not validating against the Schema in the Schema Registry? Description I have a small question regarding the JsonSerializer - it extracts the Schema from the generic as follows: this.schema = NJsonSchema.JsonSchema.FromType() https://github.com/confluentinc/confluent-kafka-dotnet/blob/5331f7af6063efcebbc3b830c59ded8f22ceb611/src/Confluent.SchemaRegistry.Serdes.Json/JsonSerializer.cs#L100 https://github.com/confluentinc/confluent-kafka-dotnet/blob/5331f7af6063efcebbc3b830c59ded8f22ceb611/src/Confluent.SchemaRegistry.Serdes.Json/JsonSerializer.cs#L101 Since we have a Schema Registry configured, shouldn't the JsonSerializer retrieve the Schema from the Schema Registry and perform the validation against it? Am I missing something? Because if someone modifies the Schema, but doesn't modify the model in his application, the validation will always pass. How to reproduce ================\ Checklist Please provide the following information: [ ] A complete (i.e. we can run it), minimal program demonstrating the problem. No need to supply a project file. [ ] Confluent.Kafka nuget version. [ ] Apache Kafka version. [ ] Client configuration. [ ] Operating system. [ ] Provide logs (with "debug" : "..." as necessary in configuration). [ ] Provide broker log excerpts. [ ] Critical issue. this is effectively checked in the SerializeAsync method.
gharchive/issue
2021-09-15T17:23:47
2025-04-01T04:33:52.286619
{ "authors": [ "DrorOzgaon", "mhowlett" ], "repo": "confluentinc/confluent-kafka-dotnet", "url": "https://github.com/confluentinc/confluent-kafka-dotnet/issues/1686", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
322970683
Add HTTPS to services We have some services that have HTTP front ends. Those services should be secured with HTTPS in the SSL enabled configurations. https://github.com/confluentinc/cp-ansible/pull/27
gharchive/issue
2018-05-14T20:33:37
2025-04-01T04:33:52.316219
{ "authors": [ "cotedm", "coughman" ], "repo": "confluentinc/cp-ansible", "url": "https://github.com/confluentinc/cp-ansible/issues/12", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
394858042
Can't import pubkeys with passphrase properly Bug description I can't import a pubkey with passphrase created on my mac to connectbot properly. Pubkey with no passphrase is working normally. Steps to reproduce on Mac ❯ ssh-keygen -t rsa -C mbp2016 -f ~/.ssh/id_with_passphrase enter "password" as passphrase id_with_passphrase -----BEGIN OPENSSH PRIVATE KEY----- b3BlbnNzaC1rZXktdjEAAAAACmFlczI1Ni1jdHIAAAAGYmNyeXB0AAAAGAAAABDfgSeBuI XCZKYbH8rUrCgIAAAAEAAAAAEAAAEXAAAAB3NzaC1yc2EAAAADAQABAAABAQDTBLWJWhnX /G8Ub5W7U5/b9TsEzya4v96R+O/3G0dPXCcJ/aprhGA/qL0sf1TNJtPkanJ6EyMNR2wiIG 4xU07o72cvZoYlopx5rc7KxTwGcQgYkNcdA+mzX5dvNiwOgLU9qQ1ZoIXSi6UGqYcpgD/7 OyCGifuGag+KxKI/r4LIHmk9f53jn2sXIqB4f8XECYla0j9ra8c5T0iPLMGGSXegokVJ4b 0hLZX+fCLEMiNxYfSAGWIPsm+jcLeTUkjCvPknfECkP9FB27WWDYzlSaNnD9IG7d93FUF5 hAERQKbhpDNdy5EJbqX2nZ44L531h470Gd7DqsIzZ+MlNVM5QTHnAAADwPLaRW9NKUankz GEYgz9ZDfTpYtDJh7uzsNisXq1C8A+mzvRSuho2OGT77WIF5GDQREbikxZy8KU77YXkj6g 9IGQ3TL5eERAGDMcfFq+5kfUCMd1pgXNp00P2By8mKrV434rpr9FRet7oawJNO2EE6HR2e /4KfrnoxIXrDu+GCARiGDYsJ0jsrMRblDsryCOuyZgbxPLcOJp0f+chb5E7lceQG+3QxLr ptpEC58P+MTg5EudB3VkIBhQSP3C+O2eIsQb9cOFzVSLITkd2PvZEGxt4NVTH/pXnS7tvj JW+glMcpnMWA3eVe95HUGD+hgE6YJ3cZnf2NnYUWDDqvldcY6vy38b/1dMRjPAFur10MmZ /uKwNrZVshhrlL6O8zsrdOQ01pxL7+MjiWZfwSRC5/pBZPdyIM4aQmzKNqx4C8V0cqJFzm X4s3ePSUXAY1ibyP4NQ0xe02CSjyxQe6zfqQap8VgrX+TYqKvkp0/QNbJvAZaH7mmjrFY1 XC0CYneljRxOWUfftgPLb81pqZsr5lpNSGuW8juio75xjdko48b2wjjVZaXx1GqnYv9YAI /xwNJzIJODnfRZFlVpOSStWPRW7hByvB7hl4qMU+PGDsPK+QxQR2H84jnYlt3dqnYSENfy HEGEq0CXnvf8LbUaLn775V6TM4MlfG9iPwlG4Qfp9wD19CiddvsqrRGaIIsvKSSy18Eav4 URk/4c0yAUL10NmroXS2mtMpIqkuzxVol2Olw2hXF+ZRYDN6xKutD16+G6ZSGtZ3146ZI/ /vIuqCHFywk1pI4TZkDtl8gwAvG9VCQWBYa9WHQ63u7/yh9EFUdeLN22wkWt1ONAE2xaXn TTITEYZtoW9HUUzXah+yNQIIjy9mbrAN+s4WBZSn+tHu5UJVv2mA/tZ3Dn3eBmkVTsWgPo kT+j6fTWhdG138kdWo7ygDcGLHzDhgMeC8TmIK6C7djZKIaYloiJhqSzTP1QXRUwfBr+Pl KYQaetOAnSushRV5QVmrZCuKx9l7J1ega1vxaCqUaVhf7LmP8uWjkA5NTyojM0JKp12v07 RdSlhMfPBexxjtDAz06jA5YM+y+7xovM5PamwLyqHHxmaqUsVd9930B+thYUCfrGNKSxZ3 Bz8xVylMZrIAYjLZpHsYy5gq2m61fj4KU19ISZiE/kLsz9pFBdJdYKd63346mK7HPCbr6f 1gf4mMAUXy/nM1kKhB8kwwurmNNRjFyb9hhc3uP+DcJmPlFHbl9iabmzkk6ASfWLWm4VEk 2bN3xK1g== -----END OPENSSH PRIVATE KEY----- id_with_passphrase.pub ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDTBLWJWhnX/G8Ub5W7U5/b9TsEzya4v96R+O/3G0dPXCcJ/aprhGA/qL0sf1TNJtPkanJ6EyMNR2wiIG4xU07o72cvZoYlopx5rc7KxTwGcQgYkNcdA+mzX5dvNiwOgLU9qQ1ZoIXSi6UGqYcpgD/7OyCGifuGag+KxKI/r4LIHmk9f53jn2sXIqB4f8XECYla0j9ra8c5T0iPLMGGSXegokVJ4b0hLZX+fCLEMiNxYfSAGWIPsm+jcLeTUkjCvPknfECkP9FB27WWDYzlSaNnD9IG7d93FUF5hAERQKbhpDNdy5EJbqX2nZ44L531h470Gd7DqsIzZ+MlNVM5QTHn mbp2016 copy id_with_passphrase to /sdcard/SSH Open Connectbot app, Go to Manege Pubkeys, And import the key. Click the imported key, and enter the password and tap 'unlock key' Toast "Bad password for key 'primary:SSH/id_with_passphrase. Authentication failed" appeared and can't open the key properly. Screenshots first key is created by mac with passphrase. not working. second key is created by mac with no passphrase. working. third key is created by Connectbot with passphrase. working fine. Android device Device: Galaxy S8 OS: Android 8.0.0 ConnectBot Version: latest 1.9.5 Mac OS: macOS Mojave 10.14.2 SSH Software and Version: OpenSSH_7.9p1, LibreSSL 2.7.3 Hello, Same issue here: Device: Oneplus 6T on OxygenOS 9.0.13 (Android 9) with ConnectBot 1.9.5-oss Key generated with OpenSSH_8.0p1, OpenSSL 1.1.1b FIPS 26 Feb 2019 Same issue Me too. Can we please get an answer as to why this could be happening? Thank you Sam issue here, as well same error. Used following to create key: ssh-keygen -t rsa -b 4096 used key with other clients and works fine. the same issue has not been solved for so long. Same issue with LineageOS 17.1, Android 10 (lineage_beryllium-userdebug 10 QQ2A); RSA 3072 bit key with passphrase, generated with openSSH 1:8.2p1-4 on ubuntu-20.04 Same issue with private key generated by Putty. It might be connected to not recognizing aes256-ctr cypher by sshlib connectbot/sshlib#37 that is used by OpenSSL 7.6+ as default and also PuttyGen for Ed25519 keys. This is fixed in 1.9.7, but this version is not yet available in Play Store or F-Droid as of 14.04.2021 The workaround is to convert key to aes256-cbc before load ssh-keygen -p -f id_ed25519-foobar -Z aes256-cbc It seems as though connectbot cannot handle OpenSSH's now private key format properly. As a workaround, adding the -m PEM switch while generating new or updating existing keys did the trick for me. Same issue the error still exist?
gharchive/issue
2018-12-30T09:18:44
2025-04-01T04:33:52.369072
{ "authors": [ "32u-nd", "H3mul", "IvanGorinich-GitHub", "KangarooCreativeTeam", "Tealk", "benediktneumayr", "remyj38", "rkuk", "sppmasterspp", "trekkie690", "umlx5h", "ziomzbronxu" ], "repo": "connectbot/connectbot", "url": "https://github.com/connectbot/connectbot/issues/684", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2329533824
[ST] Send button disappeared when bridging Describe the bug When bridging, after clicking "SEND", the UI showed "checking approval" and the button disappeared. Upon monitoring the console, we found this error message: Where did the user initiate this transfer? Connext UI Screenshots @oncall I also being able to reproduce this error by trying to bridge NEXT from Ethereum mainnet to Arbitrum However, on my console, I faced a different error message as shown below: All tickets on it: https://app.intercom.com/a/inbox/qge2vd8z/inbox/tickets/1183/linked-conversations?show_bulk_edit_modal=false&view=List
gharchive/issue
2024-06-02T08:28:59
2025-04-01T04:33:52.381011
{ "authors": [ "ekbainova", "tann9949" ], "repo": "connext/monorepo", "url": "https://github.com/connext/monorepo/issues/6250", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
706014738
connection::Connection should transmit the Error packet Right now it just uses the Error packet to create an io::Error. It'd be best to transmit the courtesy Error packet before returning the io::Error. Do this for put and get. Hi, I'm interested in working on this. To clarify, do you want to send an Error packet on io::{Read, Write} errors? Hi @PurpleMyst! Thank you! 🙂 To clarify, do you want to send an Error packet on io::{Read, Write} errors in the given reader/writer? No, we want to send the packet over the UDP socket, not the Read or Write objects. Depending on whether we're talking about get or put: The Read object is the data we are sending over the socket. So we read from the Read object, create a packet, and send it. The Write object is the place where we store the data that we receive. So we read from the socket, take the data out of the packet, and use the Write object to save it. Right now, get will attempt to parse an Error packet if it failed to parse a Data packet it's expecting. If parsing the Error packet fails, then it returns early with an error. Likewise, put will attempt to parse an Error packet if it doesn't parse the Ack packet it's expecting. If parsing the Error packet fails, then it returns early with an error. For this issue, I don't want it to return early without sending an Error packet to its peer. That is: IF we fail to parse the packet we're expecting AND we fail to parse an Error packet, that means we've received a packet that doesn't make sense and the connection should not continue. In this case, we will create an Error packet, send it to the peer, and then return an error. Does that help any? Understood. So my PR is totally unrelated to this issue, but does fix another issue. Should I tack on this change to my already-existing PR or create another one? A separate PR would be best. Thanks! 🙂 Alright, I've unlinked #41 from this issue and created a new, linked, pull request numbered #43 The two are independent from eachother, so you can pick and choose when and if to merge them :)
gharchive/issue
2020-09-22T01:22:34
2025-04-01T04:33:52.431697
{ "authors": [ "PurpleMyst", "connorkuehl" ], "repo": "connorkuehl/tftp", "url": "https://github.com/connorkuehl/tftp/issues/11", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
263602922
update to rasterio1.0 Update clover to use rasterio1.0 Coverage decreased (-0.05%) to 40.079% when pulling d28b06115bbd32db1793be485c6e87a40bc5f3ef on rasterio1.0 into d33fbc4d90121a09652806e259fd651c248f4f1a on master. Coverage decreased (-0.05%) to 40.079% when pulling d5e5ed295908ce950e3c3107e285771d767851ff on rasterio1.0 into d33fbc4d90121a09652806e259fd651c248f4f1a on master.
gharchive/pull-request
2017-10-07T00:24:41
2025-04-01T04:33:52.436289
{ "authors": [ "coveralls", "kennino" ], "repo": "consbio/clover", "url": "https://github.com/consbio/clover/pull/51", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
343458359
submit wft WFchain is a time-sharing wireless network value sharing ecosystem, and its main chain named WFchain is developed based on Ethernet Intelligent contract ERC2.0.The aim is to establish a centralization, collective governance, rule driven ecological trust system, and to achieve trustworthy and non tamperable data sharing on this trust system.The role of participation in the ecosystem: Hot users, sharing customers, advertisers can get reasonable incentives and rewards. 你好,请向我们提供更多项目信息,以便我们更好的了解你们的项目,例如主流区块链媒体相关报道,谢谢。
gharchive/pull-request
2018-07-23T01:20:45
2025-04-01T04:33:52.438042
{ "authors": [ "wenyanwang", "wifikim" ], "repo": "consenlabs/token-profile", "url": "https://github.com/consenlabs/token-profile/pull/1264", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
615188624
Add DMMM token Project background: Official website: https://d-mmm.github.io Official announcement: https://d-mmm.github.io/announcements/token-info-for-imtoken.md Team background: We are an international team of tech geeks from Russia, China and India. Project basic information: The dMMM game is the first classic dApp on the dMMM ecosystem. In the dMMM game, players can participate in the game by paying tokens and earn income when they wait in line to end the game. In the process of continuously participating in the game, the player account can be upgraded to obtain more interesting and tangible benefits or rewards. Recommended Gas Limit for transaction: 60000 请在官网添加跳转至官方公告渠道的链接。 请在官网添加跳转至官方公告渠道的链接。 Here is the Google Translate result of your reply content: Please add a link to the official announcement channel on the official website. If it is correct, the work had been done before. Please see the bottom of the official site Your website was https://dmmm.io Your website was https://dmmm.io Thanks for your reply, but this is the dApp link, not the OFFICIAL WEBSITE. The dApp link was designed for users to interact with dApp in crypto wallet and it's not suitable for adding some other links there, like announcemet or something else, or it will look wired. As I mentioned before, our official website is: https://d-mmm.github.io Yours. so you should replace the website address with "https://d-mmm.github.io",in the json file so you should replace the website address with "https://d-mmm.github.io",in the json file Thank you for your kind reply, I've fixed this problem. Welcome to recommend imToken 2.0 to your token holders to manage their token.
gharchive/pull-request
2020-05-09T14:33:11
2025-04-01T04:33:52.445894
{ "authors": [ "d-mmm", "sfsy520" ], "repo": "consenlabs/token-profile", "url": "https://github.com/consenlabs/token-profile/pull/5743", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
758246803
add wozx token 团队背景: Efforce的创始人是赫赫有名的,与乔布斯携手创办苹果公司的教父级人物沃兹尼亚克;CEO是拥有哈佛大学经济学学历的 Jacopo Visetti,项目其他主要成员背景多样化,包括有精通能源领域金融衍生品的 Maurizio Casetelli,2016 年被意大利总统授予创新奖的编程高手 Jacopo Vanetti,熟悉亚洲市场背景的 Anyu Chen 等。 项目基本情况: EFFORCE 是基于区块链技术的节能量交易平台。这个平台将希望提高其建筑物或工业设施的能源效率的企业,以及希望通过以代币形式获得能效改造项目所产出的节能量的买家汇集到一起。 更多项目信息 官网:https://www.efforce.io/ Medium:https://medium.com/efforce Telegram:https://t.me/EFFORCE_ICO Twitter:https://twitter.com/efforceofficial 微博:https://weibo.com/7510957207 媒体报道 金色财经 一文读懂由前苹果公司创始人沃兹创办的Efforce项目 https://www.jinse.com/news/blockchain/860838.html 收录的交易所 霍比特(HBTC) 代币转账调用智能合约需要的 Gas limit 60000 认证的社交媒体渠道发布公告: https://weibo.com/7510957207/JxmTiC6um 欢迎在社区中推荐用户使用 imToken 2.0 ,imToken 支持比特币, 以太坊, Comos, EOS, 波场, Nervos, 比特币现金, 莱特币,波卡等资产管理;同时支持去中心化币币兑换功能 Tokenlon;转账页面支持 ETH, USDT, EOS, BTC OTC 功能;另外对于没有 ETH 矿工费的用户,也提供 ETH 手续费充值功能。
gharchive/pull-request
2020-12-07T07:30:34
2025-04-01T04:33:52.450289
{ "authors": [ "lkbtboy", "sfsy520" ], "repo": "consenlabs/token-profile", "url": "https://github.com/consenlabs/token-profile/pull/6552", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2081259670
🛑 tasp.resterdigne.net is down In dd50764, tasp.resterdigne.net (https://tasp.resterdigne.net) was down: HTTP code: 0 Response time: 0 ms Resolved: tasp.resterdigne.net is back up in 20ee766.
gharchive/issue
2024-01-15T05:06:35
2025-04-01T04:33:52.473458
{ "authors": [ "trivoallan" ], "repo": "constructions-incongrues/status", "url": "https://github.com/constructions-incongrues/status/issues/1939", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1454446877
🛑 computertruck.parishq.net is down In 70a59f0, computertruck.parishq.net (https://computertruck.parishq.net) was down: HTTP code: 0 Response time: 0 ms Resolved: computertruck.parishq.net is back up in 8426f74.
gharchive/issue
2022-11-18T04:59:03
2025-04-01T04:33:52.476500
{ "authors": [ "trivoallan" ], "repo": "constructions-incongrues/status", "url": "https://github.com/constructions-incongrues/status/issues/285", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2303849564
🛑 traefik.pastis-hosting.net is down In 7a3d2fa, traefik.pastis-hosting.net (https://traefik.pastis-hosting.net) was down: HTTP code: 0 Response time: 0 ms Resolved: traefik.pastis-hosting.net is back up in a4d02f7.
gharchive/issue
2024-05-18T04:48:18
2025-04-01T04:33:52.479691
{ "authors": [ "trivoallan" ], "repo": "constructions-incongrues/status", "url": "https://github.com/constructions-incongrues/status/issues/3785", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2421717916
🛑 thisisradioclash.org is down In 6241e33, thisisradioclash.org (https://thisisradioclash.org) was down: HTTP code: 0 Response time: 0 ms Resolved: thisisradioclash.org is back up in 92d63b9 after 5 hours, 27 minutes.
gharchive/issue
2024-07-22T01:26:47
2025-04-01T04:33:52.482667
{ "authors": [ "trivoallan" ], "repo": "constructions-incongrues/status", "url": "https://github.com/constructions-incongrues/status/issues/5614", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2472023354
🛑 Ouïedire is down In 9e13872, Ouïedire (https://www.ouiedire.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: Ouïedire is back up in 2879464 after 9 minutes.
gharchive/issue
2024-08-18T15:23:29
2025-04-01T04:33:52.484986
{ "authors": [ "trivoallan" ], "repo": "constructions-incongrues/status", "url": "https://github.com/constructions-incongrues/status/issues/5858", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1212302824
Allow to run tests in parallel This commit makes it possible to: tox -e test -- -n=N Signed-off-by: Chenxiong Qi cqi@redhat.com Maintainers will complete the following section [x] Commit messages are descriptive enough n/a Code coverage from testing does not decrease and new code is covered n/a Python type annotations added to new code n/a JSON/YAML configuration changes are updated in the relevant schema n/a Changes to metadata also update the documentation for the metadata n/a Pull request has a link to an osbs-docs PR for user documentation updates n/a New feature can be disabled from a configuration file Shall we run them in parallel also in CI?
gharchive/pull-request
2022-04-22T12:29:37
2025-04-01T04:33:52.499838
{ "authors": [ "MartinBasti", "tkdchen" ], "repo": "containerbuildsystem/atomic-reactor", "url": "https://github.com/containerbuildsystem/atomic-reactor/pull/1898", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
452043933
containerd rpm build on RHEL s390x: Getting error Trying to build containerd rpm on RHEL s390x. Thought able to build binaries, but blocked to build the rpm. The spec file to build rpm is not available. Used https://src.fedoraproject.org/rpms/containerd/blob/master/f/containerd.spec file to build rpm. But getting below error. Any immediate response would help a lot. Followed below steps to build the rpm mkdir -p ~/rpmbuild/{BUILD,BUILDROOT,RPMS,SOURCES,SPECS,SRPMS} copied the source tarball `wget https://github.com/containerd/containerd/archive/v1.2.4.tar.gz` under SOURCES dir and https://src.fedoraproject.org/rpms/containerd/blob/master/f/containerd.spec file under SPECS dir. Tried building [root@hurlnxa0 rpmbuild]# rpmbuild -ba SPECS/containerd.spec Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.XNx7jY + umask 022 + cd /root/rpmbuild/BUILD + cd /root/rpmbuild/BUILD + rm -rf containerd-1.2.4 + /usr/bin/gzip -dc /root/rpmbuild/SOURCES/containerd-1.2.4.tar.gz + /usr/bin/tar -xf - + STATUS=0 + '[' 0 -ne 0 ']' + cd containerd-1.2.4 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.MPP8vK + umask 022 + cd /root/rpmbuild/BUILD + cd containerd-1.2.4 + GO_BUILD_PATH=/root/rpmbuild/BUILD/containerd-1.2.4/_build ++ dirname /root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd/containerd + install -m 0755 -vd /root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd install: creating directory '/root/rpmbuild/BUILD/containerd-1.2.4/_build' install: creating directory '/root/rpmbuild/BUILD/containerd-1.2.4/_build/src' install: creating directory '/root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com' install: creating directory '/root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd' + ln -fs /root/rpmbuild/BUILD/containerd-1.2.4 /root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd/containerd + cd /root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd/containerd + install -m 0755 -vd _bin install: creating directory '_bin' + export PATH=/root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd/containerd/_bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/root/bin:/var/go/bin:/usr/local/bin/go:/var/go/bin:/usr/local/go/bin/:/root/rpmbuild/bin + PATH=/root/rpmbuild/BUILD/containerd-1.2.4/_build/src/github.com/containerd/containerd/_bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/root/bin:/var/go/bin:/usr/local/bin/go:/var/go/bin:/usr/local/go/bin/:/root/rpmbuild/bin + export 'GOPATH=/root/rpmbuild/BUILD/containerd-1.2.4/_build:%{gopath}' + GOPATH='/root/rpmbuild/BUILD/containerd-1.2.4/_build:%{gopath}' + export 'LDFLAGS=-X github.com/containerd/containerd/version.Version=1.2.4' + LDFLAGS='-X github.com/containerd/containerd/version.Version=1.2.4' ++ head -c20 /dev/urandom ++ od -An -tx1 ++ tr -d ' \n' + go build -buildmode pie -compiler gc '-tags=rpm_crashtraceback ' -ldflags '-X github.com/containerd/containerd/version.Version=1.2.4 -B 0xea4426d1f0b196d3664bfdb83f1965030451334b -extldflags '\''-Wl,-z,relro '\''' -a -v -x -o _bin/containerd github.com/containerd/containerd/cmd/containerd go: GOPATH entry is relative; must be absolute path: "%{gopath}". For more details see: 'go help gopath' error: Bad exit status from /var/tmp/rpm-tmp.MPP8vK (%build) RPM build errors: Bad exit status from /var/tmp/rpm-tmp.MPP8vK (%build) [root@hurlnxa0 rpmbuild]# The error is right in your output: go: GOPATH entry is relative; must be absolute path: "%{gopath}". For more details see: 'go help gopath' You just need to fix up your gopath You can build using make from the repo or download the binaries off of the releases page if you need something.
gharchive/issue
2019-06-04T14:52:37
2025-04-01T04:33:52.512229
{ "authors": [ "AnitaNayak", "crosbymichael" ], "repo": "containerd/containerd", "url": "https://github.com/containerd/containerd/issues/3320", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
605589297
Allow more time for containerd to start for integration Test hardcoded 2 seconds and in GitHub actions we see a cancel randomly right at 2 seconds even though containerd is within milliseconds of being ready. Example: FAIL github.com/containerd/containerd 2.043s Signed-off-by: Phil Estes estesp@linux.vnet.ibm.com Codecov Report Merging #4208 into master will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #4208 +/- ## ======================================= Coverage 38.34% 38.34% ======================================= Files 90 90 Lines 12728 12728 ======================================= Hits 4881 4881 Misses 7181 7181 Partials 666 666 Flag Coverage Δ #windows 38.34% <ø> (ø) Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update e094d36...74be7d4. Read the comment docs. https://www.youtube.com/watch?v=JMfxQq1GCMg LGTM!
gharchive/pull-request
2020-04-23T14:32:16
2025-04-01T04:33:52.519516
{ "authors": [ "codecov-io", "dims", "estesp" ], "repo": "containerd/containerd", "url": "https://github.com/containerd/containerd/pull/4208", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2479656402
Rename rollback on error This comes from #3341 Failure to rename will leave the system in a broken state. This PR implements a rollback mechanism to revert whatever we can. This is not a solution for the problems linked to #3341, but at least it will stop bricking systems (hopefully...). Needs rebase Rebased. Pending CI.
gharchive/pull-request
2024-08-22T02:53:41
2025-04-01T04:33:52.521209
{ "authors": [ "AkihiroSuda", "apostasie" ], "repo": "containerd/nerdctl", "url": "https://github.com/containerd/nerdctl/pull/3342", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2761073024
Fix image inspect error on dnf Issue Fixes regression inadvertently introduced with https://github.com/containerd/nerdctl/pull/3017 for image inspect command exit code when image does not exist. Description This change updates the image inspect utility function to serialize an array of encountered errors. This aligns with the behavior of nerdctl inspect. Testing Adds simple test case for image inspect on images that do not exist. Required updates to test cases which assumed exit code 0 for image inspects on images that do not exist. What is "dnf" here? Doesn't seem related to Fedora's dnf @AkihiroSuda, oops sorry about that. I had shorthanded "did not find" not meaning to confuse dnf package manager. Lastest commit should have more clear wording.
gharchive/pull-request
2024-12-27T16:35:47
2025-04-01T04:33:52.524083
{ "authors": [ "AkihiroSuda", "austinvazquez" ], "repo": "containerd/nerdctl", "url": "https://github.com/containerd/nerdctl/pull/3792", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2721853011
benches: add a new wasi-demo-{app, oci} benchmark This commit adds a new benchmark for running the shims end-to-end on wasi-demo-{app, oci} to print 'hello'. /lgtm It doesn't seem like GitHub has recognized this as an approval, but I will see it as one!
gharchive/pull-request
2024-12-06T01:47:22
2025-04-01T04:33:52.525554
{ "authors": [ "Mossaka" ], "repo": "containerd/runwasi", "url": "https://github.com/containerd/runwasi/pull/764", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
411204505
Add 3rd chunk of Cobra examples Signed-off-by: TomSweeneyRedHat tsweeney@redhat.com Adds the third chunk of Cobra examples to the cli help. As we're putting together a release tomorrow, tried to hit the heavy commands with this PR. space are going to be the death of me yet. I think I've all the gitvalidate/space issues cleared. LGTM /lgtm
gharchive/pull-request
2019-02-17T16:16:06
2025-04-01T04:33:52.551375
{ "authors": [ "TomSweeneyRedHat", "mheon", "rhatdan" ], "repo": "containers/libpod", "url": "https://github.com/containers/libpod/pull/2357", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
460413709
Set correct SELinux label on restored containers Restoring a container from a checkpoint archive mounted the container's root file-system with a new SELinux mountlabel which resulted in a container running with a different label then the root file-system. This tells the storage setup to use the mountlabel which is stored in the checkpoint archive. The easiest way to reproduce this bug is to run 'podman exec' on a restored container. This also includes a new checkpoint/restore test case doing a 'podman exec' after 'podman container restore'. Unfortunately the test does not trigger the error as the tests running under ginkgo are not under the same SELinux confinement as the actual containers. At least that is how understand it. /approve @adrianreber please run make validate locally and --amend -s Can this be merged? I have a two open PRs which depend on this and two more locally I am waiting to publish based on this. /lgtm
gharchive/pull-request
2019-06-25T13:07:09
2025-04-01T04:33:52.553475
{ "authors": [ "adrianreber", "baude", "rhatdan" ], "repo": "containers/libpod", "url": "https://github.com/containers/libpod/pull/3425", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
637289378
[CI:DOCS] Fix carriage returns in API v2 header ReadTheDocs was wrapping lines and removing leading spaces unless there a blank line in between. This adds the blank lines to make the examples more readable on https://docs.podman.io/en/latest/_static/api.html Signed-off-by: TomSweeneyRedHat tsweeney@redhat.com LGTM, tests red though Well I didn't expect that @mheon. I'll try retesting, if that fails, any thoughts @baude? /lgtm
gharchive/pull-request
2020-06-11T20:05:59
2025-04-01T04:33:52.555772
{ "authors": [ "TomSweeneyRedHat", "mheon", "rhatdan" ], "repo": "containers/libpod", "url": "https://github.com/containers/libpod/pull/6581", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2185726817
bug: podman_images_* series include metrics for non-existing images Hello, I've created a dashboard to monitor amount of images on the podman host and noticed that the metrics dont line up with the host. There is a lot of <none> images that do not show up in cli when running podman images Where are these coming from ? (I've tried to run podman system prune but that made no difference as there were no dangling images on the system) root@podman1:~>podman system prune WARNING! This command removes: - all stopped containers - all networks not used by at least one container - all dangling images - all dangling build cache Are you sure you want to continue? [y/N] y Total reclaimed space: 0B root@podman1:~> podman images -> 14 items root@podman1:~>podman images REPOSITORY TAG IMAGE ID CREATED SIZE refinst-docker-dev-local.artifactory.io/grafana-agent1 latest eabd977236ca 2 days ago 457 MB refinst-docker-dev-local.artifactory.io/grafana_agent_latest latest ea71004b1aaf 2 days ago 457 MB refinst-docker-dev-local.artifactory.io/vmware-exporter latest 5e883f1bedfc 5 days ago 937 MB refinst-docker-dev-local.artifactory.io/veeam-em-exporter latest e09053d9feab 2 weeks ago 883 MB refinst-docker-dev-local.artifactory.io/hpilo-exporter latest c81f066fce6d 2 weeks ago 882 MB refinst-docker-dev-local.artifactory.io/python39 latest c801b1ebfa4b 2 weeks ago 880 MB refinst-docker-dev-local.artifactory.io/web_discard latest 516038f49add 2 weeks ago 404 MB refinst-docker-dev-local.artifactory.io/netapp-harvest-exporter latest 50b147ce0c4b 2 weeks ago 316 MB refinst-docker-dev-local.artifactory.io/web_dashboard latest 24dd88e557b0 5 weeks ago 319 MB refinst-docker-dev-local.artifactory.io/web_tq latest 9d2260fd7289 6 weeks ago 413 MB refinst-docker-dev-local.artifactory.io/web_gapcheck latest 7a64a7f08f58 6 weeks ago 413 MB refinst-docker-dev-local.artifactory.io/web_rds latest 3a07e63ee435 7 weeks ago 414 MB refinst-docker-dev-local.artifactory.io/web_rdf latest 8a5e4434c4b0 8 weeks ago 416 MB localhost/podman-pause 4.6.1-1692961697 4ce25834cda0 4 weeks ago 810 kB root@podman1:~> podman_image_info -> 38 items root@podman1:~>curl -s http://127.0.0.1:9882/metrics | grep -i "podman_image_info" # HELP podman_image_info Image information. # TYPE podman_image_info gauge podman_image_info{digest="sha256:0a6ed5c7ac19ac85ea3408247032e36590c97fcf9b92202131dde940994cbc9e",id="2905c0f74e2b",parent_id="ec7bf4cae86f",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:0aca56789602469679f03cd9a74abcc5cfb23e505f6a9dade9dac872fc7bf9d3",id="4ce25834cda0",parent_id="",repository="localhost/podman-pause",tag="4.6.1-1692961697"} 1 podman_image_info{digest="sha256:0c033e7a4a25b4bb3857aae3a9985eb6cc51674fa6841c5d6ab6e520b744037c",id="6f7fa9a134b1",parent_id="897420c14ff9",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:113446045e5beb631735d4b78606865b417d43aec3f5d0e10cce7138da10fe20",id="7a64a7f08f58",parent_id="",repository="refinst-docker-dev-local.artifactory.io/web_gapcheck",tag="latest"} 1 podman_image_info{digest="sha256:1673345dc3f19269ae8c87c337e6d0bbfc42ab3596abbfc2635fc5a33d8a1987",id="9a2807da1e7b",parent_id="ef5d4631a596",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:33442ed707926c7a5f95cee14ef921c79e01f6f0b6faa8c52821b4f89b9c1040",id="6d03d1c6deee",parent_id="ed57561df995",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:33c34a64e8b27039fbc546580395ef509f8943b55a2d4382c9efc75654bc2e1c",id="ed57561df995",parent_id="a207e1233108",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:3b76c4fa1298ce14181b89e09f336db94241698ef03571bda82618fd5ca33794",id="24dd88e557b0",parent_id="",repository="refinst-docker-dev-local.artifactory.io/web_dashboard",tag="latest"} 1 podman_image_info{digest="sha256:3fd696005267537e7901cea0e49d9498cea6889fc0ee17e5ea90be418d6f09b8",id="ec7bf4cae86f",parent_id="9a2807da1e7b",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:488e9d0d8eed2abacbc8e9aea9355e87e65f6a0c51acb387cf8994249202d741",id="a207e1233108",parent_id="",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:48bd1c5e531374e5763509cc21d0973ace66b43b0278a4f32645eb56264ee22c",id="8a5e4434c4b0",parent_id="",repository="refinst-docker-dev-local.artifactory.io/web_rdf",tag="latest"} 1 podman_image_info{digest="sha256:4fbe71d302dfcdd052541087701895dd0897cc05fa5f94d112efdad50a138021",id="50b147ce0c4b",parent_id="6f29191cc431",repository="refinst-docker-dev-local.artifactory.io/netapp-harvest-exporter",tag="latest"} 1 podman_image_info{digest="sha256:656bc3649d5854472637fdf47a46c56234b8bb11ad6ff2296349a61b360ce0a3",id="ea71004b1aaf",parent_id="",repository="refinst-docker-dev-local.artifactory.io/grafana_agent_latest",tag="latest"} 1 podman_image_info{digest="sha256:680500d57432c0d0efb312f7b4419b74e38db15ce54a85677beec73107d17432",id="516038f49add",parent_id="06e50700f88a",repository="refinst-docker-dev-local.artifactory.io/web_discard",tag="latest"} 1 podman_image_info{digest="sha256:7016d3a807af13f15cf6d06e06e7ca0c6457eabe795ad48ed8a074a7e438142f",id="87131e84e92f",parent_id="32c82d478acf",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:72fbd7c8d40fba25881926da0320c928bc6d96fa2e4ecf82c55c2e98948049e5",id="75547269d5dd",parent_id="c801b1ebfa4b",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:7435f71b84f40e8e11a79a00ef6ddf5b3a87fac4831c53b827d404e77eccf9f7",id="3a07e63ee435",parent_id="",repository="refinst-docker-dev-local.artifactory.io/web_rds",tag="latest"} 1 podman_image_info{digest="sha256:84fb87a5fdf985050f71d9bdea1478ee38216efc5ce575a543ec33fa54a82e48",id="a1534bf3bdb4",parent_id="2905c0f74e2b",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:8ea031fa8ae8966d5c98da5befd753adde29e6d456db4159302a8829da9a172f",id="32c82d478acf",parent_id="b4dcaebd7ad6",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:8ffa2a6e92204235f7c352c84101c1cba46cbbce6441c99d2abbf8cbd289c031",id="ef5d4631a596",parent_id="c801b1ebfa4b",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:92902126a72240481acc03525b8ead478a5a73750c29bf866d46f2e79a6b83af",id="c79f51b7c547",parent_id="5a7233b9a946",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:95bfa65e3fc9b2537d59f701edd5dc02db0fe4238e2ff2b68ff7bfe4889a516c",id="513fe78ae290",parent_id="ea71004b1aaf",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:986d1e26e6aed7df4ede6b236cc7fb6f3ac6f83bae752cfd13a0315f89fb08c1",id="c801b1ebfa4b",parent_id="",repository="refinst-docker-dev-local.artifactory.io/python39",tag="latest"} 1 podman_image_info{digest="sha256:996de721756a9656eda92bf31adadeedfec4246ecb4747accf301cba97e43806",id="d87698835e56",parent_id="c79f51b7c547",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:a08b74bfd4dee20fede7459800d26de0e65bf9ea6ad328083e600e3b1731ff9b",id="6f29191cc431",parent_id="87131e84e92f",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:a5f22f326e8c3376014ccbd0a3e81edda656f9e603e0367b25a40470e3e1e4d2",id="5a7233b9a946",parent_id="5d2e027dce4f",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:aad54800aa00ff2837730b51be85d00da21edb40dcd0f5e41c5a6d4334fffa65",id="55cb705da813",parent_id="c801b1ebfa4b",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:aed453013c0f102e953bacebb05dfa3d1355947f572665813834d439b5d67dc3",id="9d2260fd7289",parent_id="",repository="refinst-docker-dev-local.artifactory.io/web_tq",tag="latest"} 1 podman_image_info{digest="sha256:bce8cc01a18804d07163078d78febfe89144c3a6501900ad04e2a79473ac0b1a",id="897420c14ff9",parent_id="75547269d5dd",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:c364e2faffae61d860b23651baad3d73586c58a779387305cfe6e2af1c609f1d",id="06e50700f88a",parent_id="6d03d1c6deee",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:d0192458dbbbf72f0fe9f59cb0c98fc46cc3573918e866a0331d3177429a448f",id="c81f066fce6d",parent_id="bf4fc3d5cec8",repository="refinst-docker-dev-local.artifactory.io/hpilo-exporter",tag="latest"} 1 podman_image_info{digest="sha256:d601b23ecef1ceafefcfe2f122858aebb7d2873586d47fc9afd0fe5af67c80fd",id="bf4fc3d5cec8",parent_id="a1534bf3bdb4",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:e03b62dae1dc0163b6b9c90e1defe4d3f4a1808306ca133d8bdb8ee5aad58c9f",id="b4dcaebd7ad6",parent_id="4399df180324",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:e05ca2be96483cfce3ebc425ce478d24a4da427b065d744d1e9ff5a038861d46",id="5d2e027dce4f",parent_id="55cb705da813",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:e7f4b3705e8575ddd73bb326cb756686b1e874ba2257bb01178ba9d1d29a69ce",id="e09053d9feab",parent_id="6f7fa9a134b1",repository="refinst-docker-dev-local.artifactory.io/veeam-em-exporter",tag="latest"} 1 podman_image_info{digest="sha256:f80f755d353c7367cabf625b1b7723bfe8f850dbab3f72c8ee3b07d037ad8c53",id="4399df180324",parent_id="",repository="<none>",tag="<none>"} 1 podman_image_info{digest="sha256:f81453457f17aff712282c95d5c6ddb6407f487ae94f5c5c4b281291b6f2e055",id="5e883f1bedfc",parent_id="d87698835e56",repository="refinst-docker-dev-local.artifactory.io/vmware-exporter",tag="latest"} 1 podman_image_info{digest="sha256:fff48b6cab26df80408134707a6d8fcf10227c5dc6cf3c59755201f383985b4e",id="eabd977236ca",parent_id="513fe78ae290",repository="refinst-docker-dev-local.artifactory.io/grafana-agent1",tag="latest"} 1 root@podman1:~> Hi @wally007 can you also send me podman image ls --all command output. Regards Hi @wally007 can you also send me podman image ls --all command output. Regards Yes, these show the "missing" images root@podman1:~>podman images --all REPOSITORY TAG IMAGE ID CREATED SIZE refinst-docker-dev-local.artifactory.io/grafana-agent1 latest eabd977236ca 3 days ago 457 MB <none> <none> 513fe78ae290 3 days ago 457 MB refinst-docker-dev-local.artifactory.io/grafana_agent_latest latest ea71004b1aaf 3 days ago 457 MB refinst-docker-dev-local.artifactory.io/vmware-exporter latest 5e883f1bedfc 5 days ago 937 MB <none> <none> d87698835e56 5 days ago 937 MB <none> <none> c79f51b7c547 5 days ago 937 MB <none> <none> 5a7233b9a946 5 days ago 880 MB <none> <none> 5d2e027dce4f 2 weeks ago 880 MB <none> <none> 55cb705da813 2 weeks ago 880 MB refinst-docker-dev-local.artifactory.io/veeam-em-exporter latest e09053d9feab 2 weeks ago 883 MB <none> <none> 6f7fa9a134b1 2 weeks ago 883 MB <none> <none> 897420c14ff9 2 weeks ago 883 MB <none> <none> 75547269d5dd 2 weeks ago 880 MB refinst-docker-dev-local.artifactory.io/hpilo-exporter latest c81f066fce6d 2 weeks ago 882 MB <none> <none> bf4fc3d5cec8 2 weeks ago 882 MB <none> <none> a1534bf3bdb4 2 weeks ago 882 MB <none> <none> 2905c0f74e2b 2 weeks ago 880 MB <none> <none> ec7bf4cae86f 2 weeks ago 880 MB <none> <none> 9a2807da1e7b 2 weeks ago 880 MB <none> <none> ef5d4631a596 2 weeks ago 880 MB refinst-docker-dev-local.artifactory.io/python39 latest c801b1ebfa4b 2 weeks ago 880 MB refinst-docker-dev-local.artifactory.io/web_discard latest 516038f49add 2 weeks ago 404 MB <none> <none> 06e50700f88a 2 weeks ago 404 MB <none> <none> 6d03d1c6deee 2 weeks ago 404 MB <none> <none> ed57561df995 2 weeks ago 404 MB <none> <none> a207e1233108 2 weeks ago 404 MB refinst-docker-dev-local.artifactory.io/netapp-harvest-exporter latest 50b147ce0c4b 2 weeks ago 316 MB <none> <none> 6f29191cc431 3 weeks ago 316 MB <none> <none> 87131e84e92f 3 weeks ago 316 MB <none> <none> 32c82d478acf 3 weeks ago 316 MB <none> <none> b4dcaebd7ad6 3 weeks ago 316 MB <none> <none> 4399df180324 3 weeks ago 316 MB localhost/podman-pause 4.6.1-1692961697 4ce25834cda0 4 weeks ago 810 kB refinst-docker-dev-local.artifactory.io/web_dashboard latest 24dd88e557b0 6 weeks ago 319 MB refinst-docker-dev-local.artifactory.io/web_tq latest 9d2260fd7289 6 weeks ago 413 MB refinst-docker-dev-local.artifactory.io/web_gapcheck latest 7a64a7f08f58 6 weeks ago 413 MB refinst-docker-dev-local.artifactory.io/web_rds latest 3a07e63ee435 7 weeks ago 414 MB refinst-docker-dev-local.artifactory.io/web_rdf latest 8a5e4434c4b0 8 weeks ago 416 MB root@podman1:~> What are these image ? How come these are not pruned with podman system prune ? We prune images on schedule in our dev env. and do not want any dangling images - yet it seems there are still some left.
gharchive/issue
2024-03-14T08:21:07
2025-04-01T04:33:52.626053
{ "authors": [ "navidys", "wally007" ], "repo": "containers/prometheus-podman-exporter", "url": "https://github.com/containers/prometheus-podman-exporter/issues/207", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1636111537
Using typescript instead of javascript Docusaurus is written in typescript and provides first-class support: https://docusaurus.io/docs/typescript-support It would be beneficial for us to use typescript throughout the project / no javascript as other teams are familiar with typescript implementations. Will tag #15 to this, which should be good to merge tomorrow.
gharchive/issue
2023-03-22T16:23:37
2025-04-01T04:33:52.642313
{ "authors": [ "cdrage", "lilyx13" ], "repo": "containers/website-new", "url": "https://github.com/containers/website-new/issues/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
519872839
Run in single namespace What does this PR do This PR makes Maesh able to run in a single namespace. Fixes: #253 How to test this Install maesh in the same namespace of your applications, and enjoy your shiny service mesh as usual :tada: How does it work Now, all generated mesh services are labeled with app=maesh too. Then the controller is now able to a list of apps, right now maesh and jaeger are hardcoded, and we might want to change that. But it makes maesh ignore the maesh services. For the initial createMaeshService I used fields selectors instead of filtering everything in memory. I also made the filtering more time efficient using a map instead of iterating over the services, as we can have a lot of services this can be important. Additional information I kind of hacked my way to get it working, that is definitely not a final implementation. Also It doesn't support CNI just yet. Draft for the moment as CI is super broken and it is not finished. I ❤️ Where this is going. Design LGTM. Switching to review. This is looking really good. I look forwards to working on a follow up to expand this!
gharchive/pull-request
2019-11-08T08:13:03
2025-04-01T04:33:52.647212
{ "authors": [ "dtomcej", "jlevesy" ], "repo": "containous/maesh", "url": "https://github.com/containous/maesh/pull/348", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
283312816
Traefik 500 errors and zero length http responses Do you want to request a feature or report a bug? Bug What did you do? One of the backends in my traefik deployment is an Ubuntu package mirror. Traefik and the mirror are running in a docker swarm. What did you expect to see? An apt-get update should iterate through each provider and collect package information. What did you see instead? apt-get sporadically prints 500 Internal server error when crawling the package index. I tried running the mirror without Traefik, and with Traefik:v1.4.5 and both worked fine. For the requests that return 500s, the backend doesn't seem to have any logs of the request. Shell output from Ubuntu instance output: # apt-get update Get:1 http://<MIRROR>/ubuntu zesty InRelease [243 kB] Get:2 http://<MIRROR>/ubuntu zesty-updates InRelease [89.2 kB] Get:3 http://<MIRROR>/ubuntu zesty-backports InRelease [89.2 kB] Get:4 http://<MIRROR>/ubuntu zesty-security InRelease [89.2 kB] Get:5 http://<MIRROR>/ubuntu zesty/universe Sources [10.7 MB] Ign:6 http://<MIRROR>/ubuntu zesty/restricted amd64 Packages Ign:7 http://<MIRROR>/ubuntu zesty/multiverse amd64 Packages Ign:8 http://<MIRROR>/ubuntu zesty/universe amd64 Packages Ign:9 http://<MIRROR>/ubuntu zesty/main amd64 Packages Ign:10 http://<MIRROR>/ubuntu zesty-updates/universe Sources Ign:11 http://<MIRROR>/ubuntu zesty-updates/main amd64 Packages Ign:12 http://<MIRROR>/ubuntu zesty-updates/multiverse amd64 Packages Ign:13 http://<MIRROR>/ubuntu zesty-updates/universe amd64 Packages Get:14 http://<MIRROR>/ubuntu zesty-updates/restricted amd64 Packages [3604 B] Ign:15 http://<MIRROR>/ubuntu zesty-backports/universe amd64 Packages Ign:16 http://<MIRROR>/ubuntu zesty-backports/main amd64 Packages Ign:6 http://<MIRROR>/ubuntu zesty/restricted amd64 Packages Ign:7 http://<MIRROR>/ubuntu zesty/multiverse amd64 Packages Ign:8 http://<MIRROR>/ubuntu zesty/universe amd64 Packages Ign:9 http://<MIRROR>/ubuntu zesty/main amd64 Packages Get:10 http://<MIRROR>/ubuntu zesty-updates/universe Sources [85.0 kB] Ign:11 http://<MIRROR>/ubuntu zesty-updates/main amd64 Packages Ign:12 http://<MIRROR>/ubuntu zesty-updates/multiverse amd64 Packages Ign:13 http://<MIRROR>/ubuntu zesty-updates/universe amd64 Packages Ign:15 http://<MIRROR>/ubuntu zesty-backports/universe amd64 Packages Ign:16 http://<MIRROR>/ubuntu zesty-backports/main amd64 Packages Ign:6 http://<MIRROR>/ubuntu zesty/restricted amd64 Packages Ign:7 http://<MIRROR>/ubuntu zesty/multiverse amd64 Packages Get:8 http://<MIRROR>/ubuntu zesty/universe amd64 Packages [8068 kB] Ign:9 http://<MIRROR>/ubuntu zesty/main amd64 Packages Ign:11 http://<MIRROR>/ubuntu zesty-updates/main amd64 Packages Ign:12 http://<MIRROR>/ubuntu zesty-updates/multiverse amd64 Packages Ign:13 http://<MIRROR>/ubuntu zesty-updates/universe amd64 Packages Ign:15 http://<MIRROR>/ubuntu zesty-backports/universe amd64 Packages Ign:16 http://<MIRROR>/ubuntu zesty-backports/main amd64 Packages Err:6 http://<MIRROR>/ubuntu zesty/restricted amd64 Packages 500 Internal Server Error Ign:7 http://<MIRROR>/ubuntu zesty/multiverse amd64 Packages Ign:17 http://<MIRROR>/ubuntu zesty-security/universe Sources Ign:18 http://<MIRROR>/ubuntu zesty-security/universe amd64 Packages Ign:19 http://<MIRROR>/ubuntu zesty-security/main amd64 Packages Ign:20 http://<MIRROR>/ubuntu zesty-security/multiverse amd64 Packages Ign:21 http://<MIRROR>/ubuntu zesty-security/restricted amd64 Packages Ign:9 http://<MIRROR>/ubuntu zesty/main amd64 Packages Err:11 http://<MIRROR>/ubuntu zesty-updates/main amd64 Packages 500 Internal Server Error Ign:12 http://<MIRROR>/ubuntu zesty-updates/multiverse amd64 Packages Ign:13 http://<MIRROR>/ubuntu zesty-updates/universe amd64 Packages Err:15 http://<MIRROR>/ubuntu zesty-backports/universe amd64 Packages 500 Internal Server Error Ign:16 http://<MIRROR>/ubuntu zesty-backports/main amd64 Packages Get:17 http://<MIRROR>/ubuntu zesty-security/universe Sources [35.4 kB] Ign:18 http://<MIRROR>/ubuntu zesty-security/universe amd64 Packages Ign:19 http://<MIRROR>/ubuntu zesty-security/main amd64 Packages Ign:20 http://<MIRROR>/ubuntu zesty-security/multiverse amd64 Packages Get:21 http://<MIRROR>/ubuntu zesty-security/restricted amd64 Packages [3221 B] Ign:18 http://<MIRROR>/ubuntu zesty-security/universe amd64 Packages Ign:19 http://<MIRROR>/ubuntu zesty-security/main amd64 Packages Ign:20 http://<MIRROR>/ubuntu zesty-security/multiverse amd64 Packages Err:18 http://<MIRROR>/ubuntu zesty-security/universe amd64 Packages 500 Internal Server Error Ign:19 http://<MIRROR>/ubuntu zesty-security/main amd64 Packages Ign:20 http://<MIRROR>/ubuntu zesty-security/multiverse amd64 Packages Fetched 19.4 MB in 0s (27.1 MB/s) Reading package lists... Done E: Failed to fetch http://<MIRROR>/ubuntu/dists/zesty/restricted/binary-amd64/Packages 500 Internal Server Error E: Failed to fetch http://<MIRROR>/ubuntu/dists/zesty-updates/main/binary-amd64/Packages 500 Internal Server Error E: Failed to fetch http://<MIRROR>/ubuntu/dists/zesty-backports/universe/binary-amd64/Packages 500 Internal Server Error E: Failed to fetch http://<MIRROR>/ubuntu/dists/zesty-security/universe/binary-amd64/Packages 500 Internal Server Error E: Some index files failed to download. They have been ignored, or old ones used instead. Output of traefik version: Version: v1.5.0-rc2 Codename: cancoillotte Go version: go1.9.2 Built: 2017-12-06_03:07:42PM OS/Arch: linux/amd64 What is your environment & configuration (arguments, toml, provider, platform, ...)? Using docker 17.06-ce in swarm mode Arguments --debug --web --docker.swarmmode --docker.domain=localhost --docker.watch --insecureskipverify --docker.endpoint=unix://var/run/docker.sock -c /traefik.toml -l DEBUG Toml defaultEntryPoints = ["http", "https"] [entryPoints] [entryPoints.http] address = ":80" [entryPoints.https] address = ":443" [entryPoints.https.tls] [[entryPoints.https.tls.certificates]] certFile = "<REMOVED>" keyFile = "<REMOVED>" Docker compose labels labels: - "traefik.port=80" - "traefik.frontend.rule=PathPrefix: <REMOVED>" - "traefik.docker.network=<REMOVED>" If applicable, please paste the log output in debug mode (--debug switch) Traefik's logs don't seem to be of much help, the best clue I can find are lines where it completes a request to a valid URL with a 0 length message: kds_traefik.1.ixhcsuyq884t@<REMOVED> | time="2017-12-19T17:04:59Z" level=debug msg="vulcand/oxy/forward: competed ServeHttp on request" Request="{"Method":"GET","URL":{"Scheme":"http","Opaque":"","User":null,"Host":"<REMOVED>","Path":"","RawPath":"","ForceQuery":false,"RawQuery":"","Fragment":""},"Proto":"HTTP/1.1","ProtoMajor":1,"ProtoMinor":1,"Header":{"Cache-Control":["max-age=0"],"User-Agent":["Debian APT-HTTP/1.3 (1.4.6)"]},"ContentLength":0,"TransferEncoding":null,"Host":"<REMOVED>","Form":null,"PostForm":null,"MultipartForm":null,"Trailer":null,"RemoteAddr":"<REMOVED>","RequestURI":"<REMOVED>/ubuntu/dists/zesty/universe/binary-amd64/Packages.gz","TLS":null}" Do you have a specific configuration on your apt mirror ? I can't reproduce with my local configuration @Juliens The hosted part of the mirror consists of an nginx container that mounts a file tree. Its specific configuration only consists of the docker labels above. Using a raw port or Traefik 1.4.5 doesn't produce the errors. I tried 1.5.0-rc1 this morning and it also produces 500 errors. Here is the output from a single apt-get update, I notice a mismatch in the number of 200 and 404 errors between traefik and nginx, as well as a complete lack of 500 errors from nginx. ... E: Failed to fetch http://<REMOVED>/ubuntu/dists/artful/universe/binary-amd64/Packages 500 Internal Server Error E: Failed to fetch http://<REMOVED>/ubuntu/dists/artful-updates/main/binary-amd64/Packages 500 Internal Server Error E: Failed to fetch http://<REMOVED>/ubuntu/dists/artful-security/universe/source/Sources 500 Internal Server Error E: Some index files failed to download. They have been ignored, or old ones used instead. From Traefik: time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 236651, duration: 4.178873ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 78587, duration: 1.269ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 72190, duration: 1.134896ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 78589, duration: 1.69313ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 11088957, duration: 107.378104ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 86.266µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 159.201µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 107.681µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 88.657µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 108.499µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 92.627µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 95.12µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 87.138µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 97.053µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 2758, duration: 1.395755ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 54.008µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 98.622µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 95.997µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 111.272µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 138.217µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 19338, duration: 1.390618ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 53.384µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 89.036µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 74.685µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 74.187µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 97.288µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 88.486µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 1070580, duration: 8.251336ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 73.428µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 109.913µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 83.265µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 70.759µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 67.075µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 153.04µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 200, Length: 1508, duration: 1.454586ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 79.577µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 126.11µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 104.56µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 160.094µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 107.673µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 164.141µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 158.469µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 132.799µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 104.063µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 108.4µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 404, Length: 169, duration: 1.15217ms" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 78.092µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 250.998µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 109.711µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 134.268µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 118.006µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 129.568µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 186.691µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 94.264µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 188.066µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 121.087µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 500, Length: 21, duration: 123.473µs" time="2017-12-20T18:10:46Z" level=info msg="vulcand/oxy/forward/http: Round trip: http://<REMOVED>:80, code: 404, Length: 169, duration: 727.297µs" From the nginx http server: <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful/InRelease HTTP/1.1" 200 236651 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-updates/InRelease HTTP/1.1" 200 78587 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-backports/InRelease HTTP/1.1" 200 72190 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-security/InRelease HTTP/1.1" 200 78589 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful/universe/source/by-hash/SHA256/1e9c7378d247296ef9a74eec30b0b46ed42eb384022e0c3808d0216f904debfa HTTP/1.1" 200 11088957 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-backports/universe/binary-amd64/by-hash/SHA256/62624ad3e9a38d5effe0dd354689ed24041b8909e5a6b11382ac6e32096cc2fd HTTP/1.1" 200 2758 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-backports/main/binary-amd64/by-hash/SHA256/61907b316172fe3baf1bd777e4b531ff97ec96a15d79b0724493bdc97f0ac89b HTTP/1.1" 200 1470 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-updates/universe/source/Sources.gz HTTP/1.1" 200 19338 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-updates/main/binary-amd64/Packages.gz HTTP/1.1" 200 61268 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful/main/binary-amd64/Packages.xz HTTP/1.1" 200 1070580 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-backports/main/binary-amd64/Packages.xz HTTP/1.1" 200 1508 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful/universe/binary-amd64/Packages HTTP/1.1" 404 169 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-updates/universe/binary-amd64/Packages HTTP/1.1" 404 169 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-security/universe/source/Sources.gz HTTP/1.1" 200 7770 "-" "Debian APT-HTTP/1.3 (1.5.1)" <REMOVED> [20/Dec/2017:18:10:46 +0000] "GET /mirror/ubuntu/dists/artful-security/universe/binary-amd64/Packages HTTP/1.1" 404 169 "-" "Debian APT-HTTP/1.3 (1.5.1)" @ValentinGerlach seems not related because #2771 started in 1.3.0. @ValentinGerlach @sknigh I still can't reproduce, if you have anything to help me reproduce it, it would help. 😉 After some tests, the problem is related to HTTP 1.1 pipelining that is not fully support in golang. I asked for the issue to be reopen here. https://github.com/golang/go/issues/20617 As workaround, you can configure your apt to disable pipelining Acquire::http::Pipeline-Depth 0;
gharchive/issue
2017-12-19T17:36:27
2025-04-01T04:33:52.663822
{ "authors": [ "Juliens", "sknigh" ], "repo": "containous/traefik", "url": "https://github.com/containous/traefik/issues/2593", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
149190654
Use CSS transitions for the taskmanager popup hide and show fade ins/fade outs I want to get rid of jQuery if possible because we don't really need it and I think it's nicer using transitions anyway for that small gimmick. I've added the CSS animation and state handling for console. We also need to toggle classes on the app content, not sure how the console should do that? Register to the displayTaskPopup and hideTaskPopup events. Can you do that? Or tell me where the classes have to be set and what the names are? :D The classes are blur-in and blur-out on #content Implemented in https://github.com/contao/package-manager/commit/c6e925a1070d7b58cd8b0decd3ef7d4f8bffb388
gharchive/issue
2016-04-18T15:47:15
2025-04-01T04:33:52.697653
{ "authors": [ "Toflar", "aschempp" ], "repo": "contao/package-manager", "url": "https://github.com/contao/package-manager/issues/8", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1046830753
FTX: Fix PostAccountLeverage URL /api prefix was missing, quick fix Unfortunately this PR hasn't been approved/merged for several months.
gharchive/pull-request
2021-11-07T19:32:55
2025-04-01T04:33:52.698553
{ "authors": [ "Cypher01" ], "repo": "contek-io/io.contek.invoker", "url": "https://github.com/contek-io/io.contek.invoker/pull/52", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1702782170
Change contentful-management version constraint Had some issues with Dependabot updating the contentful-management in a really weird way, picking a very old version of this gem to fulfill the version constraints. I have not tested this at all extensively, but it seems to work. Hi @alexceder, thanks for the PR. I will take a look at it next week and merge if there are no issues. Cheers
gharchive/pull-request
2023-05-09T21:30:47
2025-04-01T04:33:52.717063
{ "authors": [ "alexceder", "rubydog" ], "repo": "contentful/contentful_model", "url": "https://github.com/contentful/contentful_model/pull/148", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
382665100
Missing documentation On page: https://www.contentful.com/developers/docs/android/tutorials/offline-persistence-with-vault/#generatorjava In the code snippet: // Get instance of Vault Vault vault = Vault.with( ----> context <----, DroidStorySpace.class); // Create instance of CDAClient (use a singleton!) CDAClient client = CDAClient.builder() .setSpace("space-id") .setToken("delivery-api-token") .build(); // Trigger sync vault.requestSync(SyncConfig.builder().setClient(client).build()); What is context as it's not mentioned anywhere else on the page. it looks like it might be a typo of client but that comes after. I'm confused. Hello, the context mentioned there is the Android Context used for creating Vault. Greetings, Mario I came from reading the documentation for the CDA API so found the introduction of this unknown variable confusing. Perhaps you could update the documentation anyway for others? I'd do you a PR but can't find where the source is. Hello, We will be thinking on how to highlight this being only for Android Apps even more, but for now I don't see an imitate action. Greetings, Mario
gharchive/issue
2018-11-20T13:33:04
2025-04-01T04:33:52.724037
{ "authors": [ "davidnewcomb", "mariobodemann" ], "repo": "contentful/generator.java", "url": "https://github.com/contentful/generator.java/issues/15", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2016933014
Authorization Bearer token Validations [X] I believe this is a way to improve. I'll try to join the Continue Discord for questions [X] I'm not able to find an open issue that requests the same enhancement Problem I'm using a remote llama.cpp server but I need to specify an API_KEY token. It would be very useful to secure the access to the endpoint so it can be publish on the internet and accessed from anywhere. I see in: /continuedev/continue/tree/main/server/continuedev/libs/llm/ggml.py line 67: headers["Authorization"] = f"Bearer {self.api_key}" But llamacpp.py do not implement this. I cannot use the GGML provider "openai-aiohttp" as it expects openAI API compatibility. Would you have a suggestion? Would it be possible to make the modification? Solution implement: headers["Authorization"] = f"Bearer {self.api_key}" in llamacpp.py @alfredwallace7 We have a 'headers' property in config.json that you can use for this. For example: { "models": [{ "provider": "llamacpp", "model": "...", "request_options": { "headers": { "Authorization": "Bearer <API_KEY>" } }] } Awesome! Thanks you so much for the tip! Awesome! Thanks you so much for the tip!
gharchive/issue
2023-11-29T16:09:11
2025-04-01T04:33:52.741321
{ "authors": [ "alfredwallace7", "sestinj" ], "repo": "continuedev/continue", "url": "https://github.com/continuedev/continue/issues/643", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2135703601
PostgreSQL Context Provider Adds a Postgres context provider that prints Table schema A few sample rows The context provider needs to be configured to a specific DB, and may be scoped to a specific schema. The user can choose a specific table, or all tables. The all tables case will generate a set of independent schema files, which are the same as querying each table individually. This makes it easy to use an LLM when writing SQL. E.g. migrations or queries, or even for writing database adjacent code such as ORM models. Demo Launch Postgres instance version: '3.9' services: postgres: container_name: cats-postgres image: postgres:16-alpine ports: - 5436:5432 volumes: - ~/apps/postgres:/var/lib/postgresql/data environment: - POSTGRES_PASSWORD=catsarecool - POSTGRES_USER=myuser - POSTGRES_DB=animals Connect and insert some data psql -h localhost -p 5436 -U myuser -d animals -- Table schema for cats CREATE TABLE cats ( id SERIAL PRIMARY KEY, name VARCHAR(100) NOT NULL, breed VARCHAR(100) NOT NULL ); -- Table schema for dogs CREATE TABLE dogs ( id SERIAL PRIMARY KEY, name VARCHAR(100) NOT NULL, breed VARCHAR(100) NOT NULL ); -- Inserting data into the cats table INSERT INTO cats (name, breed) VALUES ('Whiskers', 'Siamese'), ('Felix', 'Maine Coon'), ('Garfield', 'Persian'); -- Inserting data into the dogs table INSERT INTO dogs (name, breed) VALUES ('Buddy', 'Golden Retriever'), ('Max', 'Labrador Retriever'), ('Bella', 'German Shepherd'); Add the config { "name": "postgres", "params": { "host": "localhost", "port": 5436, "user": "myuser", "password": "catsarecool", "database": "animals", "schema": "public" } } See the results The llama.cpp model in use here is deepseek33b Practical use-case with ORM @NinjaPerson24119 this is glorious nothing to change, but do you want to add a quick description to docs/docs/customization/context-providers.md? Your demo here is longer than the others in that file, but if you put it in a tag it could just be copy/pasted Added some docs. Wasn't quite sure what you meant by "tag". I added a link to the PR since the entire description is basically a demo. looks great. thanks again!
gharchive/pull-request
2024-02-15T05:52:52
2025-04-01T04:33:52.747752
{ "authors": [ "NinjaPerson24119", "sestinj" ], "repo": "continuedev/continue", "url": "https://github.com/continuedev/continue/pull/859", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
743844875
Closing the dataset while an async operation is running can end with a segfault const gdal = require('gdal-next'); const ds = gdal.open('sample.tif'); const band = ds.bands.get(1) band.pixels.readAsync(0, 0, 10, 10); ds.close(); This is not expected to work, but the user should be protected against a segfault. @mmomtchev I think the correct behavior would be for ds.close() to throw if work is still ongoing, thoughts? In the next version ds.close() will block until the currently running I/O finishes (it cannot be canceled) and then it will cancel all operations that have not yet started - they will end up rejecting with Dataset closed Frankly, if I had seen this sooner, I would have probably settled for it 😄 But I already implemented a locking hierarchy on top of Brian Revis excellent PtrManager which manages the lifetime of all datasets
gharchive/issue
2020-11-16T13:39:43
2025-04-01T04:33:52.750011
{ "authors": [ "contra", "mmomtchev" ], "repo": "contra/node-gdal-next", "url": "https://github.com/contra/node-gdal-next/issues/38", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
238106458
Shorthand for "BREAKING CHANGE:" Could we add a "BREAK:" shorthand for "BREAKING CHANGE:"? Reason being if I have a long component there's not a whole lot of space left to write a meaningful description for the CHANGELOG. It got pointed out to me on twitter that you put BREAKING CHANGE: in the footer to signify a major version. @koddsson 👍 I haven't actually played with the footer stanza on my OSS projects (although I'd noticed it discussed in the Angular docs, and made sure to faithfully define its behavior). Out of curiosity, what tool are you using for CHANGELOG generation, did using the footer work well? I'm using lerna with the conventional-commits flag. It uses the conventional-commits-cli utility under the hood. I haven't made a breaking change commit yet, but I think it should work :)
gharchive/issue
2017-06-23T10:41:15
2025-04-01T04:33:52.752635
{ "authors": [ "bcoe", "koddsson" ], "repo": "conventional-changelog/conventionalcommits.org", "url": "https://github.com/conventional-changelog/conventionalcommits.org/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2504558606
之前有自动更新,之后每天不自动更新了,手动更新可以,是什么原因呢 如题,盼大神赐教 看一下服务端日志,定时任务是否运行或者报什么错误
gharchive/issue
2024-09-04T07:43:20
2025-04-01T04:33:52.761172
{ "authors": [ "cooderl", "lifangguangnian" ], "repo": "cooderl/wewe-rss", "url": "https://github.com/cooderl/wewe-rss/issues/241", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1243328171
Running dockers with user privilages inside docker-compose. What happened? When I'm running docker-compose with project created with django cookie cutter, i'm getting this warning/security message: You're running the worker with superuser privileges: this is absolutely not recommended! What should've happened instead? I think that all of docker-compose stuff should be run with dedicated user inside it I believe that's in local development only, right? Are those root in the container or root in the host? See also #1410 and #2284 Closing, because of duplication of previous issues.
gharchive/issue
2022-05-20T15:40:56
2025-04-01T04:33:52.768784
{ "authors": [ "browniebroke", "pfk-beta", "xjlin0" ], "repo": "cookiecutter/cookiecutter-django", "url": "https://github.com/cookiecutter/cookiecutter-django/issues/3715", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1313590777
Can't play base64 audio Be sure to search for your issue before opening a new one. Current Behavior Expected Behavior Steps to Reproduce Environment URL attempting to play: Browser: Operating system: jsFiddle example: https://jsfiddle.net/sv5x3ug1 Other Information Would it be possible to provide an example? I've haven't heard of base64 audio before, can you provide some documentation or a specification? When it comes to audio, react-player uses the audio element so if base64 audio isn't supported by the element then there is nothing react-player can do to support the format/codec. Here is the list of supported codecs.
gharchive/issue
2022-07-21T17:38:58
2025-04-01T04:33:52.777772
{ "authors": [ "Taar", "geniux-dev" ], "repo": "cookpete/react-player", "url": "https://github.com/cookpete/react-player/issues/1482", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2487811998
Feat: Added a warning to use a fresh VPS This warning is intended to help beginners and to ensure that a fresh VPS is always used and not a running one with a bunch of stuff on it. Fixes: https://github.com/coollabsio/coolify/discussions/1864 Thank you for the PR!
gharchive/pull-request
2024-08-26T21:22:57
2025-04-01T04:33:52.782174
{ "authors": [ "andrasbacsai", "peaklabs-dev" ], "repo": "coollabsio/documentation-coolify", "url": "https://github.com/coollabsio/documentation-coolify/pull/49", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1364492202
Update poetry to 1.2.0. Poetry v1.1 is unmaintained. https://python-poetry.org/blog/announcing-poetry-1.2.0/#will-11-still-be-maintained-how-long-will-12-be-maintained LGTM, Is there a reason dependabot is not doing this? I ran it yesterday, and it did not resolve dependencies. https://github.com/coopnorge/engineering-docker-images/network/updates/453980414 updater | INFO <job_453980414> Checking if poetry 1.1.15 needs updating proxy | 2022/09/07 05:57:56 [022] GET https://pypi.org:443/simple/poetry/ proxy | 2022/09/07 05:57:56 [022] 200 https://pypi.org:443/simple/poetry/ updater | INFO <job_453980414> Latest version is 1.2.0 and then updater | INFO <job_453980414> No update possible for poetry 1.1.15 I believe because a new version of poetry is using a pre-release version of one of the libraries: https://github.com/coopnorge/engineering-docker-images/pull/296/files#diff-3c409943ba0f8f7f3d4fd450e716ae74569eb3ddef4bd3bdbb1a041dc63f7eb5R93
gharchive/pull-request
2022-09-07T11:03:36
2025-04-01T04:33:52.825582
{ "authors": [ "nikitabarskov" ], "repo": "coopnorge/engineering-docker-images", "url": "https://github.com/coopnorge/engineering-docker-images/pull/296", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1140660864
Auto-populate for PDAs using Methods namespace doesn't work for defined types If a seed is a defined type the Auto-populate PDA's feature is failing which makes sense due to this. I tried to give this a shot but couldn't make it work reliably. Decided to submit the issue in case somebody else is working on the same issue. Fixed in https://github.com/coral-xyz/anchor/pull/2824.
gharchive/issue
2022-02-16T22:38:46
2025-04-01T04:33:52.889942
{ "authors": [ "acheroncrypto", "danmt" ], "repo": "coral-xyz/anchor", "url": "https://github.com/coral-xyz/anchor/issues/1458", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
261103907
Change ConvertAs2_0 to ConvertAs2_1 See https://github.com/coreos/bugs/issues/2171 Missed a lot of thing, not done Ok, should have got them all now. What is the point of renaming these functions? It seems like having the version numbers in the function names has no tangible benefits that I can see. Why shouldn't we just remove them and then never have to deal with this search and replace? I'd be fine with that. I think originally it wasn't clear if we'd support multiple output versions, so it was versioned, then when we bumped to 2.1.0, the function names were missed, leading to this PR. @dgonyeo do you have a opinion on the matter? I would personally be in favor of ripping the band-aid off and just dropping the version numbers altogether. I assumed you'd be providing conversions to multiple Ignition spec versions, but I don't think that's the case. If the intent is to always produce a recent version (e.g. Ignition 2.1.0), I agree with @sdemos about dropping the numbers - vendoring the ct package provides the pinning. I'm in favor of dropping the version numbers. Unless we want to revisit the decision of ct only producing a recent version (as opposed to being able to produce multiple versions), there's no reason for these names to be tied to the Ignition schema version number. @ajeddeloh do you have time to do that with this pr or should I take care of it and open a new one? I'm working on a grub/fat/something in very early boot problem right now, and wont have time for this until that's done (timeboxed on this friday). If you want to pick it up, feel free. work to drop version numbers is in #127. closing this pr.
gharchive/pull-request
2017-09-27T20:16:34
2025-04-01T04:33:52.934635
{ "authors": [ "ajeddeloh", "dghubble", "dgonyeo", "sdemos" ], "repo": "coreos/container-linux-config-transpiler", "url": "https://github.com/coreos/container-linux-config-transpiler/pull/114", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
199314323
Use "tolerate-unready-endpoints" to check etcd pod readiness ref: https://github.com/kubernetes/kubernetes/issues/39207 https://github.com/kubernetes/kubernetes/issues/39363 When etcd starts, it has a bootstrap phase that talks to other peers: Start -> bootstrapping -> running/serving If "bootstrap" phase failed, it is the same as dead because no data is available, no quorum information is known. However, if we start etcd in a Kubernetes Pod, even if it starts running, it doesn't mean etcd is running. It is important for the operator to know the status of the etcd member in this phase. We can only proceed with this member after it's "ready". There is a field in Kubernetes to decouple readiness and endpoints called TolerateUnreadyEndpointsAnnotation = "service.alpha.kubernetes.io/tolerate-unready-endpoints" Making use of this will help us differentiate the "bootstrapping" phase. We cannot use alpha feature (as a general rule). We can start to try it out when it becomes beta. Let's keep our workaround for now. Another use case of this: On self hosted etcd, etcd pods restart could lead to pod endpoints removed from service. This is dangerous and unnecessary. Because etcd pod should restart and recover. It shouldn't remove such endpoints unless pod is deleted. For example, say we have 3 members of etcd cluster, three of them died. In such case, the etcd service will have no endpoints and self hosted kubernetes cluster won't be able to recover itself. However, if service can tolerate such unready pods and don't remove the endpoints, etcd pods will restart and recover itself. The field is still alpha in k8s 1.6 . Some real world experience: Due to some issue on the node, e.g. node pressure or network partition, etcd pod was restarted and endpoint gets removed. This could be better tolerated. this is already done.
gharchive/issue
2017-01-06T23:18:18
2025-04-01T04:33:52.978407
{ "authors": [ "hongchaodeng", "xiang90" ], "repo": "coreos/etcd-operator", "url": "https://github.com/coreos/etcd-operator/issues/622", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
234717905
selfHosted: print warning if not enough nodes to scale [skip ci] fixed lgtm OK. I am gonna test it tomorrow. Tested: time="2017-06-09T16:49:34Z" level=warning msg="can't scale to size (5), only have 3 nodes ([gke-htariq-test-default-pool-0786cc20-3072 gke-htariq-test-default-pool-0786cc20-5bgb gke-htariq-test-default-pool-0786cc20-v1v9])" cluster-name=example pkg=cluster And I can still scale it down and delete it.
gharchive/pull-request
2017-06-09T04:21:38
2025-04-01T04:33:52.980457
{ "authors": [ "hongchaodeng", "xiang90" ], "repo": "coreos/etcd-operator", "url": "https://github.com/coreos/etcd-operator/pull/1184", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
68755420
etcdserver.Cluster may have inconsistent view with store for newly joined member when catching up the progress Restriction: rafthttp doesn't support talking to unknown member Current workflow for newly joined member: add new member to the cluster start the new member with name and initial cluster(which contains name and peerURLs of each member) the new member grabs the latest cluster info and etcd index from remote members it assigns member ID to local cluster info, and record etcd index. It is important to keep latest peer URLs in the cluster here because it guides rafthttp how to talk to working members when they haven't appeared in the store. its cluster has inconsistent view with store from now on its cluster rejects to be recovered from a snapshot whose index is smaller than recorded etcd index(#2348) It may miss the latest Client URLs here. its cluster rejects to apply conf change whose index is smaller than recorded etcd index(#2556) its cluster should omit invalid updateAttributes request whose index is smaller than recorded etcd index(#2683, #2690 ) it catches up the progress and its cluster have consistent view with the store In the process for newly joined member to catch up the progress and have consistent view with the store, its cluster runs in this way: its ID and peerURLs are as new as remote members its ClientURLs of known IDs are consistent with the store, and it doesn't have client URLs of unknown IDs(the ID that has been removed in the cluster) its Names start from what are given in initial-cluster, and apply name change on known IDs. Known problems for this approach: users may get inconsistent member view during the process to catch up etcd may fail to restart if killed during the process The probability to meet the problem is low because 1. people don't add/remove member a lot 2. the catch-up process is short So the solution is ok today. If we want to fully resolve the problem, we need to remove the restriction(make rafthttp talk to unknown member). It may need 1-2 week to fully support it. If we do this, cluster can always have consistent view with the store. /cc @xiang90 @barakmich @yichengq I think this is growing out of control. I think we should explore the original approach more. Instead of building a future view of the cluster, we simply add id->addr map to rafthttp and just enough for sending a rpc message (no streaming or any sort of connection keep alive). When adding a member from raft wal/conf snap, we do what we are doing in today's rafthttp. When removing a mbmer from raft wal/conf snap, we do what we are doing in today's rafthttp. The only change we need to make is to maintain the id->addr map in rafthttp instead of cluster. So cluster always represents the current view of the cluster and always be consistent with wal/snap.
gharchive/issue
2015-04-15T18:28:08
2025-04-01T04:33:52.987342
{ "authors": [ "xiang90", "yichengq" ], "repo": "coreos/etcd", "url": "https://github.com/coreos/etcd/issues/2689", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
54239401
integration: add TestForceNewCluster for #1467 refreshed. lgtm
gharchive/pull-request
2015-01-13T19:52:37
2025-04-01T04:33:52.988810
{ "authors": [ "xiang90", "yichengq" ], "repo": "coreos/etcd", "url": "https://github.com/coreos/etcd/pull/2084", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
87843045
rafthttp: use buffered channel as recv/prop chan So it ensures that the message will not be discarded because the receive side has not been ready, which happens easily in multiple core test. Use log.fatal instead of log.error. The test exits when there is something wrong because the error may affect following test cases. fixes #2961 LGTM
gharchive/pull-request
2015-06-12T20:35:38
2025-04-01T04:33:52.990108
{ "authors": [ "xiang90", "yichengq" ], "repo": "coreos/etcd", "url": "https://github.com/coreos/etcd/pull/2970", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
124259727
*: update api proto Update proto to reflect two changes: add lease field to kv: fix https://github.com/coreos/etcd/issues/3919 do not allow modifying a key several times in one txn After this pull request, we also need to clean up api document, add more examples, regenerate protobuf files. /cc @heyitsanthony @gyuho API LGTM as we discussed at the meeting. Are you going to change all Go code in the same PR as well? @gyuho Yes. In next pull request probably. There might be other places I want to change while I clean doc this afternoon. lgtm, rmw-free leases can be added later if needed
gharchive/pull-request
2015-12-29T19:50:38
2025-04-01T04:33:52.992687
{ "authors": [ "gyuho", "heyitsanthony", "xiang90" ], "repo": "coreos/etcd", "url": "https://github.com/coreos/etcd/pull/4092", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
786322939
release: Add a ToStreamArchitectures API This drains the primary logic from https://github.com/coreos/fedora-coreos-stream-generator into this repo, so it can be more easily shared with coreos-assembler. Demo patch for the generator commit 0805b718650603bc3face2974ce5cc5afd7f79c6 Author: Colin Walters <walters@verbum.org> AuthorDate: Thu Jan 14 20:39:21 2021 +0000 Commit: Colin Walters <walters@verbum.org> CommitDate: Thu Jan 14 20:42:54 2021 +0000 wip diff --git a/go.mod b/go.mod index 86a4364..9dbe85d 100644 --- a/go.mod +++ b/go.mod @@ -3,3 +3,6 @@ module github.com/coreos/fedora-coreos-stream-generator go 1.12 require github.com/coreos/stream-metadata-go v0.0.0-20210107232620-d808ce9d237c + +replace "github.com/coreos/stream-metadata-go" v0.0.0-20210107232620-d808ce9d237c => "../stream-metadata-go/" + diff --git a/main.go b/main.go index 00ab1b4..43ba63b 100644 --- a/main.go +++ b/main.go @@ -55,173 +55,6 @@ func getReleaseURL(releaseIndexURL string) (string, error) { return relIndex.Releases[len(relIndex.Releases)-1].MetadataURL, nil } -func mapArtifact(ra *release.Artifact) *stream.Artifact { - if ra == nil { - return nil - } - return &stream.Artifact{ - Location: ra.Location, - Signature: ra.Signature, - Sha256: ra.Sha256, - } -} - -func mapFormats(m map[string]release.ImageFormat) map[string]stream.ImageFormat { - r := make(map[string]stream.ImageFormat) - for k, v := range m { - r[k] = stream.ImageFormat{ - Disk: mapArtifact(v.Disk), - Kernel: mapArtifact(v.Kernel), - Initramfs: mapArtifact(v.Initramfs), - Rootfs: mapArtifact(v.Rootfs), - } - } - return r -} - -func releaseToStream(releaseArch *release.Arch, rel release.Release) stream.Arch { - artifacts := make(map[string]stream.PlatformArtifacts) - cloudImages := stream.Images{} - if releaseArch.Media.Aws != nil { - artifacts["aws"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Aws.Artifacts), - } - awsAmis := stream.AwsImage{ - Regions: make(map[string]stream.AwsRegionImage), - } - if releaseArch.Media.Aws.Images != nil { - for region, ami := range releaseArch.Media.Aws.Images { - ri := stream.AwsRegionImage{Release: rel.Release, Image: ami.Image} - awsAmis.Regions[region] = ri - - } - cloudImages.Aws = &awsAmis - } - } - - if releaseArch.Media.Azure != nil { - artifacts["azure"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Azure.Artifacts), - } - - // Not enabled right now - // if az := releaseArch.Media.Azure.Images; az != nil && az.Global != nil && az.Global.Image != nil { - // azureImage := StreamCloudImage{} - // azureImage.Image = fmt.Sprintf("Fedora:CoreOS:%s:latest", rel.Stream) - // cloudImages.Azure = &azureImage - // } - } - - if releaseArch.Media.Aliyun != nil { - artifacts["aliyun"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Aliyun.Artifacts), - } - } - - if releaseArch.Media.Exoscale != nil { - artifacts["exoscale"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Exoscale.Artifacts), - } - } - - if releaseArch.Media.Vultr != nil { - artifacts["vultr"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Vultr.Artifacts), - } - } - - if releaseArch.Media.Gcp != nil { - artifacts["gcp"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Gcp.Artifacts), - } - - if releaseArch.Media.Gcp.Image != nil { - cloudImages.Gcp = &stream.GcpImage{ - Name: releaseArch.Media.Gcp.Image.Name, - Family: releaseArch.Media.Gcp.Image.Family, - Project: releaseArch.Media.Gcp.Image.Project, - } - } - } - - if releaseArch.Media.Digitalocean != nil { - artifacts["digitalocean"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Digitalocean.Artifacts), - } - - /* We're producing artifacts but they're not yet available - in DigitalOcean as distribution images. - digitalOceanImage := stream.CloudImage{Image: fmt.Sprintf("fedora-coreos-%s", release.Stream)} - cloudImages.Digitalocean = &digitalOceanImage - */ - } - - if releaseArch.Media.Ibmcloud != nil { - artifacts["ibmcloud"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Ibmcloud.Artifacts), - } - } - - // if releaseArch.Media.Packet != nil { - // packet := StreamMediaDetails{ - // Release: rel.Release, - // Formats: releaseArch.Media.Packet.Artifacts, - // } - // artifacts.Packet = &packet - - // packetImage := StreamCloudImage{Image: fmt.Sprintf("fedora_coreos_%s", rel.Stream)} - // cloudImages.Packet = &packetImage - // } - - if releaseArch.Media.Openstack != nil { - artifacts["openstack"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Openstack.Artifacts), - } - } - - if releaseArch.Media.Qemu != nil { - artifacts["qemu"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Qemu.Artifacts), - } - } - - // if releaseArch.Media.Virtualbox != nil { - // virtualbox := StreamMediaDetails{ - // Release: rel.Release, - // Formats: releaseArch.Media.Virtualbox.Artifacts, - // } - // artifacts.Virtualbox = &virtualbox - // } - - if releaseArch.Media.Vmware != nil { - artifacts["vmware"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Vmware.Artifacts), - } - } - - if releaseArch.Media.Metal != nil { - artifacts["metal"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Metal.Artifacts), - } - } - - return stream.Arch{ - Artifacts: artifacts, - Images: cloudImages, - } -} func overrideData(original, override interface{}) interface{} { switch override1 := override.(type) { @@ -305,15 +138,15 @@ func run() error { return fmt.Errorf("Error while decoding json: %v", err) } - streamArch := make(map[string]stream.Arch) - for arch, releaseArch := range rel.Architectures { - streamArch[arch] = releaseToStream(&releaseArch, rel) + streamArches, err := rel.ToStreamArchitectures() + if err != nil { + return err } streamMetadata := stream.Stream{ Stream: rel.Stream, Metadata: stream.Metadata{LastModified: time.Now().UTC().Format(time.RFC3339)}, - Architectures: streamArch, + Architectures: streamArches, } if overrideFilename != "" { Demo patch for the generator commit 0805b718650603bc3face2974ce5cc5afd7f79c6 Author: Colin Walters <walters@verbum.org> AuthorDate: Thu Jan 14 20:39:21 2021 +0000 Commit: Colin Walters <walters@verbum.org> CommitDate: Thu Jan 14 20:42:54 2021 +0000 wip diff --git a/go.mod b/go.mod index 86a4364..9dbe85d 100644 --- a/go.mod +++ b/go.mod @@ -3,3 +3,6 @@ module github.com/coreos/fedora-coreos-stream-generator go 1.12 require github.com/coreos/stream-metadata-go v0.0.0-20210107232620-d808ce9d237c + +replace "github.com/coreos/stream-metadata-go" v0.0.0-20210107232620-d808ce9d237c => "../stream-metadata-go/" + diff --git a/main.go b/main.go index 00ab1b4..43ba63b 100644 --- a/main.go +++ b/main.go @@ -55,173 +55,6 @@ func getReleaseURL(releaseIndexURL string) (string, error) { return relIndex.Releases[len(relIndex.Releases)-1].MetadataURL, nil } -func mapArtifact(ra *release.Artifact) *stream.Artifact { - if ra == nil { - return nil - } - return &stream.Artifact{ - Location: ra.Location, - Signature: ra.Signature, - Sha256: ra.Sha256, - } -} - -func mapFormats(m map[string]release.ImageFormat) map[string]stream.ImageFormat { - r := make(map[string]stream.ImageFormat) - for k, v := range m { - r[k] = stream.ImageFormat{ - Disk: mapArtifact(v.Disk), - Kernel: mapArtifact(v.Kernel), - Initramfs: mapArtifact(v.Initramfs), - Rootfs: mapArtifact(v.Rootfs), - } - } - return r -} - -func releaseToStream(releaseArch *release.Arch, rel release.Release) stream.Arch { - artifacts := make(map[string]stream.PlatformArtifacts) - cloudImages := stream.Images{} - if releaseArch.Media.Aws != nil { - artifacts["aws"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Aws.Artifacts), - } - awsAmis := stream.AwsImage{ - Regions: make(map[string]stream.AwsRegionImage), - } - if releaseArch.Media.Aws.Images != nil { - for region, ami := range releaseArch.Media.Aws.Images { - ri := stream.AwsRegionImage{Release: rel.Release, Image: ami.Image} - awsAmis.Regions[region] = ri - - } - cloudImages.Aws = &awsAmis - } - } - - if releaseArch.Media.Azure != nil { - artifacts["azure"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Azure.Artifacts), - } - - // Not enabled right now - // if az := releaseArch.Media.Azure.Images; az != nil && az.Global != nil && az.Global.Image != nil { - // azureImage := StreamCloudImage{} - // azureImage.Image = fmt.Sprintf("Fedora:CoreOS:%s:latest", rel.Stream) - // cloudImages.Azure = &azureImage - // } - } - - if releaseArch.Media.Aliyun != nil { - artifacts["aliyun"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Aliyun.Artifacts), - } - } - - if releaseArch.Media.Exoscale != nil { - artifacts["exoscale"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Exoscale.Artifacts), - } - } - - if releaseArch.Media.Vultr != nil { - artifacts["vultr"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Vultr.Artifacts), - } - } - - if releaseArch.Media.Gcp != nil { - artifacts["gcp"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Gcp.Artifacts), - } - - if releaseArch.Media.Gcp.Image != nil { - cloudImages.Gcp = &stream.GcpImage{ - Name: releaseArch.Media.Gcp.Image.Name, - Family: releaseArch.Media.Gcp.Image.Family, - Project: releaseArch.Media.Gcp.Image.Project, - } - } - } - - if releaseArch.Media.Digitalocean != nil { - artifacts["digitalocean"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Digitalocean.Artifacts), - } - - /* We're producing artifacts but they're not yet available - in DigitalOcean as distribution images. - digitalOceanImage := stream.CloudImage{Image: fmt.Sprintf("fedora-coreos-%s", release.Stream)} - cloudImages.Digitalocean = &digitalOceanImage - */ - } - - if releaseArch.Media.Ibmcloud != nil { - artifacts["ibmcloud"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Ibmcloud.Artifacts), - } - } - - // if releaseArch.Media.Packet != nil { - // packet := StreamMediaDetails{ - // Release: rel.Release, - // Formats: releaseArch.Media.Packet.Artifacts, - // } - // artifacts.Packet = &packet - - // packetImage := StreamCloudImage{Image: fmt.Sprintf("fedora_coreos_%s", rel.Stream)} - // cloudImages.Packet = &packetImage - // } - - if releaseArch.Media.Openstack != nil { - artifacts["openstack"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Openstack.Artifacts), - } - } - - if releaseArch.Media.Qemu != nil { - artifacts["qemu"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Qemu.Artifacts), - } - } - - // if releaseArch.Media.Virtualbox != nil { - // virtualbox := StreamMediaDetails{ - // Release: rel.Release, - // Formats: releaseArch.Media.Virtualbox.Artifacts, - // } - // artifacts.Virtualbox = &virtualbox - // } - - if releaseArch.Media.Vmware != nil { - artifacts["vmware"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Vmware.Artifacts), - } - } - - if releaseArch.Media.Metal != nil { - artifacts["metal"] = stream.PlatformArtifacts{ - Release: rel.Release, - Formats: mapFormats(releaseArch.Media.Metal.Artifacts), - } - } - - return stream.Arch{ - Artifacts: artifacts, - Images: cloudImages, - } -} func overrideData(original, override interface{}) interface{} { switch override1 := override.(type) { @@ -305,15 +138,15 @@ func run() error { return fmt.Errorf("Error while decoding json: %v", err) } - streamArch := make(map[string]stream.Arch) - for arch, releaseArch := range rel.Architectures { - streamArch[arch] = releaseToStream(&releaseArch, rel) + streamArches, err := rel.ToStreamArchitectures() + if err != nil { + return err } streamMetadata := stream.Stream{ Stream: rel.Stream, Metadata: stream.Metadata{LastModified: time.Now().UTC().Format(time.RFC3339)}, - Architectures: streamArch, + Architectures: streamArches, } if overrideFilename != "" {
gharchive/pull-request
2021-01-14T20:42:46
2025-04-01T04:33:53.027073
{ "authors": [ "cgwalters" ], "repo": "coreos/stream-metadata-go", "url": "https://github.com/coreos/stream-metadata-go/pull/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
715764319
Permit "version" string in content-type header, rebased to 3.4 + test, v2 Replacement for PR #1877. Closing in favour of PR #1901.
gharchive/pull-request
2020-10-06T15:03:12
2025-04-01T04:33:53.029146
{ "authors": [ "azurit" ], "repo": "coreruleset/coreruleset", "url": "https://github.com/coreruleset/coreruleset/pull/1900", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1907778448
Add Prettier Pre-Commit Hook Summary Adds a pre-commit hook that runs prettier Test Plan Make some changes to ts file and make changes, see fixes when committing Wait I couldn't replicate this. Also, I'm assuming if there are discrepancies between the two, it's due to .editorconfig, which there doesn't seem to be a way for the pre-commit hook to follow https://stackoverflow.com/questions/71503396/how-to-use-editorconfig-for-code-formatting-in-git-pre-commit-hook
gharchive/pull-request
2023-09-21T21:17:09
2025-04-01T04:33:53.047324
{ "authors": [ "andxu282" ], "repo": "cornell-dti/course-plan", "url": "https://github.com/cornell-dti/course-plan/pull/854", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
430402550
There were some errors in compilation . There were some errors when I compile the Makefile in bnn/cpp. Here is the log #:~/work/accel/bnn-fpga-master/cpp$ make -j4 make -C minizip make[1]: Entering directory '/home/null/work/accel/bnn-fpga-master/cpp/minizip' cc -O -I../.. -DHAVE_AES -o miniunz miniunz.o unzip.o ioapi.o libaes.a -lz cc -O -I../.. -DHAVE_AES -o minizip minizip.o zip.o ioapi.o libaes.a -lz /usr/bin/ld: libaes.a(aescrypt.o): relocation R_X86_64_32 against symbol t_fn' can not be used when making a PIE object; recompile with -fPIC /usr/bin/ld: libaes.a(aeskey.o): relocation R_X86_64_32 against symbol t_fl' can not be used when making a PIE object; recompile with -fPIC /usr/bin/ld: final link failed: Nonrepresentable section on output collect2: error: ld returned 1 exit status Makefile:23: recipe for target 'miniunz' failed make[1]: *** [miniunz] Error 1 make[1]: *** Waiting for unfinished jobs.... /usr/bin/ld: libaes.a(entropy.o): relocation R_X86_64_32 against .rodata.str1.1' can not be used when making a PIE object; recompile with -fPIC /usr/bin/ld: libaes.a(aescrypt.o): relocation R_X86_64_32 against symbol t_fn' can not be used when making a PIE object; recompile with -fPIC /usr/bin/ld: libaes.a(aeskey.o): relocation R_X86_64_32 against symbol `t_fl' can not be used when making a PIE object; recompile with -fPIC /usr/bin/ld: final link failed: Nonrepresentable section on output collect2: error: ld returned 1 exit status Makefile:26: recipe for target 'minizip' failed make[1]: *** [minizip] Error 1 make[1]: Leaving directory '/home/null/work/accel/bnn-fpga-master/cpp/minizip' Makefile:5: recipe for target 'all' failed make: *** [all] Error 2 I met the same problem, have you solved this issue?
gharchive/issue
2019-04-08T11:40:18
2025-04-01T04:33:53.052145
{ "authors": [ "TruNcat3", "jiangwx" ], "repo": "cornell-zhang/bnn-fpga", "url": "https://github.com/cornell-zhang/bnn-fpga/issues/25", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2036435147
[Bug] VOI LUT images not rendered correctly Describe the Bug Images with a VOI LUT don't get their levels computed correctly when rendering, so they display incorrectly. Cornerstone3D version: 1.36.3 This works as expected when using Cornerstone JS 2.6.1 (wado image loader 3.3.2) Attached is a file which demonstrates the issue. hand.dcm.zip Correct rendering Current rendering (ignore the green annotation in the top right corner) Tags relevant to VOI LUT (0028,3010) SQ (Sequence with undefined length #=1) # u/l, 1 VOILUTSequence (fffe,e000) na (Item with undefined length #=6) # u/l, 1 Item (0028,3002) US 16384\0\14 # 6, 3 LUTDescriptor (0028,3003) LO [User Selection] # 14, 1 LUTExplanation (0028,3006) OW 0000\0000\0000\0000\0000\0000\0000\0000\0000\0000\0000\0000\0000... # 32768, 1 LUTData (07a3,0010) LO [ELSCINT1] # 8, 1 PrivateCreator (07a3,109d) DS [2332] # 4, 1 Unknown Tag & Data (07a3,109e) DS [1149] # 4, 1 Unknown Tag & Data Steps to Reproduce Load an image with a VOI LUT The current behavior The image gets rendered with incorrect levels The expected behavior The image renders with correct levels. This used to work in the "old" cornerstone. OS Linux Node version 18.18.2 Browser Chromium 119.0.6045.159, Firefox 120.0b9 Unfortunately, we currently do not have support for using the predefined VOI LUT in DICOM. Do you have an idea how hard it’d be to port the logic from the “old” cornerstone? Just to know if it’s something I, as a total outsider, could consider doing… Hmm, was it working in the old cs? Yup, there it worked without doing anything special Hello, Are there any plans to implement a predefined VOI LUT in DICOM? @sedghi are there any plans to support voi luts, dicom tag 0028 3010? It is really important that it works. It's one of the requirements in law that software for rtg dicoms diagnosis has it. https://isap.sejm.gov.pl/isap.nsf/download.xsp/WDU20230000195/O/D20230195.pdf page 9 4) c).
gharchive/issue
2023-12-11T20:04:22
2025-04-01T04:33:53.058755
{ "authors": [ "AnnaKuduk", "abustany", "kubanm3", "sedghi" ], "repo": "cornerstonejs/cornerstone3D", "url": "https://github.com/cornerstonejs/cornerstone3D/issues/938", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
626291507
Adding default values to configuration classes to prevent null pointe… Adding default values to configuration classes to prevent null pointer exceptions on empty config file Also adding some comments to config values I created a new pull request because I could not get the old one rebased to current master. Sorry for the confusion, you can delete this pull or leave it as documentation https://github.com/corona-warn-app/cwa-verification-server/pull/130
gharchive/pull-request
2020-05-28T07:43:12
2025-04-01T04:33:53.063999
{ "authors": [ "martinschefflerTSI" ], "repo": "corona-warn-app/cwa-verification-server", "url": "https://github.com/corona-warn-app/cwa-verification-server/pull/92", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }