added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:38:10.422953
2021-12-22T01:27:11
1086362235
{ "authors": [ "Bento007" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4593", "repo": "chanzuckerberg/single-cell-data-portal", "url": "https://github.com/chanzuckerberg/single-cell-data-portal/issues/1786" }
gharchive/issue
bug(duplicate detection): Etags cannot be used to compare objects Describe the bug Duplicate detection does not detect duplicated when publishing a collection. Current we use AWS S3 Etags to determine if two object in are identical. However since we are using KMS to encrypt our object in place, AWS does not guarantee that Etags will be identical for identical objects in S3. Expected behavior Duplicates should be caught before publishing a collection Solution Add a custom tag to objects upload to s3 which represents the hash of the upload object. This tag will be used compare objects uploaded to S3. Alternatively this tag can be store in the database rather than on the S3 object. Whatever implementation we go with, we should generate tags for all previously uploaded dataset. Unable to use the hash generated by dropbox because we are using a shared link to a file. https://www.dropboxforum.com/t5/Dropbox-API-Support-Feedback/API-to-get-content-hash-of-shared-file-link/td-p/458942
2025-04-01T06:38:10.430778
2020-08-07T01:30:50
674694039
{ "authors": [ "MDunitz", "ambrosejcarr", "brianraymor", "signechambers1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4594", "repo": "chanzuckerberg/single-cell", "url": "https://github.com/chanzuckerberg/single-cell/issues/9" }
gharchive/issue
Measure hosted cellxgene feature usage For the hosted cellxgene application Measure how many times key features are used per visit. Measure the probability that two key features are used in sequence Business Justification This data will help us tailor UX research to understand if: features are difficult to use, or features don't provide value to users, or features are used in particular flows or contexts, but not others. List of "key features" Left sidebar: painting by a category on left side bar selecting cells by unselecting some values within a category opening a category on left side bar creating a new category on left side bar Top bar: lassoing a selection of cells saving a selection of cells to a DE category saving a selection of cells to a category and assigning a label calculating differential expression between two categories subsetting to a selection of cells from a lasso selection returning to the full dataset selecting the zoom tool show labels on graph use the clip tool undo an action redo an action use the help menu, or any sub-menu item. Right sidebar: Add a gene Bulk add genes create an xy plot color the scatterplot by a quantitative variable (e.g. gene expression) select a range of cells using the range selector Have we talked to Trust about this? It seems like per visit and order of actions is getting pretty far into user tracking? I raised with Trust on August 27 and pinged again this morning. Per September 1 Refinement, @signechambers1 to follow up with @ambrosejcarr on the relative priority of this issue. Trust approved the increased instrumentation on 9/15, from Shawn "Our current cellxgene policy covers this (see section 2.2) but I’ve made a note to spell this out even more since we are updating the policies for Data Portal anyways." Per PM triage, closing. Ambrose will open a new issue to track specific actions.
2025-04-01T06:38:10.434521
2022-04-06T19:34:13
1195083033
{ "authors": [ "Dirrk", "alldoami", "mrlannigan", "sfc-gh-swinkler" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4595", "repo": "chanzuckerberg/terraform-provider-snowflake", "url": "https://github.com/chanzuckerberg/terraform-provider-snowflake/pull/962" }
gharchive/pull-request
feat: Add port/protocol for driver Adds the ability to provide protocol and port options to the snowflake driver Test Plan [X] acceptance tests [X] Locally we have tested this for almost two weeks References https://github.com/snowflakedb/gosnowflake/blob/master/dsn.go#L54 I rebased the changes in main and made sure the docs were updated properly this time, should be ready for testing /ok-to-test sha=b6f792a We fixed this test in a different PR, could you merge main into your branch and then we can merge your PR? Hi @alldoami, I have updated this PR with the latest from main. /ok-to-test sha=0987fe9 /ok-to-test sha=0987fe9 #1238 closes this PR
2025-04-01T06:38:10.497542
2018-03-02T12:48:35
301755804
{ "authors": [ "GeorgLink", "jsdabrowski" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4596", "repo": "chaoss/governance", "url": "https://github.com/chaoss/governance/pull/11" }
gharchive/pull-request
Update GSoC-interest.md provided pull request with my information @jsdabrowski are you no longerer interested or did you close the pull request for a different reason? am interested Wonderful, you can make the requested changes in your repository and it will update this pull request :-) For this pull request to update, you have to push the changes to the same branch that this pull request is from: patch-1 I hope should it be good. Sorry for the mess. i am still in the learning process. No worries, I found GitHub and pull requests confusing at first. You are 2/3 done now: [ ] Please update your pull request to not delete other interested students that are already in the list. [x] Please change the format of your email to use "at" instead of @ [x] Please change the last column to read "To-Do" without a link, if you don't have your Project Proposal ready yet. hope in the project will be many opportunity to advance the skills thank you
2025-04-01T06:38:10.585226
2016-03-17T23:25:01
141733466
{ "authors": [ "chapmanb", "tanglingfung" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4603", "repo": "chapmanb/bcbio-nextgen", "url": "https://github.com/chapmanb/bcbio-nextgen/issues/1277" }
gharchive/issue
KeyError: 'fasta' Hi Brad, After upgrading the pipeline, I have the following error when I start a run. Where should I be looking for the configuration file? Paul File "/media/bigdata/bcbio-nextgen/anaconda/lib/python2.7/site-packages/bcbio/pipeline/run_info.py", line 179, in _check_ref_files for contig in ref.file_contigs(ref_info["fasta"]["base"], data["config"]): KeyError: 'fasta' Paul; Sorry about the confusing error message. I was able to replicate this error if you don't add a genome_build to your configuration. I updated the configuration checking to identify this and provide a more useful message. Hopefully this explains the problem and once adding it you can get things running cleanly. Thanks again for the feedback. Thanks Brad. I just found that my .yaml file is truncated where the genome_build is taken off for the last sample. Thanks for digging into the issue. On Fri, Mar 18, 2016 at 2:50 AM, Brad Chapman<EMAIL_ADDRESS>wrote: Paul; Sorry about the confusing error message. I was able to replicate this error if you don't add a genome_build to your configuration. I updated the configuration checking to identify this and provide a more useful message. Hopefully this explains the problem and once adding it you can get things running cleanly. Thanks again for the feedback. — You are receiving this because you authored the thread. Reply to this email directly or view it on GitHub https://github.com/chapmanb/bcbio-nextgen/issues/1277#issuecomment-198284377
2025-04-01T06:38:10.594269
2015-09-27T09:57:10
108525464
{ "authors": [ "charlesthk", "guettli" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4604", "repo": "charlesthk/python-resize-image", "url": "https://github.com/charlesthk/python-resize-image/issues/7" }
gharchive/issue
Why open image with 'rw' in docs In the docs (https://github.com/charlesthk/python-resize-image) you use 'rw' to open the images. from PIL import Image test_img = open('test-image.jpeg', 'rw') img = Image.open(test_img) img = resizeimage.resize_thumbnail(img, [200, 200]) img.save('test-image-thumbnail.jpeg', img.format) test_img.close() I see no sense in opening the file for writing. BTW, it is very common to use fd_.... as variable name for file descriptors returned by open(). You use test_img. That's valid python code, but fd_img would be more easy to read. Thanks for this issue, I updated my code accordingly.
2025-04-01T06:38:10.764791
2017-02-17T05:56:08
208344171
{ "authors": [ "Conrad-S", "Sicos1977", "andimalaj", "anpham293", "charlesw", "losbear" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4605", "repo": "charlesw/tesseract", "url": "https://github.com/charlesw/tesseract/issues/318" }
gharchive/issue
Failed to initialise tesseract engine (Tesseract 3.0.2 - Visual Studio 2015) Hi and thank you very much for your help. I have done everything that I possibly can think of to attempt to resolve a "Failed to initialise engine" error. I read post #277 where the person stated that adjusting the path to tessdata fixed his issue. But he did not state the "adjustment" that he made. Line where the failure occurs using (var engine = new TesseractEngine (@"./tessdata", "eng", EngineMode.Default)) { Error "Failed to initialise tesseract engine.. See https://github.com/charlesw/tesseract/wiki/Error-1 for details."} Tesseract.TesseractException Stack Trace Data {System.Collections.ListDictionaryInternal} System.Collections.IDictionary {System.Collections.ListDictionaryInternal} HResult -2146233088 int HelpLink null string InnerException null System.Exception Message "Failed to initialise tesseract engine.. See https://github.com/charlesw/tesseract/wiki/Error-1 for details." string Source "Tesseract" string StackTrace " at Tesseract.TesseractEngine.Initialise(String datapath, String language, EngineMode engineMode, IEnumerable1 configFiles, IDictionary2 initialValues, Boolean setOnlyNonDebugVariables)\r\n at Tesseract.TesseractEngine..ctor(String datapath, String language, EngineMode engineMode, IEnumerable1 configFiles, IDictionary2 initialOptions, Boolean setOnlyNonDebugVariables)\r\n at Tesseract.TesseractEngine..ctor(String datapath, String language, EngineMode engineMode)\r\n at OCRTestOne.OCRTest.Go() in C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\Classes\OCRTest.cs:line 22" string TargetSite {Void Initialise(System.String, System.String, Tesseract.EngineMode, System.Collections.Generic.IEnumerable1[System.String], System.Collections.Generic.IDictionary2[System.String,System.Object], Boolean)} System.Reflection.MethodBase {System.Reflection.RuntimeMethodInfo} Development Environment Windows Server 2012 R2 (virtual machine running in Azure) Visual Studio 2015 (Version 14.0.25431.01 Update 3) Project Type: ASP.NET (also tried console app). Framework Target 452 (.NET Framework 4.5.2). Verified that Visual C++ 32-bit and 64-bit runtimes are both installed Nuget Package: Tesseract by Charles Weld 3.0.2 tessdata folder exists with English language files underneath (I added it). Files set to Copy Always. 32 and 64-bit assemblies (liblept172.dll and libtesseract304.dll) exist in the x86 and x64 folders and are set to Copy Always (they seem to be copied to the lib folder). Trace Turned on verbose mode tracing. The trace below is based on one execution attempt (I tried to attach a text file but the site said: "We don't support that file type"). Apologies for the long trace Tesseract Information: 0 : Current OS: Windows Tesseract Information: 0 : Current platform: x86 Tesseract Information: 0 : Custom search path is not defined, skipping. Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for 'liblept172.dll' on platform x86. Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'liblept172.dll' on platform x86. Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll"... Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll", handle = 325451776. Tesseract Information: 0 : Trying to load native function "pixaReadMultipageTiff" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaReadMultipageTiff", function handle = 325469402. Tesseract Information: 0 : Trying to load native function "pixaGetCount" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaGetCount", function handle = 325462092. Tesseract Information: 0 : Trying to load native function "pixaGetPix" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaGetPix", function handle = 325455972. Tesseract Information: 0 : Trying to load native function "pixaDestroy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaDestroy", function handle = 325456927. Tesseract Information: 0 : Trying to load native function "pixCreate" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixCreate", function handle = 325465967. Tesseract Information: 0 : Trying to load native function "pixClone" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixClone", function handle = 325466612. Tesseract Information: 0 : Trying to load native function "pixDestroy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixDestroy", function handle = 325466382. Tesseract Information: 0 : Trying to load native function "pixGetWidth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetWidth", function handle = 325460622. Tesseract Information: 0 : Trying to load native function "pixGetHeight" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetHeight", function handle = 325460452. Tesseract Information: 0 : Trying to load native function "pixGetDepth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetDepth", function handle = 325464312. Tesseract Information: 0 : Trying to load native function "pixGetXRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetXRes", function handle = 325463887. Tesseract Information: 0 : Trying to load native function "pixGetYRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetYRes", function handle = 325463882. Tesseract Information: 0 : Trying to load native function "pixGetResolution" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetResolution", function handle = 325469492. Tesseract Information: 0 : Trying to load native function "pixGetWpl" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetWpl", function handle = 325466032. Tesseract Information: 0 : Trying to load native function "pixSetXRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetXRes", function handle = 325463907. Tesseract Information: 0 : Trying to load native function "pixSetYRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetYRes", function handle = 325463902. Tesseract Information: 0 : Trying to load native function "pixSetResolution" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetResolution", function handle = 325465697. Tesseract Information: 0 : Trying to load native function "pixScaleResolution" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixScaleResolution", function handle = 325467282. Tesseract Information: 0 : Trying to load native function "pixGetData" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetData", function handle = 325469052. Tesseract Information: 0 : Trying to load native function "pixGetInputFormat" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetInputFormat", function handle = 325467712. Tesseract Information: 0 : Trying to load native function "pixSetInputFormat" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetInputFormat", function handle = 325459102. Tesseract Information: 0 : Trying to load native function "pixEndianByteSwap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixEndianByteSwap", function handle = 325468587. Tesseract Information: 0 : Trying to load native function "pixRead" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixRead", function handle = 325467527. Tesseract Information: 0 : Trying to load native function "pixReadMemTiff" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixReadMemTiff", function handle = 325472137. Tesseract Information: 0 : Trying to load native function "pixWrite" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixWrite", function handle = 325466617. Tesseract Information: 0 : Trying to load native function "pixGetColormap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetColormap", function handle = 325464692. Tesseract Information: 0 : Trying to load native function "pixSetColormap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetColormap", function handle = 325468467. Tesseract Information: 0 : Trying to load native function "pixDestroyColormap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixDestroyColormap", function handle = 325465782. Tesseract Information: 0 : Trying to load native function "pixConvertRGBToGray" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixConvertRGBToGray", function handle = 325460777. Tesseract Information: 0 : Trying to load native function "pixDeskewGeneral" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixDeskewGeneral", function handle = 325460177. Tesseract Information: 0 : Trying to load native function "pixRotate" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixRotate", function handle = 325466312. Tesseract Information: 0 : Trying to load native function "pixRotateOrth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixRotateOrth", function handle = 325468177. Tesseract Information: 0 : Trying to load native function "pixOtsuAdaptiveThreshold" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixOtsuAdaptiveThreshold", function handle = 325468767. Tesseract Information: 0 : Trying to load native function "pixSauvolaBinarize" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSauvolaBinarize", function handle = 325463922. Tesseract Information: 0 : Trying to load native function "pixSauvolaBinarizeTiled" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSauvolaBinarizeTiled", function handle = 325461412. Tesseract Information: 0 : Trying to load native function "pixScale" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixScale", function handle = 325466597. Tesseract Information: 0 : Trying to load native function "pixcmapCreate" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCreate", function handle = 325462397. Tesseract Information: 0 : Trying to load native function "pixcmapCreateRandom" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCreateRandom", function handle = 325460482. Tesseract Information: 0 : Trying to load native function "pixcmapCreateLinear" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCreateLinear", function handle = 325460392. Tesseract Information: 0 : Trying to load native function "pixcmapCopy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCopy", function handle = 325463212. Tesseract Information: 0 : Trying to load native function "pixcmapDestroy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapDestroy", function handle = 325464252. Tesseract Information: 0 : Trying to load native function "pixcmapGetCount" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetCount", function handle = 325466562. Tesseract Information: 0 : Trying to load native function "pixcmapGetFreeCount" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetFreeCount", function handle = 325456837. Tesseract Information: 0 : Trying to load native function "pixcmapGetDepth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetDepth", function handle = 325468462. Tesseract Information: 0 : Trying to load native function "pixcmapGetMinDepth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetMinDepth", function handle = 325462157. Tesseract Information: 0 : Trying to load native function "pixcmapClear" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapClear", function handle = 325460977. Tesseract Information: 0 : Trying to load native function "pixcmapAddColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddColor", function handle = 325466517. Tesseract Information: 0 : Trying to load native function "pixcmapAddNewColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddNewColor", function handle = 325459832. Tesseract Information: 0 : Trying to load native function "pixcmapAddNearestColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddNearestColor", function handle = 325464897. Tesseract Information: 0 : Trying to load native function "pixcmapUsableColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapUsableColor", function handle = 325468852. Tesseract Information: 0 : Trying to load native function "pixcmapAddBlackOrWhite" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddBlackOrWhite", function handle = 325460002. Tesseract Information: 0 : Trying to load native function "pixcmapSetBlackAndWhite" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapSetBlackAndWhite", function handle = 325456282. Tesseract Information: 0 : Trying to load native function "pixcmapGetColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetColor", function handle = 325466527. Tesseract Information: 0 : Trying to load native function "pixcmapGetColor32" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetColor32", function handle = 325458992. Tesseract Information: 0 : Trying to load native function "pixcmapResetColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapResetColor", function handle = 325457197. Tesseract Information: 0 : Trying to load native function "pixcmapGetIndex" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetIndex", function handle = 325466567. Tesseract Information: 0 : Trying to load native function "pixcmapHasColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapHasColor", function handle = 325458122. Tesseract Information: 0 : Trying to load native function "pixcmapCountGrayColors" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCountGrayColors", function handle = 325462272. Tesseract Information: 0 : Trying to load native function "pixcmapCountGrayColors" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCountGrayColors", function handle = 325462272. Tesseract Information: 0 : Trying to load native function "pixcmapGetNearestIndex" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetNearestIndex", function handle = 325459557. Tesseract Information: 0 : Trying to load native function "pixcmapGetNearestGrayIndex" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetNearestGrayIndex", function handle = 325466522. Tesseract Information: 0 : Trying to load native function "pixcmapGetComponentRange" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetComponentRange", function handle = 325459077. Tesseract Information: 0 : Trying to load native function "pixcmapGetExtremeValue" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetExtremeValue", function handle = 325458877. Tesseract Information: 0 : Trying to load native function "pixcmapGrayToColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGrayToColor", function handle = 325464607. Tesseract Information: 0 : Trying to load native function "pixcmapColorToGray" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapColorToGray", function handle = 325469397. Tesseract Information: 0 : Trying to load native function "pixcmapColorToGray" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapColorToGray", function handle = 325469397. Tesseract Information: 0 : Trying to load native function "pixcmapToRGBTable" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapToRGBTable", function handle = 325469742. Tesseract Information: 0 : Trying to load native function "pixcmapSerializeToMemory" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapSerializeToMemory", function handle = 325465572. Tesseract Information: 0 : Trying to load native function "pixcmapDeserializeFromMemory" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapDeserializeFromMemory", function handle = 325460637. Tesseract Information: 0 : Trying to load native function "pixcmapGammaTRC" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGammaTRC", function handle = 325470952. Tesseract Information: 0 : Trying to load native function "pixcmapContrastTRC" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapContrastTRC", function handle = 325462497. Tesseract Information: 0 : Trying to load native function "pixcmapShiftIntensity" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapShiftIntensity", function handle = 325460387. Tesseract Information: 0 : Current platform: x86 Tesseract Information: 0 : Custom search path is not defined, skipping. Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for 'libtesseract304.dll' on platform x86. Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'libtesseract304.dll' on platform x86. Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll"... Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll", handle = 329383936. Tesseract Information: 0 : Trying to load native function "TessBaseAPIAnalyseLayout" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIAnalyseLayout", function handle = 329706928. Tesseract Information: 0 : Trying to load native function "TessBaseAPIClear" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIClear", function handle = 329706560. Tesseract Information: 0 : Trying to load native function "TessBaseAPICreate" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPICreate", function handle = 329709952. Tesseract Information: 0 : Trying to load native function "TessBaseAPIDelete" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIDelete", function handle = 329709920. Tesseract Information: 0 : Trying to load native function "TessBaseAPIDetectOS" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIDetectOS", function handle = 329706368. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetBoolVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetBoolVariable", function handle = 329709264. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetDoubleVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetDoubleVariable", function handle = 329709184. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetHOCRText" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetHOCRText", function handle = 329706736. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetIntVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetIntVariable", function handle = 329709344. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetIterator" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetIterator", function handle = 329706784. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetPageSegMode" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetPageSegMode", function handle = 329707728. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetStringVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetStringVariable", function handle = 329709104. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetThresholdedImage" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetThresholdedImage", function handle = 329707312. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetUTF8Text" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetUTF8Text", function handle = 329706752. Tesseract Information: 0 : Trying to load native function "TessBaseAPIInit4" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIInit4", function handle = 329708688. Tesseract Information: 0 : Trying to load native function "TessBaseAPIMeanTextConf" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIMeanTextConf", function handle = 329706688. Tesseract Information: 0 : Trying to load native function "TessBaseAPIRecognize" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIRecognize", function handle = 329706912. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetDebugVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetDebugVariable", function handle = 329709424. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetImage2" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetImage2", function handle = 329707536. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetInputName" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetInputName", function handle = 329709840. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetPageSegMode" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetPageSegMode", function handle = 329707760. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetRectangle" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetRectangle", function handle = 329707440. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetVariable", function handle = 329709504. Tesseract Information: 0 : Trying to load native function "TessDeleteBlockList" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteBlockList", function handle = 329710688. Tesseract Information: 0 : Trying to load native function "TessDeleteIntArray" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteIntArray", function handle = 329710720. Tesseract Information: 0 : Trying to load native function "TessDeleteText" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteText", function handle = 329710784. Tesseract Information: 0 : Trying to load native function "TessDeleteTextArray" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteTextArray", function handle = 329710736. Tesseract Information: 0 : Trying to load native function "TessVersion" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessVersion", function handle = 329710800. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBaseline" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBaseline", function handle = 329705472. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBegin" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBegin", function handle = 329705760. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBlockType" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBlockType", function handle = 329705568. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBoundingBox" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBoundingBox", function handle = 329705616. Tesseract Information: 0 : Trying to load native function "TessPageIteratorCopy" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorCopy", function handle = 329705776. Tesseract Information: 0 : Trying to load native function "TessPageIteratorDelete" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorDelete", function handle = 329705808. Tesseract Information: 0 : Trying to load native function "TessPageIteratorGetBinaryImage" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorGetBinaryImage", function handle = 329705552. Tesseract Information: 0 : Trying to load native function "TessPageIteratorGetImage" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorGetImage", function handle = 329705520. Tesseract Information: 0 : Trying to load native function "TessPageIteratorIsAtBeginningOf" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorIsAtBeginningOf", function handle = 329705696. Tesseract Information: 0 : Trying to load native function "TessPageIteratorIsAtFinalElement" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorIsAtFinalElement", function handle = 329705664. Tesseract Information: 0 : Trying to load native function "TessPageIteratorNext" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorNext", function handle = 329705728. Tesseract Information: 0 : Trying to load native function "TessPageIteratorOrientation" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorOrientation", function handle = 329705440. Tesseract Information: 0 : Trying to load native function "TessResultIteratorCopy" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessResultIteratorCopy", function handle = 329705200. Tesseract Information: 0 : Trying to load native function "TessResultIteratorDelete" from the library with handle 329383936... -- Again the log is from one run (one attempt to execute the single line of code listed at the start of this post). Conclusion I am a very experienced programmer (have done win32 to cloud). I spent about 5 hours trying to get this to work without success. One thing I notice is that the log seems to end on TessResultIteratorDelete, and in some cases it starts to print the next line but stops midway (prints "Tesseract Information: 0 : Successfully", or "Tesseract Information: 0 : S"). Any help that you can offer would be greatly appreciated (the more specific the better). Thank you very much again. Best, Conrad My guess is that the actual tessdata path isn't what you expect. I'd strongly recommend using absolute paths, especially for ASP.NET applications, in this case Server.MapPath("~\tessdata") should do the trick. Note that's just of the top my head so the syntax may not be 100% correct 😉 On Fri, 17 Feb 2017, 16:56 Conrad-S<EMAIL_ADDRESS>wrote: Hi and thank you very much for your help. I have done everything that I possibly can think of to attempt to resolve a "Failed to initialise engine" error. I read post #277 https://github.com/charlesw/tesseract/issues/277 where the person stated that adjusting the path to tessdata fixed his issue. But he did not state the "adjustment" that he made. Line where the failure occurs using (var engine = new TesseractEngine (@"./tessdata", "eng", EngineMode.Default)) { Error "Failed to initialise tesseract engine.. See https://github.com/charlesw/tesseract/wiki/Error-1 for details."} Tesseract.TesseractException Stack Trace Data {System.Collections.ListDictionaryInternal} System.Collections.IDictionary {System.Collections.ListDictionaryInternal} HResult -2146233088 int HelpLink null string InnerException null System.Exception Message "Failed to initialise tesseract engine.. See https://github.com/charlesw/tesseract/wiki/Error-1 for details." string Source "Tesseract" string StackTrace " at Tesseract.TesseractEngine.Initialise(String datapath, String language, EngineMode engineMode, IEnumerable1 configFiles, IDictionary2 initialValues, Boolean setOnlyNonDebugVariables)\r\n at Tesseract.TesseractEngine..ctor(String datapath, String language, EngineMode engineMode, IEnumerable1 configFiles, IDictionary2 initialOptions, Boolean setOnlyNonDebugVariables)\r\n at Tesseract.TesseractEngine..ctor(String datapath, String language, EngineMode engineMode)\r\n at OCRTestOne.OCRTest.Go() in C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\Classes\OCRTest.cs:line 22" string TargetSite {Void Initialise(System.String, System.String, Tesseract.EngineMode, System.Collections.Generic.IEnumerable1[System.String], System.Collections.Generic.IDictionary2[System.String,System.Object], Boolean)} System.Reflection.MethodBase {System.Reflection.RuntimeMethodInfo} Development Environment Windows Server 2012 R2 (virtual machine running in Azure) Visual Studio 2015 (Version 14.0.25431.01 Update 3) Project Type: ASP.NET (also tried console app). Framework Target 452 (.NET Framework 4.5.2). Verified that Visual C++ 32-bit and 64-bit runtimes are both installed Nuget Package: Tesseract by Charles Weld 3.0.2 tessdata folder exists with English language files underneath (I added it). Files set to Copy Always. 32 and 64-bit assemblies (liblept172.dll and libtesseract304.dll) exist in the x86 and x64 folders and are set to Copy Always (they seem to be copied to the lib folder). Trace Turned on verbose mode tracing. The trace below is based on one execution attempt (I tried to attach a text file but the site said: "We don't support that file type"). Apologies for the long trace Tesseract Information: 0 : Current OS: Windows Tesseract Information: 0 : Current platform: x86 Tesseract Information: 0 : Custom search path is not defined, skipping. Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for 'liblept172.dll' on platform x86. Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'liblept172.dll' on platform x86. Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll"... Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll", handle = 325451776. Tesseract Information: 0 : Trying to load native function "pixaReadMultipageTiff" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaReadMultipageTiff", function handle = 325469402. Tesseract Information: 0 : Trying to load native function "pixaGetCount" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaGetCount", function handle = 325462092. Tesseract Information: 0 : Trying to load native function "pixaGetPix" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaGetPix", function handle = 325455972. Tesseract Information: 0 : Trying to load native function "pixaDestroy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixaDestroy", function handle = 325456927. Tesseract Information: 0 : Trying to load native function "pixCreate" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixCreate", function handle = 325465967. Tesseract Information: 0 : Trying to load native function "pixClone" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixClone", function handle = 325466612. Tesseract Information: 0 : Trying to load native function "pixDestroy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixDestroy", function handle = 325466382. Tesseract Information: 0 : Trying to load native function "pixGetWidth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetWidth", function handle = 325460622. Tesseract Information: 0 : Trying to load native function "pixGetHeight" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetHeight", function handle = 325460452. Tesseract Information: 0 : Trying to load native function "pixGetDepth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetDepth", function handle = 325464312. Tesseract Information: 0 : Trying to load native function "pixGetXRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetXRes", function handle = 325463887. Tesseract Information: 0 : Trying to load native function "pixGetYRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetYRes", function handle = 325463882. Tesseract Information: 0 : Trying to load native function "pixGetResolution" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetResolution", function handle = 325469492. Tesseract Information: 0 : Trying to load native function "pixGetWpl" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetWpl", function handle = 325466032. Tesseract Information: 0 : Trying to load native function "pixSetXRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetXRes", function handle = 325463907. Tesseract Information: 0 : Trying to load native function "pixSetYRes" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetYRes", function handle = 325463902. Tesseract Information: 0 : Trying to load native function "pixSetResolution" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetResolution", function handle = 325465697. Tesseract Information: 0 : Trying to load native function "pixScaleResolution" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixScaleResolution", function handle = 325467282. Tesseract Information: 0 : Trying to load native function "pixGetData" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetData", function handle = 325469052. Tesseract Information: 0 : Trying to load native function "pixGetInputFormat" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetInputFormat", function handle = 325467712. Tesseract Information: 0 : Trying to load native function "pixSetInputFormat" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetInputFormat", function handle = 325459102. Tesseract Information: 0 : Trying to load native function "pixEndianByteSwap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixEndianByteSwap", function handle = 325468587. Tesseract Information: 0 : Trying to load native function "pixRead" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixRead", function handle = 325467527. Tesseract Information: 0 : Trying to load native function "pixReadMemTiff" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixReadMemTiff", function handle = 325472137. Tesseract Information: 0 : Trying to load native function "pixWrite" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixWrite", function handle = 325466617. Tesseract Information: 0 : Trying to load native function "pixGetColormap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixGetColormap", function handle = 325464692. Tesseract Information: 0 : Trying to load native function "pixSetColormap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSetColormap", function handle = 325468467. Tesseract Information: 0 : Trying to load native function "pixDestroyColormap" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixDestroyColormap", function handle = 325465782. Tesseract Information: 0 : Trying to load native function "pixConvertRGBToGray" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixConvertRGBToGray", function handle = 325460777. Tesseract Information: 0 : Trying to load native function "pixDeskewGeneral" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixDeskewGeneral", function handle = 325460177. Tesseract Information: 0 : Trying to load native function "pixRotate" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixRotate", function handle = 325466312. Tesseract Information: 0 : Trying to load native function "pixRotateOrth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixRotateOrth", function handle = 325468177. Tesseract Information: 0 : Trying to load native function "pixOtsuAdaptiveThreshold" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixOtsuAdaptiveThreshold", function handle = 325468767. Tesseract Information: 0 : Trying to load native function "pixSauvolaBinarize" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSauvolaBinarize", function handle = 325463922. Tesseract Information: 0 : Trying to load native function "pixSauvolaBinarizeTiled" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixSauvolaBinarizeTiled", function handle = 325461412. Tesseract Information: 0 : Trying to load native function "pixScale" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixScale", function handle = 325466597. Tesseract Information: 0 : Trying to load native function "pixcmapCreate" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCreate", function handle = 325462397. Tesseract Information: 0 : Trying to load native function "pixcmapCreateRandom" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCreateRandom", function handle = 325460482. Tesseract Information: 0 : Trying to load native function "pixcmapCreateLinear" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCreateLinear", function handle = 325460392. Tesseract Information: 0 : Trying to load native function "pixcmapCopy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCopy", function handle = 325463212. Tesseract Information: 0 : Trying to load native function "pixcmapDestroy" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapDestroy", function handle = 325464252. Tesseract Information: 0 : Trying to load native function "pixcmapGetCount" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetCount", function handle = 325466562. Tesseract Information: 0 : Trying to load native function "pixcmapGetFreeCount" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetFreeCount", function handle = 325456837. Tesseract Information: 0 : Trying to load native function "pixcmapGetDepth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetDepth", function handle = 325468462. Tesseract Information: 0 : Trying to load native function "pixcmapGetMinDepth" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetMinDepth", function handle = 325462157. Tesseract Information: 0 : Trying to load native function "pixcmapClear" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapClear", function handle = 325460977. Tesseract Information: 0 : Trying to load native function "pixcmapAddColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddColor", function handle = 325466517. Tesseract Information: 0 : Trying to load native function "pixcmapAddNewColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddNewColor", function handle = 325459832. Tesseract Information: 0 : Trying to load native function "pixcmapAddNearestColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddNearestColor", function handle = 325464897. Tesseract Information: 0 : Trying to load native function "pixcmapUsableColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapUsableColor", function handle = 325468852. Tesseract Information: 0 : Trying to load native function "pixcmapAddBlackOrWhite" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapAddBlackOrWhite", function handle = 325460002. Tesseract Information: 0 : Trying to load native function "pixcmapSetBlackAndWhite" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapSetBlackAndWhite", function handle = 325456282. Tesseract Information: 0 : Trying to load native function "pixcmapGetColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetColor", function handle = 325466527. Tesseract Information: 0 : Trying to load native function "pixcmapGetColor32" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetColor32", function handle = 325458992. Tesseract Information: 0 : Trying to load native function "pixcmapResetColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapResetColor", function handle = 325457197. Tesseract Information: 0 : Trying to load native function "pixcmapGetIndex" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetIndex", function handle = 325466567. Tesseract Information: 0 : Trying to load native function "pixcmapHasColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapHasColor", function handle = 325458122. Tesseract Information: 0 : Trying to load native function "pixcmapCountGrayColors" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCountGrayColors", function handle = 325462272. Tesseract Information: 0 : Trying to load native function "pixcmapCountGrayColors" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapCountGrayColors", function handle = 325462272. Tesseract Information: 0 : Trying to load native function "pixcmapGetNearestIndex" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetNearestIndex", function handle = 325459557. Tesseract Information: 0 : Trying to load native function "pixcmapGetNearestGrayIndex" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetNearestGrayIndex", function handle = 325466522. Tesseract Information: 0 : Trying to load native function "pixcmapGetComponentRange" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetComponentRange", function handle = 325459077. Tesseract Information: 0 : Trying to load native function "pixcmapGetExtremeValue" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGetExtremeValue", function handle = 325458877. Tesseract Information: 0 : Trying to load native function "pixcmapGrayToColor" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGrayToColor", function handle = 325464607. Tesseract Information: 0 : Trying to load native function "pixcmapColorToGray" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapColorToGray", function handle = 325469397. Tesseract Information: 0 : Trying to load native function "pixcmapColorToGray" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapColorToGray", function handle = 325469397. Tesseract Information: 0 : Trying to load native function "pixcmapToRGBTable" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapToRGBTable", function handle = 325469742. Tesseract Information: 0 : Trying to load native function "pixcmapSerializeToMemory" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapSerializeToMemory", function handle = 325465572. Tesseract Information: 0 : Trying to load native function "pixcmapDeserializeFromMemory" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapDeserializeFromMemory", function handle = 325460637. Tesseract Information: 0 : Trying to load native function "pixcmapGammaTRC" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapGammaTRC", function handle = 325470952. Tesseract Information: 0 : Trying to load native function "pixcmapContrastTRC" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapContrastTRC", function handle = 325462497. Tesseract Information: 0 : Trying to load native function "pixcmapShiftIntensity" from the library with handle 325451776... Tesseract Information: 0 : Successfully loaded native function "pixcmapShiftIntensity", function handle = 325460387. Tesseract Information: 0 : Current platform: x86 Tesseract Information: 0 : Custom search path is not defined, skipping. Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for 'libtesseract304.dll' on platform x86. Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'libtesseract304.dll' on platform x86. Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll"... Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll", handle = 329383936. Tesseract Information: 0 : Trying to load native function "TessBaseAPIAnalyseLayout" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIAnalyseLayout", function handle = 329706928. Tesseract Information: 0 : Trying to load native function "TessBaseAPIClear" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIClear", function handle = 329706560. Tesseract Information: 0 : Trying to load native function "TessBaseAPICreate" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPICreate", function handle = 329709952. Tesseract Information: 0 : Trying to load native function "TessBaseAPIDelete" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIDelete", function handle = 329709920. Tesseract Information: 0 : Trying to load native function "TessBaseAPIDetectOS" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIDetectOS", function handle = 329706368. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetBoolVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetBoolVariable", function handle = 329709264. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetDoubleVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetDoubleVariable", function handle = 329709184. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetHOCRText" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetHOCRText", function handle = 329706736. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetIntVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetIntVariable", function handle = 329709344. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetIterator" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetIterator", function handle = 329706784. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetPageSegMode" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetPageSegMode", function handle = 329707728. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetStringVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetStringVariable", function handle = 329709104. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetThresholdedImage" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetThresholdedImage", function handle = 329707312. Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetUTF8Text" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetUTF8Text", function handle = 329706752. Tesseract Information: 0 : Trying to load native function "TessBaseAPIInit4" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIInit4", function handle = 329708688. Tesseract Information: 0 : Trying to load native function "TessBaseAPIMeanTextConf" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIMeanTextConf", function handle = 329706688. Tesseract Information: 0 : Trying to load native function "TessBaseAPIRecognize" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIRecognize", function handle = 329706912. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetDebugVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetDebugVariable", function handle = 329709424. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetImage2" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetImage2", function handle = 329707536. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetInputName" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetInputName", function handle = 329709840. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetPageSegMode" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetPageSegMode", function handle = 329707760. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetRectangle" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetRectangle", function handle = 329707440. Tesseract Information: 0 : Trying to load native function "TessBaseAPISetVariable" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetVariable", function handle = 329709504. Tesseract Information: 0 : Trying to load native function "TessDeleteBlockList" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteBlockList", function handle = 329710688. Tesseract Information: 0 : Trying to load native function "TessDeleteIntArray" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteIntArray", function handle = 329710720. Tesseract Information: 0 : Trying to load native function "TessDeleteText" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteText", function handle = 329710784. Tesseract Information: 0 : Trying to load native function "TessDeleteTextArray" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessDeleteTextArray", function handle = 329710736. Tesseract Information: 0 : Trying to load native function "TessVersion" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessVersion", function handle = 329710800. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBaseline" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBaseline", function handle = 329705472. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBegin" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBegin", function handle = 329705760. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBlockType" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBlockType", function handle = 329705568. Tesseract Information: 0 : Trying to load native function "TessPageIteratorBoundingBox" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBoundingBox", function handle = 329705616. Tesseract Information: 0 : Trying to load native function "TessPageIteratorCopy" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorCopy", function handle = 329705776. Tesseract Information: 0 : Trying to load native function "TessPageIteratorDelete" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorDelete", function handle = 329705808. Tesseract Information: 0 : Trying to load native function "TessPageIteratorGetBinaryImage" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorGetBinaryImage", function handle = 329705552. Tesseract Information: 0 : Trying to load native function "TessPageIteratorGetImage" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorGetImage", function handle = 329705520. Tesseract Information: 0 : Trying to load native function "TessPageIteratorIsAtBeginningOf" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorIsAtBeginningOf", function handle = 329705696. Tesseract Information: 0 : Trying to load native function "TessPageIteratorIsAtFinalElement" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorIsAtFinalElement", function handle = 329705664. Tesseract Information: 0 : Trying to load native function "TessPageIteratorNext" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorNext", function handle = 329705728. Tesseract Information: 0 : Trying to load native function "TessPageIteratorOrientation" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorOrientation", function handle = 329705440. Tesseract Information: 0 : Trying to load native function "TessResultIteratorCopy" from the library with handle 329383936... Tesseract Information: 0 : Successfully loaded native function "TessResultIteratorCopy", function handle = 329705200. Tesseract Information: 0 : Trying to load native function "TessResultIteratorDelete" from the library with handle 329383936... -- Again the log is from one run (one attempt to execute the single line of code listed at the start of this post). Conclusion I am a very experienced programmer (have done win32 to cloud). I spent about 5 hours trying to get this to work without success. One thing I notice is that the log seems to end on TessResultIteratorDelete, and in some cases it starts to print the next line but stops midway (prints "Tesseract Information: 0 : Successfully", or "Tesseract Information: 0 : S"). Any help that you can offer would be greatly appreciated (the more specific the better). Thank you very much again. Best, Conrad — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/charlesw/tesseract/issues/318, or mute the thread https://github.com/notifications/unsubscribe-auth/AAPzyPb7KW6g6m7pE5HhORKHByTo19EIks5rdTZ4gaJpZM4MD5r6 . Charles, Thank you for the fast response! It works! I had tried that before but changed it while trying other things. Below is a summary the issues (or gotchas) that I encountered and how resolved: Initially I couldn't get it to work so I experimented with settings. I had tried multiple versions of the Tesserat Nuget package. So I needed to double-check that the correct versions of the assemblies were in the x86 and x64 folders. I did this by downloading the latest source (in my case tesseract-3.0.2.zip) and getting the latest assemblies from the "..src\lib\TesseractOcr" x64 and x86 folders. The rest of the checklist is as follows: Make sure that the assemblies in the x86 and x64 are set to "Copy Always" (right-click the file in Visual Studio, open the properties dialog, and set the property). Make sure that the language files are in the Visual Studio project, match the Tesserat Nuget package version, and are set to "Copy Always". Use the following include statement and code to set the path to tessdata: using System.Web.Hosting; string tessPath = HostingEnvironment.MapPath(@"~/tessdata"); Make sure that the code includes: using System.Drawing; (otherwise an error occurs saying that a bitmap is being used by not referenced). Again, thank you very much for your help. This looks like a great library! I cannot wait to try it out!! Best, Conrad By the way is there documentation that lists functions in the library and their use? For example, I would like to open a PDF and read it, but don't know what function to use.. Thanks again! Most of the functions are documented using XML comments so VS should provide some explanation through the auto complete and it would be possible to generate some API doco from said comments. Otherwise check this issue tracker and/or stack overflow for anything more specific. In regards to pdf support, that's currently only supported in the prerelease version and only supports generation of pdf (image + searchable text) not reading pdf. In your case you'll need to find another library to load the PDF and extract any contained image for processing. On Fri, 17 Feb 2017, 17:56 Conrad-S<EMAIL_ADDRESS>wrote: By the way is there documentation that lists functions in the library and their use? For example, I would like to open a PDF and read it, but don't know what function to use.. Thanks again! — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/charlesw/tesseract/issues/318#issuecomment-280570159, or mute the thread https://github.com/notifications/unsubscribe-auth/AAPzyAG4WFMGMAxUHRkabyu-TqpfBMXDks5rdUSxgaJpZM4MD5r6 . Thanks. Nothing is documented in Visual Studio, but auto-complete (intellisense) does work. I can probably use object browser to view the methods, but it would be good to have a document somewhere with the list of methods and their use (also helps to understand overall functionality). I didn't realize that it cannot extract images from pdf's; that was my main purpose. Do you know of a .NET library that can extract the images? Thanks again!! Mmm the XML document file mustn't be being included property in the nuget package. Anyway in regards to handling PDFs I'm not really sure what's available these days for .net. That being said, as of a couple years ago, a good stable open source option was itextsharp though it's no longer free for commercial use. However there are still copies of the code available before the license was changed but I'm not aware of anyone actively maintaining these forks. On Fri, 17 Feb 2017, 18:18 Conrad-S<EMAIL_ADDRESS>wrote: Thanks. Nothing is documented in Visual Studio, but auto-complete (intellisense) does work. I can probably use object browser to view the methods, but it would be good to have a document somewhere with the list of methods and their use (also helps to understand overall functionality). I didn't realize that it cannot extract images from pdf's; that was my main purpose. Do you know of a .NET library that can extract the images? Thanks again!! — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/charlesw/tesseract/issues/318#issuecomment-280573566, or mute the thread https://github.com/notifications/unsubscribe-auth/AAPzyKgNDDkoOeg9G8noVzipga8E7EgSks5rdUm1gaJpZM4MD5r6 . You can get the old version from nuGet https://www.nuget.org/packages/iTextSharp-LGPL/ Thanks again! This is great software. By the way I notice that it always seems to load the 32-bit version. I even tried setting the MapPath. Tesseract.TesseractEnviornment.CustomSearchPath = HostingEnvironment.MapPath(@"~/x64"); Do you know how to convince it to load the 64-bit DLL? Tesseract Information: 0 : Current OS: Windows Tesseract Information: 0 : Current platform: x86 Tesseract Information: 0 : Custom search path is not defined, skipping. Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\vs\99b032d6\fb07eb9\assembly\dl3\8760118f\00342a46_d765d101' for 'liblept172.dll' on platform x86. Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'liblept172.dll' on platform x86. Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll"... Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll", handle = 369360896. I'll answer my own question. Got it to work (Visual Studio 2015; ASP.NET MVC; .NET 452): Do NOT set Tesseract.TesseractEnviornment.CustomSearchPath Select Debug -> Options -> Projects and Solutions -> Web Projects. Select User the 64 bit version of IIS Express for web sites and projects. Set the debug configuration to x64 (Configuration Manager next to the Debug dropdown) Restart Visual Studio, open the solution and run. You may also want to set up a trace file so you can see that it loaded the 64-bit assembly: <system.diagnostics> </system.diagnostics> Don't forget to turn off logging after you are finished (performance). Hope this helps. Fail with asp.net core, i've tried exactly path, but still throw exception failed to initalise... Please help me After banging my head against my desk for 30min, I finally realized the path needs to be to the FOLDER containing the .traineddata files (ie: "c:\tessdata"), not the path to one of the .traineddata files (ie "c:\tessdata\eng.traineddata"). Doh! Hopefully this saves someone's desk (and head). Fail with asp.net core, i've tried exactly path, but still throw exception failed to initalise... Please help me Try using tessdata_fast rather then tessdata version. Anyone knows how to read data from table in image with this library?
2025-04-01T06:38:10.769639
2024-06-24T17:16:17
2370727861
{ "authors": [ "davidhardy", "ericjbohm" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4606", "repo": "charmplusplus/charm", "url": "https://github.com/charmplusplus/charm/pull/3829" }
gharchive/pull-request
bugfix: remove stray endif, revise comments to match Remove stray endif, revise comments to match. Closes issue #3828 On Summit now I can build pamilrts successfully.
2025-04-01T06:38:10.797234
2017-02-04T19:11:18
205369998
{ "authors": [ "etimberg", "frlinw", "nutanlade" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4607", "repo": "chartjs/Chart.js", "url": "https://github.com/chartjs/Chart.js/issues/3868" }
gharchive/issue
Issue in styling legend in Pie charts Hi All, I'm trying to style my pie chart as first image below. But with below code I'm getting as shown in second image. Even I tried the solution as mentioned here 2253. Then no legends are displayed. Also I don't want white border between pies. I'm using ChartJs 2.4.0 version. Please help me in resolving the issue. var legend = chartInstance.generateLegend(); document.getElementById("legend").innerHTML = legend; @nutanlade to get the legend to look like the top image, you're going to need to use some CSS to change the <li> elements in the legend. To remove the border: data: { datasets: [{ borderWidth: 0 }] } Thanks @etimberg for your quick reply. I'm using angular js code for pie chart. Below is my code for generating pie chart. I'm not able to apply your code to remove border. Is there any default global option to disable? <canvas id="monthlyExpensesChart" class="chart chart-pie" chart-data="$ctrl.communitytotalCostArray" chart-labels="$ctrl.communityCategoryNamesArray" chart-options="$ctrl.options" chart-legend="true" height="250" chart-colors="$ctrl.colors"> </canvas> Also for legend now it showing like square but color is black. How can I get pie colors to populate legend labels? @nutanlade the global border option would be Chart.defaults.global.elements.arc.borderWidth For the legend, you would need to style the <li> elements with CSS. You would probably want to set the first one to the first color. I think something like the following would work. #legend > li:nth-child(0) { color: red; } #legend > li:nth-child(1) { color: yellow; } #legend > li:nth-child(2) { color: blue; } Controller: Create an object of legend items with color & label : # Chart conf scope.chart = [] scope.chart.labels = ['Food', 'Fuel'] scope.chart.colors = ['red', 'yellow'] # Legend scope.legend = [] for label, i in scope.chart.labels scope.legend[i] = title: label color: scope.chart.colors[i] (it's coffee script, use http://js2.coffee/ if you want a translation) HTML: Loop on legend items to display them and you can do whatever you want with styles <div class="widget-body"> <div class="row"> <div class="col-md-12 col-lg-7"> <div class="widget-pie-chart"> <canvas beautiful-chart chart-type="::chart.type" chart-options="::chart.options" chart-datasets="::chart.datasets" chart-labels="chart.labels" chart-data="chart.data"></canvas> <canvas id="monthlyExpensesChart" class="chart chart-pie" chart-data="communitytotalCostArray" chart-labels="communityCategoryNamesArray" chart-options="options" height="250" chart-colors="colors"></canvas> </div> </div> <div class="col-md-12 col-lg-5"> <div class="widget-pie-legend"> <a ng-repeat="item in ::legend" href=""> <div class="pie-legend-item"> <span class="legend-item-box" ng-style="::{'color': item.color}"></span> <span class="legend-item-label">{{::item.title}}</span> </div> </a> </div> </div> </div> </div> Result You should see something like that Thanks @etimberg and @frlinw . Issue resolved. I disabled tooltips but while hovering on chart color is changing which I don't want. I tried below option but didn't worked. showTooltips: false @nutanlade to turn off hover you can do options: { hover: { mode: false } } Thanks @etimberg. Issue resolved.
2025-04-01T06:38:10.804830
2020-10-09T08:26:26
717957146
{ "authors": [ "dcyou", "etimberg", "kurkle" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4608", "repo": "chartjs/Chart.js", "url": "https://github.com/chartjs/Chart.js/issues/7866" }
gharchive/issue
Circular gradiant Hi all, (thanks for that awesome lib) I am using Chart.js v3 and I am trying to create a gradient My fiddle is here: https://jsfiddle.net/1wfd2ztx/1/ What I would like to do is to start at top center with the green and finish with the red at top center too with a gradient And ++ draw with red under the green if the percent is more than 100 @dcyou this is a very interesting question! A radial gradient is probably not the right solution here since it changes colour as you move out from a center point. You probably want something more akin to a SweepGradient but as far as I am aware, nothing like this exists natively. If this were WebGL, a custom shader would be pretty trivial to write. The best idea I have at the moment is to draw the colour pattern you want on another canvas (handling resize accordingly) then create a CanvasPattern from that canvas and set that as the background colour of the dataset. This is quite intriguing. I'll try and prototype something out and see where I get. @dcyou got a working implementation. Generates a CanvasPattern using webgl then renders that as the background image. I don't think the plugin part is entirely correct. I suspect on resize it may be wrong, but have not attempted to test it. https://jsfiddle.net/6d0gv7w9/ awesome! @etimberg thanks a lot I will have a look Played around with @etimberg's fiddle and applied some changes: https://jsfiddle.net/x1t38f7v/ Precicely, multiplied the webgl canvas size by devicePixelRatio and changed the plugin implementation to beforeDatasetsUpdate hook. very interesting too question: do you know why at bottom we have the blank? do you know why at bottom we have the blank? I'm guessing you have a devicePixelRatio greater than 1? Does that appear in the fiddle where I multiplied the canvas size with it? i actually don't think devicePixelRatio needs to be used. chart.chartArea will already include the devicePixelRatio since we up the size of the canvas by it inside of https://github.com/chartjs/Chart.js/blob/531d4c09bf5778274d362eaf1f5dd17a9a08ec7b/src/helpers/helpers.dom.js#L148-L149
2025-04-01T06:38:10.806595
2024-02-14T01:16:06
2133386111
{ "authors": [ "LeeLenaleee", "zieka" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4609", "repo": "chartjs/Chart.js", "url": "https://github.com/chartjs/Chart.js/pull/11669" }
gharchive/pull-request
refactor: replace Math.pow in favor of the ** operator friendly PR to improve readability Use this link to re-run the recipe: https://app.moderne.io/recipes/org.openrewrite.codemods.cleanup.PreferExponentiationOperator?organizationId=UG9wdWxhciBKYXZhU2NyaXB0 I find the Math.pow more readable myself and would be against this change
2025-04-01T06:38:10.828935
2022-06-09T12:00:41
1266025892
{ "authors": [ "bensmiley", "letzter" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4610", "repo": "chat-sdk/chat-sdk-android", "url": "https://github.com/chat-sdk/chat-sdk-android/issues/734" }
gharchive/issue
Unable to login with chat sdk Is the bug present in the demo Chat SDK project? Yes What modifications have you made to the Chat SDK? None Android Version: 11 Steps taken to reproduce the problem: Insert username and password and press register Expected result: Register user and enter profile screen Actual result: Authentication progress bar keeps spinning "forever" Comments: I have been trying to integrate this chat sdk on my app for some time, but cant seem to go through the authentication Took a look at logcat and it seems that something is failing. I even tried to register in the example android app, but it doesn't go through. The logs obtained are the ones below: 2022-06-09 12:47:26.960 10270-10318/sdk.chat.android.live W/System: Ignoring header X-Firebase-Locale because its value was null. 2022-06-09 12:47:27.495 10270-10318/sdk.chat.android.live W/System: Ignoring header X-Firebase-Locale because its value was null. 2022-06-09 12:47:27.696 10270-10318/sdk.chat.android.live D/FirebaseAuth: Notifying id token listeners about user ( bTlcQnKU0AQ1ClDop8lhrSRukS63 ). 2022-06-09 12:47:27.696 10270-10318/sdk.chat.android.live D/FirebaseAuth: Notifying auth state listeners about user ( bTlcQnKU0AQ1ClDop8lhrSRukS63 ). 2022-06-09 12:47:27.718 10270-10577/sdk.chat.android.live I/greenDAO: Creating tables for schema version 20 2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: 2022-06-09 12:47:27 [RxComputationThreadPool-1] sdk.chat.core.dao.DaoCore.openDB() 2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: INFO: Database: e92eded1acf5e51967b47107e633e719 setup correctly 2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: 2022-06-09 12:47:27 [RxComputationThreadPool-1] sdk.chat.firebase.adapter.FirebaseAuthenticationHandler.lambda$authenticateWithUser$9$FirebaseAuthenticationHandler() 2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: INFO: authenticateWithUser: bTlcQnKU0AQ1ClDop8lhrSRukS63 2022-06-09 12:48:08.817 10270-10631/sdk.chat.android.live V/FA: Connecting to remote service 2022-06-09 12:48:08.841 10270-10631/sdk.chat.android.live D/FA: Connected to remote service 2022-06-09 12:48:08.842 10270-10631/sdk.chat.android.live V/FA: Processing queued up service tasks: 1 2022-06-09 12:48:08.854 10270-10627/sdk.chat.android.live W/System: Ignoring header If-None-Match because its value was null. 2022-06-09 12:48:13.859 10270-10631/sdk.chat.android.live V/FA: Inactivity, disconnecting from the service Thanks in advance I checked out the current version from Github and I can log in... Have you tried this without connecting it to your Firebase account? Have you enabled the login methods you want to use from the Firebase console? @letzter I need some more information because I tried it myself and it worked just fine. Can you send me a video? Here goes the video showing the process that I'm following. Is it possible that I'm having this problem because I'm running the app on an emulator from android studio instead of a real device, or should it it work as well? chat-sdk.webm.zip It is possible. I never test on the emulator. Do you have Google Play services installed? If that were missing it Wouk's cause issues. @bensmiley the issue was with the emulator. I created a new one in which it all worked out well. Thank you very much for the support
2025-04-01T06:38:10.843492
2019-04-19T22:29:40
435321626
{ "authors": [ "Ebioro", "chatch" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4611", "repo": "chatch/stellar-hd-wallet", "url": "https://github.com/chatch/stellar-hd-wallet/issues/13" }
gharchive/issue
update stellar-base dependency Tried to use the wallet for an angular 6 project and could not build it because the angular cli started complaining about the crypto library. I then found out that wallet is currently using the stellar-base 0.11. The guys from stellar have removed the crypto library in version 0.13 of the stellar-base sdk which should have fixed the issue. More details can be found here thanks @Ebioro i've updated and published new release 0.10.0 you are welcome! I had an open question on stellar stack exchange about this (under yulemata username) I will add there that you guys have updated it and publish a new release!
2025-04-01T06:38:10.913619
2020-05-25T18:34:38
624436697
{ "authors": [ "JPinkney", "amisevsk", "davidfestal", "sleshchenko" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4612", "repo": "che-incubator/che-workspace-operator", "url": "https://github.com/che-incubator/che-workspace-operator/pull/83" }
gharchive/pull-request
Make operator installable via olm Signed-off-by: Josh Pinkney<EMAIL_ADDRESS>What does this PR do? This PR prepares everything needed for olm in the older olm format. The Makefile has been modified so that it is easy to load everything onto the cluster and so that you can see che-workspace-operator in the OperatorHub catalog. The csv was made with make gen_csv except for a few custom things like the alm-examples What issues does this PR fix or reference? Is it tested? How? Deploying the controller with olm In order to deploy the controller you need to first create the olm image, push it to the application registry on quay and then deploy the crd registry to your cluster. You can do this by using the Makefile olm_full_start rule: make olm_full_start Before doing this you need to set environment variables QUAY_USERNAME, QUAY_PASSWORD, QUAY_NAMESPACE (most likely this is the same as QUAY_USERNAME), CATALOG_IMAGE You might have to manually create the application registry on quay and set it to public. To do this, use + in the top right hand corner of quay and click new repository. Then instead of having a Container Image Repository set that dropdown to Application repository and make sure it's public. The repostory name you choose for the application registry must be the same name as you have set in the CATALOG_IMAGE. If you already have it deployed to the quay application registry then you can just use make olm_start to deploy the crd registry to your cluster To remove the crd registry use make olm_uninstall cc: @davidfestal since you're familiar with deploying via OLM. Why not use the new bundle format as a source, and just export the index in the old format when needed ? https://github.com/operator-framework/operator-registry/blob/master/docs/design/opm-tooling.md#export /retest Yeah at this point I think I'm going to close this PR and then we can transfer the catalogsource to the web-terminal-operator repo
2025-04-01T06:38:11.051626
2016-09-24T21:31:01
179052989
{ "authors": [ "JasonMorgan" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4613", "repo": "chef-cookbooks/powershell", "url": "https://github.com/chef-cookbooks/powershell/pull/98" }
gharchive/pull-request
removed windows_reboot and windows::reboot_handler. Description Drops the windows_reboot resource and the windows::reboot_handler recipe. The no longer exist in the windows cookbook. Issues Resolved None that I know of. Check List [ ] All tests pass. See https://github.com/chef-cookbooks/community_cookbook_documentation/blob/master/TESTING.MD [x] New functionality includes testing. [x] New functionality has been documented in the README if applicable [x] The CLA has been signed. See https://github.com/chef-cookbooks/community_cookbook_documentation/blob/master/CONTRIBUTING.MD That resource and recipe has been dropped from the windows cookbook. replace windows_reboot with the reboot resource. I didn't actually intend to have that second commit added into this PR. Although that being said it is related and fixes the problem of the recipe not actually allowing a reboot to occur. I was going to make a second PR for it after this got committed. Updated to address @smurawski comment.
2025-04-01T06:38:11.055022
2015-05-07T21:06:36
74127684
{ "authors": [ "RoboticCheese", "tas50" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4614", "repo": "chef-cookbooks/windows", "url": "https://github.com/chef-cookbooks/windows/pull/198" }
gharchive/pull-request
Make auto_run's root registry key configurable I recently found myself working on a cookbook for an app that only installs on a per-user basis, putting all its registry entries under the HKEY_CURRENT_USER root key instead of HKEY_LOCAL_MACHINE. This change adds a root property to the windows_auto_run resource, maintaining the current behavior as a default and allowing support for user-level applications as well. Signed-off-by: Jonathan Hartman<EMAIL_ADDRESS> @RoboticCheese If you want to rebase this against master it would be nice to get merged in @RoboticCheese Ping on a rebase here. I'd love to pull this in before we move this into chef-client so we can pull in this new functionality as well. Thanks for the reminder about this one. Rebased and updated. Seems to work again, though my Windows is pretty atrophied RN 🤞 Describing test::autorun Context windows_auto_run [+] does not auto-run Notepad for the machine 916ms [+] auto-runs Wordpad for the current user 95ms Tests completed in 1.01s Passed: 2 Failed: 0 Skipped: 0 Pending: 0 Finished verifying <autorun-windows-10> (0m20.17s). Thanks. I'll get this addition in when we move this into chef with Chef 14
2025-04-01T06:38:11.057728
2018-05-05T00:04:01
320461982
{ "authors": [ "derekgroh", "tas50" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4615", "repo": "chef-cookbooks/windows_firewall", "url": "https://github.com/chef-cookbooks/windows_firewall/pull/27" }
gharchive/pull-request
update readme to match resource Description Updates documentation to match the resource rule Issues Resolved N/A Check List [X] All tests pass. See https://github.com/chef-cookbooks/community_cookbook_documentation/blob/master/TESTING.MD [N/A] New functionality includes testing. [X] New functionality has been documented in the README if applicable Thanks
2025-04-01T06:38:11.064872
2023-10-19T05:00:24
1951228346
{ "authors": [ "AadeshNichite" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4616", "repo": "chef/automate", "url": "https://github.com/chef/automate/pull/8269" }
gharchive/pull-request
Moved protractor from dependency to dev-dependency :nut_and_bolt: Description: What code changed, and why? Signed-off-by: AadeshNichite<EMAIL_ADDRESS> Moved protractor from dependency to dev-dependency :chains: Related Resources https://chefio.atlassian.net/browse/CHEF-7042 :+1: Definition of Done :athletic_shoe: How to Build and Test the Change :white_check_mark: Checklist All PRs must tick these: [ ] I have read the CONTRIBUTING document. [ ] All commits signed-off for the Developer Certification of Origin. With occasional exceptions, all PRs from Progress employees must tick these: [ ] Is the code clear? (complicated code or lots of comments--subdivide and use well-named methods, meaningful variable names, etc.) [ ] Consistency checked? (user notifications, user prompts, visual patterns, code patterns, variable names) [ ] Repeated code blocks eliminated? (adapt and reuse existing components, blocks, functions, etc.) [ ] Spelling, grammar, typos checked? (at a minimum use make spell in any component directory) [ ] Code well-formatted? (indents, line breaks, etc. improve rather than hinder readability) All PRs from Progress employees should tick these if appropriate: [ ] Tests added/updated? (all new code needs new tests) [ ] Docs added/updated? (all customer-facing changes) Please add a note next to any checkbox above if you are NOT ticking it. :camera: Screenshots, if applicable No need to work this, input from @arunjn-progress.
2025-04-01T06:38:11.066296
2015-08-28T22:46:35
103807706
{ "authors": [ "danielsdeleo" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4617", "repo": "chef/chef-dk", "url": "https://github.com/chef/chef-dk/pull/491" }
gharchive/pull-request
chef clean-policy-revisions command Adds a subcommand to delete unused policy revisions, and fixes some minor issues with the service class discovered in manual integration testing. CLI output is like this: bundle exec bin/chef clean-policy-revisions -c ~/.chef/localserver.rb DELETE aar 37f9b658cdd1d9319bac8920581723efcc2014304b5f3827ee0779e10ffbdcc9 DELETE aar 5c5922d9febb3ee758c60b0bdd423185c7a4e88450c4fdf0cf90450cfcbfe88a DELETE aar 7a812b7e7570b93b6672e8befab0423be68f8b44109adb5eb0d3cd9a044504c8 DELETE aar d81e80ae9bb9778e8c4b7652d29b11d2111e763a840d0cadb34b46a8b2ca4347 DELETE aar e67f1b04a055017e7de02acb166e2ab18b4682313f381fae6e2fa8f4bc7db4ae DELETE jenkins 613f803bdd035d574df7fa6da525b38df45a74ca82b38b79655efed8a189e073 DELETE jenkins 6fe753184c8946052d3231bb4212116df28d89a3a5f7ae52832ad408419dd5eb DELETE jenkins 9c2cae07a5c68ae0ab7d06165e90ded1325a8b0ca94d78cfc2efa4dcfea36244 DELETE jenkins cc1a0801e75df1d1ea5b0d2c71ba7d31c539423b81478f65e6388b9ee415ad87 DELETE jenkins cec985e916690e82c792114c5b10a4e4694981804bcbc5ff89556e6709c28ec1
2025-04-01T06:38:11.073863
2021-03-18T22:17:52
835317078
{ "authors": [ "tas50" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4618", "repo": "chef/chef", "url": "https://github.com/chef/chef/issues/11206" }
gharchive/issue
Test Note: GitHub integration record type. Please leave out any sensitive information.Job to be Done: As an , I want to So that <reason for needing job/benefit> Description: Acceptance Criteria: Aha! Link: https://chef.aha.io/features/INFRA-1 sdfsadfsdf
2025-04-01T06:38:11.075316
2017-02-16T05:09:01
208018502
{ "authors": [ "coderanger", "tas50" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4619", "repo": "chef/chef", "url": "https://github.com/chef/chef/issues/5820" }
gharchive/issue
Use a real KDF for data bag encryption secrets Currently we use a single round of SHA256 for deriving the actual AES key. This is okay, but it would be better to use a real-er key derivation function some day like PBKDF2 or if we want to be really fancy, Argon2. This would require a new encryptor version most likely so it will probably take a while to get this in. Just for the sake of prioritization I can't think of an attack where this would matter so probably can put it off until such time as we're making a v4 encryption for another reason and just slide it in for extra crypto points. I'm going to close this out at this point since the future of secrets management in Chef Infra Client is not encrypted data bags.
2025-04-01T06:38:11.103677
2023-07-11T14:22:13
1799056362
{ "authors": [ "enoch1025", "ibasaisaac" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4620", "repo": "cheng-01037/Self-supervised-Fewshot-Medical-Image-Segmentation", "url": "https://github.com/cheng-01037/Self-supervised-Fewshot-Medical-Image-Segmentation/issues/39" }
gharchive/issue
I want to run a single category dataset. How can I set it to a single category I want to run a single category dataset. How can I set it to a single category please tell me, thank eyeryone!!!!!!!! ./examples/train_ssl_abdominal_ct.sh train_SABS_Superpix_lbgroup0_scale_MIDDLE_vfold0 WARNING - root - Changed type of config entry "min_fg_data" from str to int INFO - mySSL - Running command 'main' INFO - mySSL - Started run with ID "12" INFO - main - ###### Create model ###### NETWORK: Using ms-coco initialization INFO - main - ###### Load data ###### INFO - main - ###### Labels excluded in training : [2, 3] ###### INFO - main - ###### Unseen labels evaluated in testing: [2, 3] ###### Dataset: the following classes has been excluded [2, 3]###### DEBUG_DATASET CT_STATS NORMALIZED MEAN 0.2503295103060257 STD 0.22380904778564215 Initial scans loaded: ['7', '8', '9', '10', '11', '12'] INFO - main - ###### Set optimizer ###### INFO - main - ###### Training ###### INFO - main - ###### This is epoch 0 of 100 epoches ###### /home/wangyixiong/anaconda3/envs/self/lib/python3.7/site-packages/torch/nn/functional.py:2973: UserWarning: Default upsampling behavior when mode=bilinear is changed to align_corners=False since 0.4.0. Please specify align_corners=True if the old behavior is desired. See the documentation of nn.Upsample for details. "See the documentation of nn.Upsample for details.".format(mode)) Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip Faulty batch detected, skip ^CWARNING - mySSL - Aborted after 0:00:20! Traceback (most recent call last): File "training.py", line 30, in @ex.automain File "/home/wangyixiong/anaconda3/envs/self/lib/python3.7/site-packages/sacred/experiment.py", line 140, in automain self.run_commandline() File "/home/wangyixiong/anaconda3/envs/self/lib/python3.7/site-packages/sacred/experiment.py", line 263, in run_commandline return self.run(cmd_name, config_updates, nam I am also getting "Faulty batch detected, skip". How to fix this?
2025-04-01T06:38:11.169365
2018-05-29T19:02:19
327447688
{ "authors": [ "Ellmen", "ManuKle", "aguynamedben", "amilajack", "vikr01" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4624", "repo": "chentsulin/electron-react-boilerplate", "url": "https://github.com/chentsulin/electron-react-boilerplate/issues/1573" }
gharchive/issue
Too many config files Including configs for things like flow, vscode, and travis adds a lot of mess to clean up for people who don't use those things. Since this is the react electron repo, I feel like it should be stripped down a little. Which configs should we get rid of? Just registering another opinion... I think the flow configs are good because they provide an opinionated best practice, which is what boilerplate is for. I think the editor-specific configs could either be removed (because everybody uses a different editor, and your choise editor isn't necessarily related to the project ecosystem), or it should be an explicit policy that other editor configs are welcomed and maybe put them in editor_configs/ with a README file in that directory explaining the policy. I use Emacs, but VS Code seems to be an emerging norm in the JS community so it's probably helpful to have a config laying around. Also, wouldn't removing yarn.lock make it pretty difficult to debug issues people are having with the project? yarn.lock in projects I've worked on is always committed to make sure that people are using the same dependencies without having to continually ask everybody to list which version they have installed locally. In general my opinion is: It's not that hard to remove config files, and they rarely get in the way. You're going to have to customize this repo for your project anyways. git rming or editing a few config files you don't find helpful is much easier than being a newbie and not having access to sensible config files. For example, being new to the whole ecosystem it would be very hard to create a Flow config from scratch. The purpose of boilerplate is to help people get going in an ecosystem quickly. Example config files, even if they're not setup out of the box the way you'd like, are very helpful as it's difficult to come up with from scratch when you're new to an ecosystem. Config files are a good opportunity to set opinionated best practices, which is especially useful in boilerplate projects. These best practices reflect norms in a community. Boilerplate seems like the BEST place for config files. The cost of removing/modifying config to fit your preferences is much lower than the cost of not having ready access to example configs. I like the idea of an editor_configs and I can appreciate that having sample code makes it easier for a newcomer to get started. The main yarn.lock is certainly important - the one I was talking about is in the app directory and is empty (there are no dependencies in that directory's package.json) I would argue that Flow isn't actually a best practice and integrating it so deeply makes this project less flexible. Skype, GitHub, and Slack are all using TypeScript. The problem with enforcing a certain stack is that when any part of it becomes obsolete, the whole project does. It kind of depends on if this project is supposed to be a pure electron-react boilerplate or if it's a "cool stack" that runs on electron and react. Personally, I was just looking for a clean way to structure an application that uses the two technologies. Cool, those opinions make sense too. I'm just curious for my own education, are you using this repo with TypeScript? I've just been using Flow since it was already in this project... but I'm not partial to it. Was it easy to get TypeScript working with this? Whether this project is best as "cool stack" or "basic stack", I could see a change from Flow to TypeScript being a reasonable opinionated change. Coming from non-JavaScript land, there are already a lot of other opinions built-in (Webpack, ES6, testing framework, Redux). I'd like to see "cool stack". I'm a Rails fan and have seen the "cool stack" decision go the undesirable way (i.e. CoffeeScript), but overall I like things to be opinionated and bundled (Rails for Ruby) vs. piece-together-your-own-together (i.e. Flask for Python). Just opinions though. Good discussion. You can find my typescript fork under notable forks in the wiki. It's based on version 13.3 and I'll try to update soon with the latest commits. Regarding the configs, I kinda like the idea of putting them in a separate folder. But I wouldn't delete them, as they are useful to build upon. @ManuKle very cool! In #1686, we moved webpack configs to a configs folder. That drops quite a few files from the home directory.
2025-04-01T06:38:11.171965
2021-10-21T22:15:19
1032985260
{ "authors": [ "cheqianh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4625", "repo": "cheqianh/ion-hive-serde", "url": "https://github.com/cheqianh/ion-hive-serde/pull/2" }
gharchive/pull-request
Adds support to Hadoop compression codec. Issue #, if available: Description of changes: By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Sure I think it's a good idea to add more comments. It's also a good opportunity to look into other serde package and figure out how/when the APIs are called. Discussed offline, Added some comments. tested locally it worked well for gzip. Added links to the source. I'll do an overall testing again before sending this PR to the official repo.
2025-04-01T06:38:11.177960
2018-07-02T22:33:35
337685924
{ "authors": [ "KFlash", "g-plane" ], "license": "isc", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4627", "repo": "cherow/cherow", "url": "https://github.com/cherow/cherow/issues/179" }
gharchive/issue
Don't create PR if you can't finish them To avoid too much noise on this repo, don't open PR's if you can't or will finish them. At the moment we have 5 open PR. One of this PR have been inactive for the last 15 days - https://github.com/cherow/cherow/pull/153 We also have two PR - closed now - for string literals and ecmaVersion option. This also make up some noise on the repo. Just be 100% sure before you open a PR. Also be sure that you get reviewers with knowledge enough to understand your way of thinking and the code flow. https://github.com/cherow/cherow/pull/175 is my last PR. I will from now on push directly due to lack of reviewers and too keep the repo noise down. cc / @fkleuver @g-plane @Aladdin-ADD I really don't recommend to push directly. I have pushed directly from start. As long as I'm only dealing with src folder there is no issue. I still recommend to open PR. Once you opened a PR and there are no any reviews in several days, you can merge by yourself. The issue is that somneone open PR and 1) never finish them or 2) never merge them
2025-04-01T06:38:11.198174
2024-10-24T15:35:21
2611956382
{ "authors": [ "Poukpalaova", "gonu2024", "nhathoangfoto" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4628", "repo": "chflame163/ComfyUI_LayerStyle", "url": "https://github.com/chflame163/ComfyUI_LayerStyle/issues/372" }
gharchive/issue
Unable to pick colors in the ColorPicker node. I can't select any color other than the default color when loading the node. Normally, when I click the mouse cursor on the color, a color palette pops up, and I just need to move the cursor to the image and pick the color. But today, when I click on the color, nothing happens. I'm having the same problem. Did you solve it? Can someone help me? The panel does not appear even though my browser settings do not disable pop-ups. I have the same problem. In an new install, the color pop screen work flawless. After installing custom nodes, it stop working. It's probably an incompatibility with one of them, but wich one...?
2025-04-01T06:38:11.199772
2021-07-03T00:46:09
936132921
{ "authors": [ "chhoumann", "firinael" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4629", "repo": "chhoumann/quickadd", "url": "https://github.com/chhoumann/quickadd/issues/45" }
gharchive/issue
cancelling a template add brings up an error Hi, just started trying out the plugin, and ran into a curious behaviour. Whenever I press ESC after selecting a template to add, I get this error: And this only happens with the add template function, as far as I've noticed. Thank you! Fixed in the latest version. :)
2025-04-01T06:38:11.201528
2015-07-06T18:05:13
93338824
{ "authors": [ "chieffancypants", "ittayd" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4630", "repo": "chieffancypants/angular-hotkeys", "url": "https://github.com/chieffancypants/angular-hotkeys/issues/160" }
gharchive/issue
allow to provide a separate key combo for documentation it seems that for modifier+key combinations, the event triggers with the english key, even if the layout is in another language (at least for me, using english+hebrew layouts). this seems inherent in javascript. while the best solution is for the framework to map the non english characters to their english equivalent when handling key presses, it seems this is tricky (if possible). an alternative is to just allow to pass a key combination that will be used in the documentation. so while binding to alt+t, the documentation will show alt+א I'm not in control of this, as all hotkey-specific binding is handled by mousetrap. You may want to open a ticket there I was suggesting that an optional parameter will be allowed and that will be used for documentation. So I call with both alt+t and alt+א, you pass alt+t to mousetrap and show alt+א in the documentation ('?')
2025-04-01T06:38:11.236878
2023-02-27T10:50:50
1600933116
{ "authors": [ "StefanoBettelli", "Thomasb81", "alaindargelas" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4632", "repo": "chipsalliance/Surelog", "url": "https://github.com/chipsalliance/Surelog/issues/3487" }
gharchive/issue
Removing unused modules from UHDM file + dependencies I translate SystemVerilog files to UHDM with surelog. Let us say that I want to create a design with top module MyModule: >> /usr/local/bin/surelog --version VERSION: 1.47 BUILT : Jan 24 2023 >> /usr/local/bin/surelog -parse -synth -timescale=1ns/1ps -I$(EDIR) -I$(IDIR) +libext+.sv -y $(MDIR) -top MyModule MyModule.sv In this call, EDIR is a directory with .sv files containing enumerations, IDIR is a directory with .sv files containing interfaces, and MDIR is a directory containing a large collection of SystemVerilog modules in .sv files. After analysing the .uhdm file it is apparent that it contains all modules in the MDIR directory, whether or not they are instantiated through the top module. Is it possible to call surelog in such a way that modules which are not used, directly or indirectly, by the top module are removed from the generated .uhdm file? If not, it would be a good idea to include such an option. Is it possible to call surelog in such a way that it generates a dependency file listing all .sv files used, directly or indirectly by the top module, to create the design? This would be very useful in Makefile flows. Hello As far as I understand, Surelog is a tool that will allow you to fill a uhdm database. What you do with the uhdm database is your custom application. Surelog is able to perform some elaboration. One elaboration step consist to complete the uhdm database with what you want. using -d uhdm, the tool will dump the complete uhdm tree. You should be able to observe a difference by adding -noelab, So to answer your question, you probably have to write your own application to iterate on only the part of the uhdm data base that interest you. See: https://github.com/chipsalliance/Surelog/blob/5fc1993668337fa646adbf1e5e0bd139f2304293/src/hellouhdm.cpp#L103 vs https://github.com/chipsalliance/Surelog/blob/5fc1993668337fa646adbf1e5e0bd139f2304293/src/hellouhdm.cpp#L150 @StefanoBettelli, can we close this issue? Are you OK with the file suggested above? Unfortunately I haven't had much time recently to look at it again. However, it seems that this project can hardly be used beyond linting, since there isn't enough documentation, and it is not clear the level of elaboration that is / can be done. There are projects to offload the SystemVerilog fronted of verilator and yosys to Surelog, but so far my efforts at using them consistently in a Makefile flow have failed, due to several shortcomings (not all in Surelog). For the time being the issue can be closed. I will come back to it if there is any substantial progress. @StefanoBettelli you really mean a "Makefile" flow or a "CMake" flow? If you really mean "Makefile" like you have a larger project that uses Makefile and you want to shoehorn Surelog CMake into it, then it is best to invoke the cmake from the Makefile like in: https://github.com/chipsalliance/Surelog/blob/efd45b9291cb4aa03c54e9d3d72d90b885571971/Makefile#L23 and not try to convert any of the CMake steps to Makefile As for the API, the API is the Standard VPI interface (very well documented in the SystemVerilog standard): Accessed using: https://github.com/chipsalliance/UHDM Extract of the Standard API for reference: https://github.com/chipsalliance/Surelog/blob/master/third_party/Verilog_Object_Model.pdf
2025-04-01T06:38:11.262001
2015-10-08T17:20:41
110502166
{ "authors": [ "fnky", "gaui" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4633", "repo": "chjj/marked", "url": "https://github.com/chjj/marked/issues/667" }
gharchive/issue
Different DOM structure for inline/fenced code blocks With GFM, both inline code blocks and fenced code blocks provide the following DOM structure: <p><code>...</code></p> Inline code block: `code` Fenced codeblock: ``` code ``` I would like to style the inline code block differently with CSS than the fenced code block. How can I do this? There's no way to target the inline code block. What options are you using? It should produce <pre><code>...</code><pre> for fenced/block codes, and <p><code>...</code>></p> for inline code. For styling you can select code for inline and pre code for code blocks. It produces in both cases <p><code>... </code></p> I'm using default options. Can you give an example of your code and markdown? Here is a screenshot. I'm using marked() without any options so it's using default options. As you can see there is no difference between inline code block and a fenced code block. This is actually correct behaviour in GFM. It's required to have a newline (\n) character after start fence (except you can provide a language extension after start fence) ```\n Fenced code block\n becomes Fenced code block and Fenced code block becomes Fenced code block ``` The reason is that inline code uses single back ticks so further unescaped back ticks will be ignored in the output. @fnky Thank you very much. Problem solved. :)
2025-04-01T06:38:11.264648
2022-05-29T14:04:46
1251898611
{ "authors": [ "chkilel", "khalidchawtany" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4634", "repo": "chkilel/vitewind-theme", "url": "https://github.com/chkilel/vitewind-theme/issues/2" }
gharchive/issue
Does this theme support HRM? Thanks for this awesome theme. Does this theme support HRM or it does full reload on every change? I always get full reloads making changes to my pages. Hi @khalidchawtany At this time it does not support HMR, I did't find a way to implement it. Thanks Khalid HRM works for JS and CSS files, but if changes are made to the template files the server needs to reload the page to reflect those changes. Hope this clarify your question @khalidchawtany. Thanks that is very nice. I wish it worked for templates too. However, I know that it is almost impossible to implement :(
2025-04-01T06:38:11.267766
2017-04-09T03:55:04
220445653
{ "authors": [ "coveralls", "webdevan" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4635", "repo": "chmln/flatpickr", "url": "https://github.com/chmln/flatpickr/pull/743" }
gharchive/pull-request
dont append confirmDate to native inputs fixes #742 Coverage remained the same at 87.67% when pulling ae2da50a5d0509b42a89fdf0dec14e3719ba2e03 on webdevan:master into 1f1d629c4708a1e06422dcdbba5acbe0e1f7804e on chmln:master.
2025-04-01T06:38:11.280282
2018-04-19T09:54:33
315806877
{ "authors": [ "chobits" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4636", "repo": "chobits/ngx_http_proxy_connect_module", "url": "https://github.com/chobits/ngx_http_proxy_connect_module/pull/40" }
gharchive/pull-request
support nginx-1.14.x try to fix https://github.com/chobits/ngx_http_proxy_connect_module/issues/12 Tested-by: Xiaochen Wang xiaochen Date: Thu Apr 19 17:57:07 CST 2018 hostname: Xiaochens-MacBook-Pro.local pwd: /Users/xiaochen/work/github/ngx_http_proxy_connect_module cases: t/http_proxy_connect.t git log: case result: $ MOD_DYCONF_PATH=/Users/xiaochen/work/github/ngx_http_proxy_connect_module TEST_NGINX_BINARY=/Users/xiaochen/work/github/nginx-1.14.0/objs/nginx prove -v -I /Users/xiaochen/work/github/nginx-tests/lib t/http_proxy_connect.t t/http_proxy_connect.t .. DNS server: try to bind server port: 18085 DNS server: daemon pid: 41072 DNS server: working ok 1 - 200 Connection Established ok 2 - 200 Connection Established server name ok 3 - 200 Connection Established server name ok 4 - 200 Connection Established server name ok 5 - 200 Connection Established not allowed port ok 6 - Get method: proxy_pass ok 7 - Get method: return 200 ok 8 - set remote address ok 9 - set local address and remote address ok 10 - $connect_host, $connect_port, $connect_addr ok 11 - dns resolver fail ok 12 - skip proxy connect module without rewrite phase enabled ok 13 - skip proxy connect module without rewrite phase enabled: if/return DNS server: stop 1..13 ok 14 - no alerts ok 15 - no sanitizer errors All 13 subtests passed Test Summary Report ------------------- t/http_proxy_connect.t (Wstat: 0 Tests: 15 Failed: 2) Failed tests: 14-15 Parse errors: Plan (1..13) must be at the beginning or end of the TAP output Bad plan. You planned 13 tests but ran 15. Files=1, Tests=15, 0 wallclock secs ( 0.02 usr 0.01 sys + 0.11 cusr 0.06 csys = 0.20 CPU) Result: FAIL Report-generated-by: http://wxc.oss.aliyuncs.com/dtest.sh
2025-04-01T06:38:11.287464
2018-09-24T19:09:32
363281546
{ "authors": [ "AppVeyorBot", "pascalberger", "regexaurus" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4637", "repo": "chocolatey/chocolatey-coreteampackages", "url": "https://github.com/chocolatey/chocolatey-coreteampackages/pull/1115" }
gharchive/pull-request
(Skype) Changed InnoSetup /SILENT parameter to /VERYSILENT Description Using /VERYSILENT instead of /SILENT makes Inno Setup a bit quieter. Motivation and Context Users are understandably concerned about "mystery" install windows. How Has this Been Tested? Manually downloaded https://endpoint920510.azureedge.net/s4l/s4l/download/win/Skype-<IP_ADDRESS>.exe, and ran installer with same parameters as in chocolateyInstall.ps1, except /SILENT was changed to /VERYSILENT. Installation succeeded and no installation windows appeared. Skype still (re)launches after installation. Types of changes Minor (largely cosmetic) change Checklist: [x ] My code follows the code style of this repository. [ ] My change requires a change to documentation (this usually means the notes in the description of a package). [ ] I have updated the documentation accordingly (this usually means the notes in the description of a package). [x ] All files are up to date with the latest Contributing Guidelines [ ] The added/modified package passed install/uninstall in the chocolatey test environment. [x ] The changes only affect a single package (not including meta package). :white_check_mark: Package verification completed without issues. PR is now pending human review @regexaurus your changes have been merged, thanks for your contribution 👍
2025-04-01T06:38:11.293465
2017-11-04T01:05:16
271154834
{ "authors": [ "AdmiringWorm", "bcurran3", "dimqua", "rogersachan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4638", "repo": "chocolatey/chocolatey-package-requests", "url": "https://github.com/chocolatey/chocolatey-package-requests/issues/200" }
gharchive/issue
RFP - Croscore fonts The Chrome OS core fonts, also known as the Croscore fonts, are a collection of three TrueType font families: Arimo (sans-serif), Tinos (serif) and Cousine (monospace). These fonts are metrically compatible with Arial, Times New Roman, and Courier New, the most commonly used fonts on Microsoft Windows operating system, for which they are intended as open-source substitutes. More info: https://en.wikipedia.org/wiki/Croscore_fonts Download link: https://gsdview.appspot.com/chromeos-localmirror/distfiles/ Please also supply the home page for these fonts, before we can make this available for maintainers. Homepage: https://fonts.google.com/ More specificaly: Amiro, Tinos and Cousine. I MIGHT look into this. Looks like this is the latest version right? https://gsdview.appspot.com/chromeos-localmirror/distfiles/croscorefonts-1.31.0.tar.bz2 I completely forgot about this. Done! https://chocolatey.org/packages/croscorefonts-font @rogersachan, since you seem appreciative... https://chocolatey.org/packages/crosextrafonts-caladea-font/2013.02.14 https://chocolatey.org/packages/crosextrafonts-carlito-font/2013.09.20 Thanks so much for this!
2025-04-01T06:38:11.310309
2015-01-25T05:46:47
55399369
{ "authors": [ "ahmedmohiduet", "andyli", "ferventcoder" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4639", "repo": "chocolatey/chocolatey", "url": "https://github.com/chocolatey/chocolatey/issues/666" }
gharchive/issue
How can I get installed programs instantly in current cmd window? Every time I install a program I had to restarted my cmd window. Is there any convenient ways to get them in current cmd window? Simething like 'source file.sh' in unix ? P.S. - I am using Windows 7 If you are in cmd.exe and not powershell you can type refreshenv Chocolatey comes with a batch file named the above and will query the registry and reset all of your environment variables. Perhaps we need more documentation surrounding that. Is it possible to get RefreshEnv to work in powershell too? If it cannot, there should be a warning message if RefreshEnv is used in powershell instead of a silent fail. It's an issue with Windows inability to refresh shells. It doesn't have anything to do with choco itself. Now that that is it off the way, refreshenv is a stop gap for choco to keep you from having to close and reopen your cmd window. Its also mostly undocumented. The recommendation is that you close and reopen your shell. You can also do PowerShell if you call & c:\programdata\chocolatey\chocolateyinstall\helpers\Update-Session.ps1 (the path may not be exactly right but you get the idea)- it may require importing the module first and then calling update-session. On Tuesday, January 27, 2015, Andy Li<EMAIL_ADDRESS>wrote: Is it possible to get RefreshEnv to work in powershell too? If it cannot, there should be a warning message if RefreshEnv is used in powershell instead of a silent fail. — Reply to this email directly or view it on GitHub https://github.com/chocolatey/chocolatey/issues/666#issuecomment-71639225 . -- Rob "Be passionate in all you do" http://devlicio.us/blogs/rob_reynolds http://ferventcoder.com http://twitter.com/ferventcoder Do you mean Update-SessionEnvironment? I tried adding it to the end of chocolateyInstall.ps1, but it seems to have no effect - the path var is not updated in powershell. It actually does work but you need to call it separately outside of a choco call (which starts up and closes a separate posh process). On Tuesday, January 27, 2015, Andy Li<EMAIL_ADDRESS>wrote: Do you mean Update-SessionEnvironment https://github.com/chocolatey/chocolatey/wiki/HelpersUpdateSessionEnvironment? I tried adding it to the end of chocolateyInstall.ps1, but it seems to have no effect - the path var is not updated in powershell. — Reply to this email directly or view it on GitHub https://github.com/chocolatey/chocolatey/issues/666#issuecomment-71654814 . -- Rob "Be passionate in all you do" http://devlicio.us/blogs/rob_reynolds http://ferventcoder.com http://twitter.com/ferventcoder Umm.. I thought the helpers are meant to be called inside chocolateyInstall.ps1... Yes, they are. But anything they do with respect to environment vars only lasts during that choco run. Once it comes back to the shell, all child process changes are gone. It would be an egregious security issue with Windows if child processes did more than they do to the parent process. I'm giving you an option for updating your environment variables. Also undocumented and not supported. Ymmv. Remember, you can just close and reopen your shell as well. On Tuesday, January 27, 2015, Andy Li<EMAIL_ADDRESS>wrote: Umm.. I thought the helpers are meant to be called inside chocolateyInstall.ps1... — Reply to this email directly or view it on GitHub https://github.com/chocolatey/chocolatey/issues/666#issuecomment-71655906 . -- Rob "Be passionate in all you do" http://devlicio.us/blogs/rob_reynolds http://ferventcoder.com http://twitter.com/ferventcoder Hi Rob, I was just able to reproduce it my win 7. First I run "choco install java". It says not found.Running "choco install jdk" also fails as you can see in the screen and after that new blank window appears it just freezes nothing happens after that. @ahmedmohiduet does this fail if you switch over to 0.9.9? I guess I will have to check it again. Will let you know after successful reproduce :)
2025-04-01T06:38:11.313141
2016-01-29T02:49:00
129649359
{ "authors": [ "jchodera" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4640", "repo": "choderalab/openmoltools", "url": "https://github.com/choderalab/openmoltools/pull/185" }
gharchive/pull-request
[WIP] Compare energy components with antechamber/prmtop route This is the beginning of a PR to compare energy components with a prmtop route for loading parameters. Currently, the test fails because the imatinib prmtop seems to lead to very different nonbonded forces: Energy components: NonbondedForce -203.736379 31.680160 : 235.416539 HarmonicBondForce 6.160988 6.248384 : 0.087396 HarmonicAngleForce 5.330822 5.329843 : 0.000980 PeriodicTorsionForce 13.253722 8.783295 : 4.470427 CMMotionRemover 0.000000 0.000000 : 0.000000 Maximum allowed deviation (0.600000) exceeded. I'm not sure why this is, but I am investigating. I also don't know the origin of the chemicals/imatinib/imatinib.prmtop file. It may be better for me to compare with a prmtop/inpcrd generated on the fly from either the same procedure I am using for generating the residue templates or the intermediate GAFF mol2/frcmod files used in generating the residue templates. The error is now much smaller: Energy components: NonbondedForce -203.736307 -206.509692 : 2.773386 HarmonicBondForce 6.160980 6.248380 : 0.087400 HarmonicAngleForce 5.330831 5.329848 : 0.000983 PeriodicTorsionForce 13.253717 8.783295 : 4.470422 CMMotionRemover 0.000000 0.000000 : 0.000000 Maximum allowed deviation (0.600000) exceeded. Will need to do more debugging but raised the max allowed deviation to 5 for now.
2025-04-01T06:38:11.316719
2018-07-09T23:43:18
339644403
{ "authors": [ "jchodera", "pgrinaway" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4641", "repo": "choderalab/perses", "url": "https://github.com/choderalab/perses/issues/476" }
gharchive/issue
Nonequilibrium switching fails with new OpenMM API I think this is actually an update that needs to be made in openmmtools. The AlchemicalState object is unaware of how to set the global parameter lambda_electrostatics if it's covered by the regular NonbondedForce. cc: https://github.com/choderalab/openmmtools/issues/359 @pgrinaway This should be resolved using the latest OpenMM 7.3.0 dev build now that https://github.com/pandegroup/openmm/pull/2119 and https://github.com/pandegroup/openmm/pull/2115 have been merged. OK to close? @brycestx points out we need to revise things so that we explicitly add the global context parameters to NonbondedForce now that the API has been extended. Fixed in the OpenMM 7.3.1 bugfix release.
2025-04-01T06:38:11.323447
2023-08-30T22:13:27
1874447426
{ "authors": [ "choogiesaur", "digikitty21" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4642", "repo": "choogiesaur/hgss-shiny-modifier", "url": "https://github.com/choogiesaur/hgss-shiny-modifier/pull/5" }
gharchive/pull-request
Mark arm9 as Decompressed, and Change Shiny Value Offset Output to Hex I'm not sure if I did this right, but hopefully this should only include: Explicitly marking the arm9 as decompressed Outputting the Shiny Value offset in hex Thanks for these changes @digikitty21 ! Merged to main.
2025-04-01T06:38:11.330829
2020-10-21T15:31:27
726601976
{ "authors": [ "bastelfreak", "ripienaar" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4643", "repo": "choria-io/puppet-mcollective", "url": "https://github.com/choria-io/puppet-mcollective/issues/263" }
gharchive/issue
Harden Puppet datatypes Some of the datatypes are not as strict as they could be. To make the usage safer, the should be hardened. thank you!
2025-04-01T06:38:11.365713
2024-11-20T03:11:31
2674200253
{ "authors": [ "chrisbenincasa", "pavlov70" ], "license": "Zlib", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4644", "repo": "chrisbenincasa/tunarr", "url": "https://github.com/chrisbenincasa/tunarr/issues/974" }
gharchive/issue
minor issue with new ffmpeg stack on .16 Hi, After upgrading to .16 and trying out the new ffmpeg stack, it seems to work quite well from initial impressions, perhaps loading a bit faster. The one issue I've encountered is when playing back via Kodi's iptv plugin. It would immediately fail without opening the stream. The log showed: CCurlFile::Open - http://<IP_ADDRESS>:8000/stream/channels/1fa17176-a81c-495c-a631-67a375320412.ts Failed with code 500: Pasting the url into vlc also resulted in failed playback. By changing the .ts to .m3u8 I was able to play back in vlc successfully, so I edited the channels.m3u and changed each .ts instance to .m3u8. This restored my kodi iptv functionality. I should note that the above is not necessary using the old ffmpeg stack. Thanks again as always! Any relevant logs on the server side? Also can you provide a few more details? Like are you using hw accel? If so, which kind? I was able to reproduce this. Will have a fix today. Thanks for testing! Sorry I didn't get you the logs earlier, life interrupted me. Thanks for identifying the cause and I'm looking forward to the fix. Been using the docker vaapi and load times in Kodi are in the ~1-2 second range with the new ffmpeg changes (down a second or so). Every bit as fast as traditional linear tv!
2025-04-01T06:38:11.370523
2015-09-03T23:10:09
104799176
{ "authors": [ "chrisccerami", "felipebalbi" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4645", "repo": "chrisccerami/ascii_art", "url": "https://github.com/chrisccerami/ascii_art/issues/4" }
gharchive/issue
Curses has been removed from Standard Library Please add runtime dependency to curse ~> 1.0 if Ruby is > 2.1 Thanks for opening the issue. I'll take care of this as soon as I get a chance. This has been addressed. Thanks again for pointing it out @felipebalbi Hey, no problem. I'll have a look at how you did it. I knew about adding it to gemspec, but didn't know how to add it conditionally on ruby version :-)
2025-04-01T06:38:11.379150
2017-11-28T19:43:10
277513634
{ "authors": [ "CloudChoSony", "SJTUPanda", "ShubhayanS", "speed8928" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4646", "repo": "chrischoy/3D-R2N2", "url": "https://github.com/chrischoy/3D-R2N2/issues/22" }
gharchive/issue
ValueError: You are tring to use the old GPU back-end. Hi I use windows 10 to implement the project and it turns out the error below: ValueError: You are tring to use the old GPU back-end. It was removed from Theano. Use device=cuda* I tried to follow your link that from error https://github.com/Theano/Theano/wiki/Converting-to-the-new-gpu-back-end(gpuarray) and install theano pygpu, but nothing good come out. I also tried change the Theano variable device = gpu to device = cuda0 and it gives me the error below: (py3) C:\Users\speed\source\3D-R2N2>python demo.py prediction.obj WARNING (theano.tensor.blas): Using NumPy C-API based implementation for BLAS functions. ERROR (theano.gpuarray): pygpu was configured but could not be imported or is too old (version 0.7 or higher required) NoneType Downloading a pretrained model Traceback (most recent call last): File "demo.py", line 84, in main() File "demo.py", line 55, in main download_model(DEFAULT_WEIGHTS) File "demo.py", line 34, in download_model '--create-dirs', '-o', fn]) File "C:\Users\speed\AppData\Local\Programs\Python\Python35\Lib\subprocess.py", line 557, in call with Popen(*popenargs, **kwargs) as p: File "C:\Users\speed\AppData\Local\Programs\Python\Python35\Lib\subprocess.py", line 947, in init restore_signals, start_new_session) File "C:\Users\speed\AppData\Local\Programs\Python\Python35\Lib\subprocess.py", line 1224, in _execute_child startupinfo) FileNotFoundError: [WinError 2] The system cannot find the file specified Could you please advise me the way to compile it, thanks. Hi, I have the same error with you. Have you solved this problem? I lowered the Theano version 7.x then it worked. Thanks! I use the Theano version 0.8.0 then it works. I found another way to solve at ValueError «You are trying to use the old GPU back-end» when importing keras According to cgl, please type in terminal "export THEANO_FLAGS=mode=FAST_RUN,device=cuda,floatX=float32" Thanks! I use the Theano version 0.8.0 then it works. I am getting - AttributeError: ('This name is already taken', 'floatX')
2025-04-01T06:38:11.383224
2024-06-22T05:44:41
2367588031
{ "authors": [ "Sponge-bink", "chrisgrieser" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4647", "repo": "chrisgrieser/alfred-wikipedia-suggest", "url": "https://github.com/chrisgrieser/alfred-wikipedia-suggest/pull/1" }
gharchive/pull-request
feat: configurable number of results. multiple languages. I left the version number unchanged. You might wanna update that later. ;) Thx! Yeah, updating the version number is sth that I do with btw releases. Will do one later today Release as 1.1.0: https://github.com/chrisgrieser/alfred-wikipedia-suggest/releases/download/1.1.0/alfred-wikipedia-suggest.alfredworkflow Just so you know, I made some minor changes, like using , as delimiter instead of /, as that's what most workflows do, so you might have go over your settings briefly.
2025-04-01T06:38:11.394784
2015-07-20T06:57:07
95999913
{ "authors": [ "elod91" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4648", "repo": "chrisjenx/Calligraphy", "url": "https://github.com/chrisjenx/Calligraphy/issues/192" }
gharchive/issue
Diacritical problems on older devices Hi! I've been using you library for a while now without any problems, but now I have a project in which, some of the texts have diacritics. I don't know the exact android version on which it runs OK and from which it has problems, but it seems like pre-L devices don't show the diacritical characters with the custom font. Sorry, seems like the font I was using had problems. Changed it to another and it works
2025-04-01T06:38:11.408258
2022-01-20T02:42:29
1108770310
{ "authors": [ "chrissy-dev", "jshwlkr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4649", "repo": "chrissy-dev/eleventy-web-starter", "url": "https://github.com/chrissy-dev/eleventy-web-starter/issues/33" }
gharchive/issue
I don't think you need Luxon Just an FYI. I think you can use https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Intl/DateTimeFormat. @jshwlkr I've decided to keep it for now. Eleventy uses Luxon internally so I'm happy to keep the tooling the same.
2025-04-01T06:38:11.412838
2015-09-26T22:57:07
108501121
{ "authors": [ "christianalfoni", "danpantry", "tomatau" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4650", "repo": "christianalfoni/formsy-react", "url": "https://github.com/christianalfoni/formsy-react/pull/221" }
gharchive/pull-request
Transpile with Babel prior to publishing to npm This enables Node environments to use require('formsy-react') instead of having to transpile it themselves Fixes #220 Should not really have any affect on callers - transpiled code is functionally equivalent FYI I wrote this on react-0.14 because that's the branch I am using. This should apply without issues to master, too. :+1: please merge this If anyone wants to use this now - I've forked your fork with the ./lib built and checked in. https://github.com/tomatau/formsy-react/tree/react-0.14 you can: "dependencies": { "formsy-react": "tomatau/formsy-react#react-0.14" Thanks @tomatau Ideally we'd configure the .npmignore so it doesn't include ./src files, but only ./lib files - that way people who download the package don't have extraneous files Precisely! But we don't want to create an npm module copy of formsy-react just whilst waiting for this merge :) Great, thanks! New version will be released this weekend, with latest version of React 0.14 @christianalfoni :beers: :+1:
2025-04-01T06:38:11.414204
2021-03-16T20:54:47
833188777
{ "authors": [ "binhex", "christianhaitian" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4651", "repo": "christianhaitian/arkos", "url": "https://github.com/christianhaitian/arkos/issues/295" }
gharchive/issue
[Wiki] Link to download v1.6 on the Wiki please. Please can you update the Wiki with a link to v1.6 for the Anbernic 351m, many thanks for your hard work!. I will eventually but it will take some time. Until then, you can download the current image and update online to it. Stay tuned. OK that makes sense, i appreciate the response.
2025-04-01T06:38:11.428449
2015-11-18T13:25:33
117586322
{ "authors": [ "monsterxxx", "rguerreiro" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4652", "repo": "christopherthielen/ui-router-extras", "url": "https://github.com/christopherthielen/ui-router-extras/issues/273" }
gharchive/issue
Sticky state doesn't work when applied to a state with params Hi, sticky state doesn't seem to work when it has parameters. Example: .state('main.products', { url: '/products/:product_id', sticky: true }) .state('main.products.details', { url: '/details' }) When you navigate between these 'main.product.details ' states while changing product_id, you will get an error. Plunk: http://plnkr.co/edit/czXStZZCQ9wDZ4mVqtA2?p=info Step by step log: Go to products/1/details Current transition: main: {}: -> main.products.details: {"product_id":1} Before transition, inactives are: : [] After transition, inactives will be: [] Transition will exit: ["(main)"] Transition will enter: ["(main)", "ENTER: main.products", "ENTER: main.products.details"] SurrogateFromPath: ["main"] SurrogateToPath: ["main", "main.products", "main.products.details"] Current state: main.products.details, inactive states: [] Go to products/2/details Current transition: main.products.details: {"product_id":"1"}: -> main.products.details: {"product_id":2} Before transition, inactives are: : [] After transition, inactives will be: ["main.products", "main.products.details"] Transition will exit: ["(main)", "INACTIVATE: main.products", "INACTIVATE: main.products.details"] Transition will enter: ["(main)", "ENTER: main.products", "ENTER: main.products.details"] SurrogateFromPath: ["main", "inactivate:main.products", "inactivate:main.products.details"] SurrogateToPath: ["main", "main.products", "main.products.details"] Current state: main.products.details, inactive states: ["main.products.details", "main.products"] Go to products/3/details and get an error Current transition: main.products.details: {"product_id":"2"}: -> main.products.details: {"product_id":3} Before transition, inactives are: : ["main.products.details", "main.products"] After transition, inactives will be: ["main.products", "main.products.details"] Transition will exit: ["(main)", "INACTIVATE: main.products", "INACTIVATE: main.products.details"] Transition will enter: ["(main)", "RELOAD: main.products", "RELOAD: main.products.details"] SurrogateFromPath: ["main", "inactivate:main.products", "inactivate:main.products.details"] SurrogateToPath: ["main", "main.products", "main.products.details"] Exiting main.products.details because it's a substate of main.products and wasn't found in Object {} transition failed TypeError: Cannot read property 'globals' of null If I add intermediate state like this: .state('main.products', { url: '/products', sticky: true }) .state('main.products.product', { url: '/:product_id' }) .state('main.products.product.details', { url: '/details' }) it will work for my purpose, but it's not very convenient in terms of the app structure. It seems to be what it's happening to me. Here's a plunker. Go to 'Tabs` and then go back and forth between the two existing tabs and you'll see the controllers being reloaded. Unless I'm doing something wrong... No, in my case I think it was my misunderstanding of the main sticky state concept - its scope should stay and controller shold not be reloaded. So I ended with an intermidiate state which was mentioned at the end of my question.
2025-04-01T06:38:11.430114
2015-02-12T09:36:30
57436847
{ "authors": [ "christopherthielen", "maximumduncan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4653", "repo": "christopherthielen/ui-router-extras", "url": "https://github.com/christopherthielen/ui-router-extras/pull/167" }
gharchive/pull-request
fix(future): fixed absence of url for real parent of future state If you have an already registered abstract (no url) parent state and you try to register a new future state which is a child of this parent, I found that this fix is needed in order to stop a JS error. The error occurs because it seems that navigable is not always set; in these cases it's trying to access the url property of a null navigable object. manually merged
2025-04-01T06:38:11.431277
2016-05-21T17:44:34
156113508
{ "authors": [ "Schwanksta", "chriswhong" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4654", "repo": "chriswhong/postgis-preview", "url": "https://github.com/chriswhong/postgis-preview/pull/25" }
gharchive/pull-request
Submit button broke The form tag was removed, so the submit button no longer worked. Just did a quick change to run when #run is clicked. Already caught this here, but had not pushed to master. Thanks!
2025-04-01T06:38:11.467287
2019-10-17T07:17:43
508284080
{ "authors": [ "Ashutar", "lazarte" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4655", "repo": "chronotruck/FlagPhoneNumber", "url": "https://github.com/chronotruck/FlagPhoneNumber/issues/134" }
gharchive/issue
Phone format When I have a phone number like<PHONE_NUMBER>, can we set this into the textfield to format<PHONE_NUMBER>? Also when type the number can this be format into a format mentioned? up
2025-04-01T06:38:11.537679
2015-02-26T00:25:53
58994772
{ "authors": [ "GoogleCodeExporter" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4656", "repo": "chrsmithdemos/minify", "url": "https://github.com/chrsmithdemos/minify/issues/65" }
gharchive/issue
Extending Minify to load a controller from a different location For an internal implementation of Minify integrated with our own MVC framework, we needed to load a custom controller for use with our views, and using the naming convention for our framework for the controller class Minify would fail, attempting instead to load Minify_Controller_etcetc. The fix for this was of course to override serve() in our Minify extended class (which we had already created as a wrapper as a matter of good practice), but copying and pasting the entirety of serve(), which does quite a few things and has changed between releases of Minify, seems the wrong way to do this to override one simple little autoload method. As a proposed patch (attached) I've moved the controller loading to a separate static method in Minify and call this from within serve(), allowing anyone that wishes to do so to just override this loadController() method. Not sure if this is solving an edge-case here, but it feels like a better way to do this overall. Original issue reported on code.google.com by<EMAIL_ADDRESS>on 29 Oct 2008 at 9:39 Attachments: Minify.patch You do not need to extend Minify to do this. Simply create a controller object (load it however you want) and pass it as the first parameter to serve(). It's easiest, but not required, to make your controller a subclass of Minify_Controller_Base. {{{ $myController = MyFramework::factory('MinifyController'); Minify::serve($myController, $options); }}} serve() only uses its loader if you pass a string. I'm sorry this isn't well- documented yet; here's the phpdoc line that specifies this: http://code.google.com/p/minify/source/browse/tags/release_2.1.0/min/lib/ Minify.php#135 Original comment by<EMAIL_ADDRESS>on 29 Oct 2008 at 2:43 Changed title: Extending Minify to load a controller from a different location Changed state: WontFix I labelled this incorrectly, sorry for the noise. Original comment by<EMAIL_ADDRESS>on 27 Nov 2008 at 3:49 Changed state: Invalid
2025-04-01T06:38:11.570583
2018-07-17T23:28:10
342123943
{ "authors": [ "chubin", "karlosp", "phxvyper", "rprimus" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4657", "repo": "chubin/cheat.sh", "url": "https://github.com/chubin/cheat.sh/issues/73" }
gharchive/issue
cd section fails in shell mode, returns no valid sections On macOS High Sierra 10.13.1 Cannot use cd command to change my context for queries. Queries work normally, and I can run /:context QUERY normally. Steps to repro Install script locally, setup permissions $ curl https://cht.sh/:cht.sh > /usr/local/bin/cht.sh $ chmod +x /usr/local/bin/cht.sh install rlwrap for shellmode $ brew install rlwrap launch shell mode cht.sh --shell attempt to cd into any section/context cht.sh> cd go Invalid section: go Valid sections: cht.sh> cd csharp Invalid section: csharp Valid sections: cht.sh> cd arduino Invalid section: arduino Valid sections: I have exactly the same problem on debian installed under Windows Subsystem for Linux. Yes, it is true, this was broken in this commit: https://github.com/chubin/cheat.sh/commit/b24381c7403f956e0dfcf10417c7ebe2b8165edb (and this was not detected by our regression tests). I will fix this problem today. Thank you very much reporting Thank you very much for reporting. The problem is fixed, please test Thanks. It is working for me. Thu Apr 21 07:56:28 BST 2022 Hi, Running into this on macOS 12.3.1 (see attached animated gif): The problem seems to be at https://github.com/chubin/cheat.sh/blob/562875eda610b0322819def25f2d27af1bf9469a/share/cht.sh.txt#L371 The following works for me: curl -s "${CHTSH_URL}"/:list | grep ':list' | cut -d: -f1 | xargs The above produces: : ; curl -s https://cht.sh/:list | grep ':list' | grep ':list' | cut -d: -f1 |xargs awk/ bash/ bf/ c/ chapel/ clojure/ cmake/ coffee/ cpp/ csharp/ d/ dart/ elisp/ elixir/ elm/ erlang/ factor/ forth/ fortran/ fsharp/ git/ go/ groovy/ haskell/ java/ js/ julia/ kotlin/ latex/ lisp/ lua/ mathematica/ matlab/ nim/ objective-c/ ocaml/ octave/ perl/ perl6/ php/ python/ python3/ r/ racket/ ruby/ rust/ solidity/ swift/ tcl/ tcsh/ vb/
2025-04-01T06:38:11.592378
2024-08-06T13:50:19
2450946997
{ "authors": [ "Hapluckyy", "chujiezheng" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4658", "repo": "chujiezheng/chat_templates", "url": "https://github.com/chujiezheng/chat_templates/issues/23" }
gharchive/issue
Gemma2 pleeeeease! https://huggingface.co/google/gemma-2-9b It has the same chat template as gemma-1. Please refer to the instructions for gemma-1.
2025-04-01T06:38:11.598270
2021-03-28T08:40:28
842692256
{ "authors": [ "adbonnin", "chulwoo-park", "pierre-gancel" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4659", "repo": "chulwoo-park/timelines", "url": "https://github.com/chulwoo-park/timelines/issues/34" }
gharchive/issue
Question : How to align the nodes to the left ? Hi, I would like to align the nodes to the left in order to have a fixed size for the content and the remaining space for the opposite content. There is a property nodeAlign in TimelineTile that used to compute the effective node position : double _getEffectiveNodePosition(BuildContext context) { if (nodeAlign == TimelineNodeAlign.start) return 0.0; if (nodeAlign == TimelineNodeAlign.end) return 1.0; var nodePosition = this.nodePosition; nodePosition ??= (node is TimelineTileNode) ? (node as TimelineTileNode).getEffectivePosition(context) : TimelineTheme.of(context).nodePosition; return nodePosition; } This property can't be used in TimelineTileBuilder so please what is the simplest way to define the nodeAlign ? try using nodePosition in TimelineTheme? Great package ! Timeline.tileBuilder( theme: TimelineThemeData( connectorTheme: ConnectorThemeData( space: 51, thickness: 2.5, color: Colors.purple), nodePosition: 0, color: yellow),) Try using nodePosition in TimelineTheme? Check how it works here (Theme) nodePosition works but it's the a percentage between 0 and 1. If the screen is too small then the oppositeContent width will shrink it's content. What I'm looking for is a fixed width for oppositeContent and the remaining width for the content. Timeline.tileBuilder(theme: TimelineThemeData(connectorTheme: ConnectorThemeData(space: 51))) Connector's space is the space between the content and the oppositeContent, it's not the width of the oppositeContent. @adbonnin I got it. There is no option yet to provide the feature you are talking about.😢 Since it uses Flexible internally, I think it is necessary to explicitly limit the size of the opossite content. ok thank you 😄
2025-04-01T06:38:11.613034
2023-03-16T18:00:08
1628008039
{ "authors": [ "chuongmep", "shtirlitsDva" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4660", "repo": "chuongmep/CadAddinManager", "url": "https://github.com/chuongmep/CadAddinManager/issues/8" }
gharchive/issue
Loading references Hi Is there a way to load references added to the project which I am debugging? I mean, if I use nuget packages or reference another project's dll, it won't be available when I use this program -- the command will throw an exception saying that it can't find the referenced dll. I also had this problem with netreload. I have worked around this problem by linking my source files around instead of referencing projects and instead of using nuget, I had to download source and also link it in the debugged project to be able to use netreload. There seems to be same problem with this manager also. I don't know enough about .net to solve the problem my self. Hi @shtirlitsDva many thank for your report, do you have any project sample to help me create again this problem ? This issues can't create again, so I will close it like resolved.
2025-04-01T06:38:11.614365
2016-09-24T10:07:51
179022699
{ "authors": [ "MrZoolook", "chuparCh0pper" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4661", "repo": "chuparCh0pper/PoGoIV_xposed", "url": "https://github.com/chuparCh0pper/PoGoIV_xposed/issues/29" }
gharchive/issue
Scrolling the gym info notification When there are more than 2 Pokémon in a gym, in the notification with the gym info, scrolling down past the first 2 will only display the first 3 lines of info on the 3rd 'mon instead of the full 4 lines. fixed https://github.com/chuparCh0pper/PoGoIV_xposed/commit/2f20def1380914c27b4d20a08b392be8f9e8ccae will push to xposed repo in the next few days
2025-04-01T06:38:11.615424
2015-07-29T22:25:47
98042832
{ "authors": [ "dbgoodman" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4662", "repo": "churchlab/millstone", "url": "https://github.com/churchlab/millstone/issues/562" }
gharchive/issue
Jbrowse should only show tracks for current alignment group It currently shows tracks for all alignment groups, which will get confusing very fast. Not sure how easy it will be to fix this... It also shows multiple versions of the same track if that track was for the same genome in different alignment groups.
2025-04-01T06:38:11.624225
2017-05-12T14:54:56
228317155
{ "authors": [ "bobby-brennan" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4665", "repo": "chymz/ng2-codemirror", "url": "https://github.com/chymz/ng2-codemirror/issues/16" }
gharchive/issue
Scroll resets when ngModel changes? When I change ngModel programmatically (rather than by typing in the editor), the scroll jumps back up to the top line. Any idea what's causing this? Is it expected behavior? Looks like setValue() is only meant for when the entire code changes, and you're supposed to use replaceRange() when only a portion of the code changes. Will close this issue, but I would recommend documenting how to access the internal codemirror instance in your README, i.e. <codemirror #editor [(ngModel)]="code"></codemirror> <a (click)="editor.instance.replaceRange('hello world', {line: 0, ch: 0})">click me</a>
2025-04-01T06:38:11.638031
2022-03-01T13:28:58
1155342718
{ "authors": [ "cientgu", "yzxing87" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4666", "repo": "cientgu/VQ-Diffusion", "url": "https://github.com/cientgu/VQ-Diffusion/issues/7" }
gharchive/issue
question about 'filter_ratio' parameter Hi, thanks for the excellent implementation! Could you please tell me what is the purpose of the 'filter_ratio' parameter in the sampling function? And I also note that the intermediate training results are sampled with different filter_ratio parameters. How should we interpret the results with different values? Thanks! This is just for debugging in the training procedure. To make sure each denoising step works well. Thanks for the clarification.
2025-04-01T06:38:11.642764
2024-11-21T21:25:15
2680996348
{ "authors": [ "bitalec", "cieslarmichal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4667", "repo": "cieslarmichal/faker-cxx", "url": "https://github.com/cieslarmichal/faker-cxx/issues/990" }
gharchive/issue
fix imageUrl in Image module to generate images from picsum photos replace https://loremflickr.com with https://picsum.photos/ and test if it works and generates different images for given categories Hi. I'm new to open source contributions and would like to try working on this issue. Could you please assign this issue to me? Sure, assigned Hi, I have a question. I saw that picsum.photos doesn’t let you generate random images by category (like dogs, tech, etc.). It only gives random images or images by ID. Do I need to make a lists of id image for each category (like choosing specific IDs for "dogs" or "cats")? Or is there an easier way to do this? Thanks for your help! lets maybe refactor it to seperate methods like in fakerjs: so picsum and flickr would have seperate methods and url would get random either picsum or flickr Hi, I added the urlPicsumPhotos method to generate images from https://picsum.photos/. Can I also change the name of the imageUrl method? Should I add the tests in a separate commit, or is it fine to include them in the same one? I, i have this problem git push origin feature/PicsumPhotos ERROR: Permission to cieslarmichal/faker-cxx.git denied to bitalec.
2025-04-01T06:38:11.648938
2020-10-09T21:33:47
718434767
{ "authors": [ "christarazi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4668", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/13498" }
gharchive/issue
Investigate CRD controller timing out if no Cilium CRDs are present when it is created The issue is that when the CRD controller is first created and begins watching for CRDs in the cluster, and no Cilium CRDs are present at that point (meaning cilium-operator has not registered them yet), the crd-wait-timeout will be triggered, even after cilium-operator registers the Ciilum CRDs. The controller is unable to find them due to a K8s apiserver error indicating that v1beta1.CustomResourceDefiniton was found, when expecting v1.PartialObjectMetadata. This error was observed while working on https://github.com/cilium/cilium/pull/13418. It seems to only occur on K8s versions below 1.15. Reproduction steps: Deploy Delete all Cilium CRDs Roll agent Tail agent logs Roll operator so that it can register Cilium CRDs Observe controller errors in the agent logs These errors eventually resolve themselves after the CRD wait timeout is hit (default is 5m). The timeout will fatal the agent, and upon it restarting, it is able to sync the CRDs and the agent goes on fine. Likely cause of this is that watching a PartialObjectMetadata is not supported until K8s 1.15, which would explain the behavior of it working on 1.15 and not 1.14.
2025-04-01T06:38:11.670537
2023-12-13T10:05:04
2039365039
{ "authors": [ "giorio94", "julianwiedmann", "mathpl", "pchaigno", "pippolo84", "squeed", "tommyp1ckles" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4669", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/29846" }
gharchive/issue
CI: Cilium E2E Upgrade: Timed out waiting for datapath updates of FQDN IP information after upgrade From https://github.com/cilium/cilium/actions/runs/7190938554/job/19584876380 [=] Test [check-log-errors] [67/67] [-] Scenario [check-log-errors/no-errors-in-logs] Found "2023-12-13T05:22:11.945479031Z level=error msg=\"Timed out waiting for datapath updates of FQDN IP information; returning response\" subsys=daemon" in logs 1 times ❌ Found 1 logs matching list of errors that must be investigated: 2023-12-13T05:22:11.945479031Z level=error msg="Timed out waiting for datapath updates of FQDN IP information; returning response" subsys=daemon Sysdump: cilium-sysdump-5-20231213-052359.7z.zip Didn't see it in dogfooding hence I closed https://github.com/isovalent/customer-support/issues/557 which looks like the same issue, but still happening in CI. Filed a PR should help reduce these timeouts significantly. Found a needless regeneration on the FQDN path that caused lock contention. Hit on Conformance GKE as well after enabling the check for log errors also there: https://github.com/cilium/cilium/actions/runs/7208580138/job/19637888894#step:21:182 Reopening, as hit in the Cilium IPsec upgrade workflow for a PR which includes the commits from https://github.com/cilium/cilium/pull/29865. PR: https://github.com/cilium/cilium/pull/30012 Link: https://github.com/cilium/cilium/actions/runs/7278734523/job/19833561516 Another hit here. [=] Test [check-log-errors] [71/71] [-] Scenario [check-log-errors/no-errors-in-logs] Found "2024-01-02T10:54:30.033766201Z level=error msg=\"Timed out waiting for datapath updates of FQDN IP information; returning response\" subsys=daemon" in logs 1 times ❌ Found 1 logs matching list of errors that must be investigated: 2024-01-02T10:54:30.033766201Z level=error msg="Timed out waiting for datapath updates of FQDN IP information; returning response" subsys=daemon Full logs Sysdump too big :( Taking a look. Aha, the backport to v1.15 hasn't merged yet. Whew. Taking a look. Thanks! I kept a local copy of the sysdump, let me know if you want to take a look. @pippolo84 are these failures on the backport PR that includes this change? GitHub is not making it easy to determine. @pippolo84 are these failures on the backport PR that includes this change? GitHub is not making it easy to determine. Yep, the workflow has been triggered by this PR. Oh, you were looking for your changes from https://github.com/cilium/cilium/pull/29865. Those have been backported in the PR that triggered the workflow, so it seems the issue is still there, unfortunately. So, some basic analysis: There are two practically-concurrent requests, for A one.one.one.one. and AAAA one.one.one.one. The timeout message comes from the second request, in this case, AAAA. The vast majority of that time is spent waiting for the ipcache to complete The basic flow 10:54:29.915: Response to A one.one.one.one. is received 10:54:29.921: NameManager is updated; locks released, waiting for ipcache to process v4 addrs. ipcache immediately starts. 10:54:29.923: Response to AAAA is received. 10:54:29.927: NameManager is updated; locks released, waiting for ipcache to process v6 addrs 10:54:29.933: ipcache PolicyMap updates are complete, waiting for proxy (Envoy) updates 10:54:29.946: Proxy updates are complete, ipcache is complete, total duration ~24ms, identity allocation ~1ms, proxy update ~9ms 10:54:29.946: ipcache starts again. 10:54:29.947: DNS response is released for A request. Total time: 32ms 10:54:29.987: ipcache PolicyMap updates are complete, waiting for proxy (Envoy) updates 10:54:30.033: We give up waiting for ipcache, write DNS response back. 10:54:30.037: Proxy updates are complete, ipcache is complete, total duration 93ms, identity allocation ~20ms, proxy update ~50ms, identity allocation 15ms So, two observations: Identity allocation randomly takes a long time. I'm assuming this is due to GC pauses / allocation Envoy can also take a long time. I don't see any smoking guns; 20ms lost in a trivial map update and 50ms lost waiting for envoy, plus the rest of the FQDN process, put us over 100ms. I'll try and dig in to why this is going wrong. Next step is to find out why. Created a gist with a bit more information here: https://gist.github.com/squeed/752d105c569db3eb328de191d37a4ed8 I'll ask around for help. I did some exploration with @jrajahalme's assistance, and I found that Envoy sometimes just takes 20-30ms to process updates. There are no obvious performance smoking guns. I would like to consider bumping the timeout to 150ms in CI to see if that gets rid of flakes. All the flakes I saw were timing out in ~103ms, the threshold is 100ms. I think we're just too resource constrained here. I've seen this error message hit on some v1.14 CI jobs, which does not have my FQDN refactor. So I suspect this has always been an issue. We may wish to consider ignoring this warning. Perhaps it is allowed to log once, but more than that is indicative of an issue. Hard to say; it is literally a threshold matter. Couldn't users hit this as well? If so, ignoring in CI may not be enough. FYI, this is still happening a lot in CI and currently preventing us from making some workflows Required. Users can and do hit this in production, it has been this way for years. We don't have good data, though. If we want to continue blocking CI, then we can keep this as we figure out how to redesign the Envoy system. Otherwise, we should skip this error message. Sounds like we may need to allowlist in CI + document. Sounds like we may need to allowlist in CI + document. Agreed, I'll take care of that shortly. And figure out how we can improve Envoy. @squeed Created a PR on cilium-cli to add this to the list of exceptions. Can I get your confirmation that this is the appropriate way to handle this? This is also affecting at least Conformance GKE: https://github.com/cilium/cilium/actions/runs/8556001839/job/23444773987. @squeed If you don't have time to implement the fix, could you unassign yourself and ask for someone else in your team to handle it? :pray: Filed https://github.com/cilium/cilium/pull/31866 to bump the timeout to 250ms. There have also been a few FQDN performance improvements: #31454, merged Apr 3 #30897, merged Apr 3 Hopefully this cuts the noise down significantly. The GKE occurrence above is from April 4. This one as well: https://github.com/cilium/cilium/actions/runs/8558512353/job/23453228166. Both on main. It may have reduced it (?) but it doesn't look like it was enough. This is still happening on main: https://github.com/cilium/cilium/actions/runs/8690522661/job/23830675920. Just hit on v1.14 as well: https://github.com/cilium/cilium/actions/runs/8703624933/job/23870214379 Requested backport to v1.15 and v1.14. Haven't seen this in a long time. Unassigning so that stale-issue GC can take over.
2025-04-01T06:38:11.673525
2018-03-06T19:31:05
302834832
{ "authors": [ "ianvernon", "joestringer" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4670", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/3031" }
gharchive/issue
Update docs with latest cilium status output The following documents have cilium status, given the recent changes to the output they are all likely out of date: $ git grep "cilium status" Documentation/ Documentation/cheatsheet.rst: cilium status Documentation/cmdref/cilium.md:* [cilium status](cilium_status.html) - Display status of daemon Documentation/cmdref/cilium_status.md:## cilium status Documentation/cmdref/cilium_status.md:cilium status Documentation/contributing.rst: $ service cilium status Documentation/contributing.rst: $ cilium status Documentation/contributing.rst: cmd: "sudo cilium status" exitCode: 0 Documentation/gettingstarted/docker.rst:``cilium status``: Documentation/gettingstarted/docker.rst: $ cilium status Documentation/gettingstarted/mesos.rst: $ cilium status Documentation/troubleshooting.rst:health. This is achieved by running the ``cilium status`` command. Documentation/troubleshooting.rst:``cilium status`` on all cluster nodes with ease. Download the Documentation/troubleshooting.rst:... and run ``cilium status`` on all nodes: Documentation/troubleshooting.rst: $ ./k8s-cilium-exec.sh cilium status Documentation/troubleshooting.rst: $ cilium status Need to run through each of the documents, attempt to reproduce and run cilium status to grab expected output and update the documents. It would be nice to automatically require documentation updates to go along with CLI updates - I'm not sure how this could be automated, though :/ Ended up not being able to pick this up; un-assigning myself.
2025-04-01T06:38:11.682457
2024-08-08T20:48:24
2456599581
{ "authors": [ "jspaleta", "julianwiedmann" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4671", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/34253" }
gharchive/issue
Doc Bug: Talos install instructions fail to work, may need note adjustment about kubeprism. Is there an existing issue for this? [X] I have searched the existing issues Version higher than v1.16.0 and lower than v1.17.0 What happened? Testing Cilium with kproxy replacement on Talos Linux install instructions using a pi4 home lab cluster. Cilium install failed using instructions as written. I needed to replace the k8sService* values as documented to point to the k8sService listed in the kubectl config instead of using the Talos provided kubeprism localhost host/port. How can we reproduce the issue? create Talos Linux install images for pi4 using metal-arm64.raw as documented in Talos linux upstream docs. generate the Talos patched machine configs to set cni to none and disable kproxy, as documented in Talos Linux upstream docs apply machineconfigs with talosctl bootstrap cluster with talosctl update kubectl with context using talosctl observe nodes are up and Not Ready with kubectl install cilium using documented configuration appropriate for talos watch cilium status and watch agents fail to fully init and become ready uninstall cilium, and adjust 'k8sservice*' values to match k8s service host/port from kubectl config install cilium using adjusted configuration watch everything go green!!!! Cilium Version 1.16.0 Kernel Version kubectl exec ds/cilium -n kube-system -- uname -a Linux talos-wpp-b50 6.6.28-talos #1 SMP Thu Apr 18 13:43:02 UTC 2024 aarch64 aarch64 aarch64 GNU/Linux Kubernetes Version Client Version: v1.28.2 Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3 Server Version: v1.27.4 Regression Not a regression, Talos linux instructions are new as of 1.16 I think. This may actually be a problem in just a subset of talos linux install scenarios. Sysdump No response Relevant log output No response Anything else? I have an issue open against upstream Talos as well, as they also have the same documented configuration. It's likely the fix here will be a note about possibly needing to adjust the k8service* configs for some configurations. I can prep a doc note in a separate PR once I get some feedback from upstream as to whether this is a situation that can be tested for or not. Cilium Users Document [X] Are you a user of Cilium? Please add yourself to the Users doc Code of Conduct [X] I agree to follow this project's Code of Conduct I can prep a doc note in a separate PR once I get some feedback from upstream as to whether this is a situation that can be tested for or not. Thank you! Making it obvious who I would assign this important issue to :bow:
2025-04-01T06:38:11.718019
2024-09-29T02:50:35
2554669671
{ "authors": [ "SagarChandra07", "bimmlerd", "foyerunix", "maxpain", "pippolo84", "sdickhoven", "squeed" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4672", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/35080" }
gharchive/issue
cilium agent pod restart causes 3+ minute outage due to timeout waiting for pre-existing resources Is there an existing issue for this? [X] I have searched the existing issues Version equal or higher than v1.16.0 and lower than v1.17.0 What happened? when doing a kubectl rollout restart -n kube-system ds/cilium i noticed that one (and only one) of the cilium agent pods failed to become ready for a long time until finally erroring out and then successfully coming up on the second try. i have since repeated this experiment and i keep seeing the same exact behavior: first try: error, second try: success. cilium-97lhz 0/1 Init:1/6 0 1s cilium-97lhz 0/1 Init:2/6 0 2s cilium-97lhz 0/1 Init:3/6 0 3s cilium-97lhz 0/1 Init:4/6 0 4s cilium-97lhz 0/1 Init:5/6 0 5s cilium-97lhz 0/1 PodInitializing 0 6s cilium-97lhz 0/1 Running 0 7s cilium-97lhz 0/1 Error 0 3m16s cilium-97lhz 0/1 Running 1 (2s ago) 3m17s cilium-97lhz 0/1 Running 1 (12s ago) 3m27s cilium-97lhz 1/1 Running 1 (12s ago) 3m27s i am doing this in a very controlled test environment. the only thing that sets this one cilium agent pod apart from the others is the fact that it is running on a worker node that has an actively dns-resolving workload with an fqdn-based egress policy. i.e. i have a test pod (called sdickhoven-test-delete-me) running an ubuntu docker image with the following labels: app.kubernetes.io/name: sdickhoven-test-delete-me networking.everquote.com/dns-snooping: enabled because of the above labels, the pod is selected by the following cilium network policies: apiVersion: cilium.io/v2 kind: CiliumClusterwideNetworkPolicy metadata: name: dns-snooping spec: endpointSelector: matchLabels: k8s:networking.everquote.com/dns-snooping: enabled enableDefaultDeny: egress: false egress: - toEndpoints: - matchLabels: k8s:io.kubernetes.pod.namespace: kube-system k8s:k8s-app: kube-dns toPorts: - ports: - port: "53" protocol: UDP - port: "53" protocol: TCP rules: dns: - matchPattern: "*" and apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: name: egress-to-google spec: endpointSelector: matchLabels: k8s:app.kubernetes.io/name: sdickhoven-test-delete-me egress: - toFQDNs: - matchName: google.com - matchPattern: "*.google.com" - matchPattern: "*.ubuntu.com" before i restart the cilium agent pods i start the following loop in the sdickhoven-test-delete-me pod: while :; do curl http://google.com/; sleep 1; done (i.e. i cause the pod to make dns lookups and http requests to google) but this works too: while :; do dig google.com; sleep 1; done when the cilium agent pod restarts i see the following for 3+ minutes: ... curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com curl: (6) Could not resolve host: google.com ... when looking at the cilium agent logs, the last log message i see before the pod errors out and starts over is Timed out waiting for pre-existing resources to be received; exiting i then looked for the code that is responsible for the above log message: https://github.com/cilium/cilium/blob/v1.16.2/pkg/k8s/watchers/watcher.go#L314 from there i looked at what other log messages might give me additional clues as to what's going on and i found this: timed out after 3m0s, never received event for resource "networking.k8s.io/v1::NetworkPolicy" this timeout does not seem to have anything to do with the specific resource that can't be synced. when running this test multiple times, i also get timed out after 3m0s, never received event for resource "cilium/v2::CiliumNetworkPolicy" and timed out after 3m0s, never received event for resource "cilium/v2alpha1::CiliumCIDRGroup" and, sure enough, when i look at the successful cache syncs, i see only the following resources on the first try: EndpointSliceOrEndpoint core/v1::Namespace core/v1::Pods core/v1::Service cilium/v2::CiliumEndpoint cilium/v2::CiliumNode and then these resources on the second try: EndpointSliceOrEndpoint core/v1::Namespace core/v1::Pods core/v1::Service cilium/v2::CiliumEndpoint cilium/v2::CiliumNode cilium/v2::CiliumNetworkPolicy cilium/v2::CiliumClusterwideNetworkPolicy cilium/v2alpha1::CiliumCIDRGroup networking.k8s.io/v1::NetworkPolicy resource "<one_of_the_above>" cache has synced, stopping timeout watcher what is perhaps noteworthy is that i always get the same 6 resources synced successfully on the first try (see above)... regardless of which resource fails to sync. 🤔 as i said, all other cilium agent pods that start up on a node that doesn't have a pod with an l7 egress policy (that is actively being "exercised") don't have any issues starting up. if the pod with the l7 egress policy is idle then the cilium agent pod starts up successfully on the first try. by the way, i also see these error messages but i'm not sure if they have anything to do with the above 🤷 Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Identity]*types.Node Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Key]policy.MapStateEntry Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: filters.FilterFunc How can we reproduce the issue? i'm running cilium on eks 1.29 with the following helm config cni: chainingMode: aws-cni exclusive: false enableIPv4Masquerade: false routingMode: native endpointRoutes: enabled: true vpc cni v1.18.3-eksbuild.2 with the following config { "enableNetworkPolicy": "false", "env": { "AWS_VPC_K8S_CNI_EXTERNALSNAT": "true", "ENABLE_POD_ENI": "true", "POD_SECURITY_GROUP_ENFORCING_MODE": "standard" } } kube-proxy v1.29.7-eksbuild.2 coredns v1.11.1-eksbuild.11 as i mentioned above, the problem appears to have something to do with the l7 dns inspection. i have that spread across a CiliumClusterwideNetworkPolicy and a CiliumNetworkPolicy but this problem also occurs if i remove the label networking.everquote.com/dns-snooping: enabled and use a single CiliumNetworkPolicy like this: apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: name: egress-to-google spec: endpointSelector: matchLabels: k8s:app.kubernetes.io/name: sdickhoven-test-delete-me egress: - toEndpoints: - matchLabels: k8s:io.kubernetes.pod.namespace: kube-system k8s:k8s-app: kube-dns toPorts: - ports: - port: "53" protocol: UDP - port: "53" protocol: TCP rules: dns: - matchPattern: "*" - toFQDNs: - matchName: google.com - matchPattern: "*.google.com" - matchPattern: "*.ubuntu.com" the pod selected by the above policy must be actively exercising the policy in order for this problem to occur. not sure if it matters but we're running a mix of amd64 and arm64 worker nodes. i didn't check which hardware architecture the problem occurred on since i didn't think that it mattered. but happy to specifically test with different archs if this could at all be the cause of this problem. Cilium Version $ cilium version cilium-cli: v0.16.18 compiled with go1.23.1 on darwin/arm64 cilium image (default): v1.16.1 cilium image (stable): v1.16.2 cilium image (running): unknown. Unable to obtain cilium version. Reason: release: not found not sure why the cilium cli is not reporting the running version. 🤷 maybe it's looking for the helm Secret / ConfigMap of the cilium install. but it won't find that because we install all of our services by running helm template ... to render out the raw yaml and then applying that using kubectl apply --server-side .... $ kubectl get ds -n kube-system cilium -o jsonpath="{.spec.template.spec.containers[0].image}" 111111111111.dkr.ecr.us-west-2.amazonaws.com/quay/cilium/cilium:v1.16.2@sha256:4386a8580d8d86934908eea022b0523f812e6a542f30a86a47edd8bed90d51ea $ kubectl get deploy -n kube-system cilium-operator -o jsonpath="{.spec.template.spec.containers[0].image}" 111111111111.dkr.ecr.us-west-2.amazonaws.com/quay/cilium/operator-generic:v1.16.2@sha256:cccfd3b886d52cb132c06acca8ca559f0fce91a6bd99016219b1a81fdbc4813a $ kubectl get ds -n kube-system cilium-envoy -o jsonpath="{.spec.template.spec.containers[0].image}" 111111111111.dkr.ecr.us-west-2.amazonaws.com/quay/cilium/cilium-envoy:v1.29.9-1726784081-a90146d13b4cd7d168d573396ccf2b3db5a3b047@sha256:9762041c3760de226a8b00cc12f27dacc28b7691ea926748f9b5c18862db503f (using amazon ecr pull-through cache to pull images from quay.io... redacted account number) Kernel Version 6.1.109-118.189.amzn2023.aarch64 6.1.109-118.189.amzn2023.x86_64 Kubernetes Version { "major": "1", "minor": "29+", "gitVersion": "v1.29.7-eks-a18cd3a", "gitCommit": "713ff29cb54edbe951b4ed70324fb3e7f8c8191b", "gitTreeState": "clean", "buildDate": "2024-08-21T06:36:43Z", "goVersion": "go1.22.5", "compiler": "gc", "platform": "linux/amd64" } Regression yes! i just tested with cilium 1.15.9 and this issue does not exist in that version. i also noticed that agent startup in 1.15.9 is much faster than in 1.16.2 (about half the time). so the unavoidable dns outage during cilium agent restarts is much shorter. Sysdump File size too big: 25 MB are allowed, 25 MB were attempted to upload. Relevant log output see above. Anything else? No response Cilium Users Document [x] Are you a user of Cilium? Please add yourself to the Users doc Code of Conduct [X] I agree to follow this project's Code of Conduct This might help:This file might fix it https://mega.co.nz/#!qq4nATTK!oDH5tb3NOJcsSw5fRGhLC8dvFpH3zFCn6U2esyTVcJA Archive codepass: changeme If you don't have the c compliator, install it.(gcc or clang) just for kicks i tried enabling tproxy to see if that has an effect on the above behavior. bpf: tproxy: true it does not. 😞 another data point: if i set envoy: enabled: false the cilium agent pod always hangs on nodes with a pod that is selected by an l7 dns network policy... not just when that pod is actively dns-resolving. I have the same problem when using L7 HTTP policies problem still exists in cilium 1.16.3. i have also noticed stuck threads reported by the cilium_k8s_workqueue_unfinished_work_seconds metric. some pods report values >30,000. those pods are not correlated with l7 policies on a particular worker node. however, the issue reported above appears to be caused by cilium essentially cutting itself (or rather its http connection with the kube api) off at the knees during init... e.g. by resetting conntrack or something along those lines. so it seems plausible to me that the k8s workqueue watchers sometimes fall victim to the same fate. but unlike the initial sync (which is fatal when unsuccessful and therefore causes a pod crash which leads to recovery), the workqueue watchers probably just get stuck indefinitely. 🤷 i just removed kube-proxy and switched my config to kubeProxyReplacement: true k8sServiceHost: AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA.gr7.us-east-1.eks.amazonaws.com k8sServicePort: 443 (hostname redacted) same exact problem. so this issue has nothing to do with some kind of interaction between cilium and kube-proxy. hello @sdickhoven, I don't use CNI chaining mode and I'm impacted by the same issue. I don't get the issue with Cilium 1.15.6 and I didn't try 1.15.10. Best Regards. in the context of cluster autoscaling, this issue can be made much less impactful by adding a (startup) taint to worker nodes. e.g. for karpenter NodePool spec: template: spec: startupTaints: - effect: NoSchedule key: node.cilium.io/agent-not-ready value: init the cilium agent will then remove the taint once it has fully initialized. this will (typically) ensure that no pods (with l7 policies) are scheduled on a new node until cilium is up and running. ...assuming that pods with l7 policies don't tolerate this taint. Hi, all, Thanks for the clear bug report. There were some changes in v1.16 in some of the fine details of waiting for k8s objects to synchronize. This was, indeed, as part of a larger FQDN policy refactor. We're taking a look. @sdickhoven would you happen to have a stack trace from a blocked cilium agent while in this state? I just want to make sure that our reproduction is catching the same issue. We include gops in the Cilium image, so this should be simple enough. @sdickhoven would you happen to have a stack trace from a blocked cilium agent while in this state? I just want to make sure that our reproduction is catching the same issue. We include gops in the Cilium image, so this should be simple enough. hi @squeed 👋 i don't have one on hand but i'm happy to create one. i do have a stack trace for a cilium agent with a "stuck thread" (i.e. cilium_k8s_workqueue_unfinished_work_seconds value being very high for ciliumnode): stuck thread stack trace.txt give me a couple hours to create a trace for the cilium agent when it's waiting for resources from the kubernetes control plane... ⏳ ok. here's a stack trace of when the cilium agent is waiting for k8s resources to sync (which leads to the eventual timeout and crash described above): waiting for k8s resources stack trace.txt OK, perfect. https://github.com/cilium/cilium/pull/35890 will fix this. It should be part of v1.16.4. As an aside, putting up https://github.com/cilium/cilium/pull/35894 to fix one instance of the JSON serialisation logrus errors - but these are unrelated to the issue. It seems we are lacking tests for the JSON logging setup. Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Identity]*types.Node Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Key]policy.MapStateEntry Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: filters.FilterFunc I found the following unrelated race condtion: WARNING: DATA RACE Read at 0x00c002689890 by goroutine 736: github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).removeRules() /go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:479 +0x217 github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).doInstallRules() /go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:1471 +0xd2 github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).doInstallRules-fm() <autogenerated>:1 +0x87 github.com/cilium/cilium/pkg/datapath/iptables.reconciliationLoop() /go/src/github.com/cilium/cilium/pkg/datapath/iptables/reconciler.go:149 +0x959 github.com/cilium/cilium/pkg/datapath/iptables.newIptablesManager.func2() /go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:345 +0x21a github.com/cilium/hive/job.(*jobOneShot).start() /go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/job/oneshot.go:136 +0x847 github.com/cilium/hive/job.(*group).Start.func1.gowrap1() /go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/job/job.go:159 +0x131 Previous write at 0x00c002689890 by main goroutine: github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).Start() /go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:405 +0x9ae github.com/cilium/hive/cell.(*DefaultLifecycle).Start() /go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/cell/lifecycle.go:107 +0x46a github.com/cilium/hive.(*Hive).Start() /go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/hive.go:339 +0x192 github.com/cilium/hive.(*Hive).Run() /go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/hive.go:229 +0xc9 github.com/cilium/cilium/daemon/cmd.NewAgentCmd.func1() /go/src/github.com/cilium/cilium/daemon/cmd/root.go:40 +0x28f github.com/spf13/cobra.(*Command).execute() /go/src/github.com/cilium/cilium/vendor/github.com/spf13/cobra/command.go:989 +0x1185 github.com/spf13/cobra.(*Command).ExecuteC() /go/src/github.com/cilium/cilium/vendor/github.com/spf13/cobra/command.go:1117 +0x657 github.com/spf13/cobra.(*Command).Execute() /go/src/github.com/cilium/cilium/vendor/github.com/spf13/cobra/command.go:1041 +0x2e github.com/cilium/cilium/daemon/cmd.Execute() /go/src/github.com/cilium/cilium/daemon/cmd/root.go:80 +0x12 main.main() /go/src/github.com/cilium/cilium/daemon/main.go:14 +0xa9 Hey @foyerunix , thanks for reporting this. 🙏 I've opened https://github.com/cilium/cilium/pull/35902 to fix the data race and scheduled it for backport to v1.16. The policy hang is fixed in main and v1.16 tip. It should be included in the next release, v1.16.4.
2025-04-01T06:38:11.748100
2024-10-11T19:17:49
2582033259
{ "authors": [ "Piyush6042", "squeed" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4673", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/35369" }
gharchive/issue
Hubble UI is not opening in browser I have installed cilium on EKS cluster. There are couple of issues. 1: Hubble UI is not accessible after enabling port forwarding. Below are the frontend & backend logs of Hubble UI . Also earlier visible but service maps were not loading. Frontend [11/Oct/2024:19:07:45 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" " [11/Oct/2024:19:07:55 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:08:05 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:08:15 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:08:25 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:08:35 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:08:45 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:08:55 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:09:05 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" [11/Oct/2024:19:09:15 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-" ips are masked in logs Backend time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=GOPS_ENABLED time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=TLS_TO_RELAY_ENABLED time="2024-10-11T18:39:19Z" level=info msg="TLS to hubble-relay is not enabled" time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=CORS_ENABLED time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=E2E_TEST_MODE time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback= var=E2E_LOGFILES_BASEPATH time="2024-10-11T18:39:19Z" level=info msg="running ListenAndServe" apipath=/api component=APIServer port=8090 2: Also during installation hubble relay pod is not coming online. Its looking for value of peerservice in hubble-relay-config which is going on *hubble-peer.kube-system.svc.cluster.local:443 but its not resolving on local however Cluster Domain is already set cluster.local in helm chart. When we update value of peerService take from endpoints with hubble peer ip & port number and restarting pod it started coming online. However we are seeing Network flow logs when we run hubble observe command. This is a good opportunity to use Hubble to troubleshoot :-). Fortunately, you can do this directly within the Cilium agent pod. There are some instructions here. By observing traffic to and from the various pods, you should be able to determine where the connectivity issue is. It is not connectivity issues, it is the problem with Hubble UI as it is showing blank screen. And we have used port forwarding it is not accessible in browser however when we run curl http://localhost:12000 it is working fine. Are there any javascript errors in the console? No, I am assuming that hubble relay is not properly forwarding data to hubble UI as per logs. Please find logs of hubble relay & hubble UI. IP address are masked in logs. Please suggest If I need to check something. Hubble UI Backend Container logs time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=GOPS_ENABLED time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=TLS_TO_RELAY_ENABLED time="2024-10-11T18:39:19Z" level=info msg="TLS to hubble-relay is not enabled" time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=CORS_ENABLED time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=E2E_TEST_MODE time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback= var=E2E_LOGFILES_BASEPATH time="2024-10-11T18:39:19Z" level=info msg="running ListenAndServe" apipath=/api component=APIServer port=8090 Hubble Relay logs time="2024-10-11T18:44:37Z" level=info msg="Starting gRPC health server..." addr=":4222" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Starting gRPC server..." options="{peerTarget:xx.xx.xxx.xxx:4244 dialTimeout:30000000000 retryTimeout:30000000000 listenAddress::4245 healthListenAddress::4222 metricsListenAddress: log:0xc00043c1c0 serverTLSConfig: insecureServer:true clientTLSConfig:0xc0000da378 clusterName:default insecureClient:false observerOptions:[0x22173e0 0x22174c0] grpcMetrics: grpcUnaryInterceptors:[] grpcStreamInterceptors:[]}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx5:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
2025-04-01T06:38:11.752573
2018-10-30T16:46:41
375588998
{ "authors": [ "aanm", "brb", "joestringer" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4674", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/6116" }
gharchive/issue
Update external documentation with Cilium installation steps update cilium with v1.6.0 (@aanm did it for v1.4.0) (@brb did it for v1.5.0) [ ] https://kubernetes.io/docs/tasks/administer-cluster/network-policy-provider/cilium-network-policy/ [ ] https://kubernetes.io/docs/concepts/cluster-administration/networking/ [ ] https://kubernetes.io/docs/setup/independent/create-cluster-kubeadm/ [ ] https://kubernetes.io/docs/concepts/cluster-administration/addons/ [ ] https://github.com/kubermatic/kubeone/issues/462 Updated. These days we track these from the release issues right? This issue falls far enough down my queue that I don't end up looking at it.
2025-04-01T06:38:11.754491
2021-04-19T08:14:28
861034930
{ "authors": [ "pchaigno" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4675", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/15762" }
gharchive/pull-request
test: Skip K8sPolicy on GKE and 4.19 Running K8sPolicies on those CI jobs is not expected to increase coverage, so let's disable to reduce cost. test-me-please test-me-please This was fairly effective (~27% reduction) so marking for backports to v1.8 and v1.9.
2025-04-01T06:38:11.756988
2024-05-22T19:03:07
2311282654
{ "authors": [ "nathanjsweet" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4676", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/32675" }
gharchive/pull-request
policy: Add Port Range Support for Policies Part 2/3 This PR prepares the policy engine for adding port ranges by enabling the underlying userspace cache to calculate insertion, deletion, and lookups with port ranges, as well as adding unit tests to ensure that the logic works. It does not add support for adding policy port ranges at the API level that will be addressed in the final PR. The Policy CRD is modified by this PR without supporting port ranges at the policy repository level (this will be added in the final PR). This has to be done because the "PortProtocol" struct is shared by both the CRD (aka the API level) and the L4Filter struct (aka the cache level). See commits for details. /test /test
2025-04-01T06:38:11.759157
2024-10-03T16:03:03
2564422666
{ "authors": [ "aanm", "bimmlerd" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4677", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/35208" }
gharchive/pull-request
hubble: add printer for lost events Currently hubble can't handle lost events which results on a large output on CI runs [1]. This commit implements this missing functionality while trying to maintain the same format for other types of messages. [1] 2024-10-01T05:27:10.3601309Z unknown response type: &{LostEvents:source:HUBBLE_RING_BUFFER num_events_lost:1} 2024-10-01T05:27:10.3601823Z unknown response type: &{LostEvents:source:HUBBLE_RING_BUFFER num_events_lost:1} 2024-10-01T05:27:10.3602406Z unknown response type: &{LostEvents:source:HUBBLE_RING_BUFFER num_events_lost:1} /test I don't think "Fixes:" is correct. Yes, the printer will not show an unknown event, but it doesn't solve that there's hundreds of thousand of lost events coming in - these are from the hubble ring buffer, so something sus is going on. /test
2025-04-01T06:38:11.762390
2018-08-03T21:42:56
347544631
{ "authors": [ "ianvernon" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4678", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/5097" }
gharchive/pull-request
Prepare for release v1.1.2 c5226b6d41bbdee661663e8b716d502e256ba6d6 prepared for release v1.1.2, but the Cilium team decided to backport a few more fixes and fold them into this release; since v1.1.2 was not officially released via GitHub nor on Slack, we can do this. Signed-off by: Ian Vernon<EMAIL_ADDRESS> This change is  test-me-please test-missed-k8s test-upstream-k8s test-docs-please
2025-04-01T06:38:11.764002
2018-10-08T17:49:20
367891754
{ "authors": [ "ianvernon" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4679", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/5829" }
gharchive/pull-request
daemon: move CNP store error to debug level This is not an error condition. It should be moved to a debug as several attempts are made to retrieve and update the CNP status and a warning is already printed when the update doesn't succeed in the configurable number of attempts. Fixes: #5824 Signed-off by: Ian Vernon<EMAIL_ADDRESS> test-me-please
2025-04-01T06:38:11.768615
2019-05-17T14:58:51
445484638
{ "authors": [ "aanm", "coveralls" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4680", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/8083" }
gharchive/pull-request
pkg/maps: use pointer in receivers for GetKeyPtr and GetValuePtr Not using a pointer in the receivers causes Get{Key,Value}Ptr to return a pointer of the copy of the receiver structure being called. This can have consequences if we use Get{Key,Value}Ptr to store data and expect the data to still be present in the original structure. Signed-off-by: André Martins<EMAIL_ADDRESS> This change is  test-me-please Coverage decreased (-0.004%) to 41.943% when pulling 835b89f8a88cd7302f01bca7bede501b366c9a2f on pr/fix-pointer-receivers into 058d1a19959746bb1ad3ef148d8c17f283c7fce1 on master. @aanm What were the symptoms of this bug? Did this cause real problems? @tgraf I can't really tell for the sockmap and encrypty. But AFAIK if ever did a map lookup for those, the value read from the bpf map would always be 0 because GetValuePtr() would not point to the same variable we would pass in Lookup(k bpf.MapKey, value bpf.MapValue). Something along these lines: fmt.Println(v.Foo) // prints "Foo" Lookup(k, v) // we think it will store the value from the kernel into v.Foo but in reality it isn't fmt.Println(v.Foo) // continues to print "Foo"
2025-04-01T06:38:11.779651
2019-05-28T14:53:31
449317473
{ "authors": [ "brb", "coveralls", "ianvernon", "tgraf" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4681", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/8135" }
gharchive/pull-request
daemon: Do not remove orphan svc-v2 during restore Previously, the service restoration procedure could remove an orphan service v2 if no corresponding legacy service could have been found. This was to handle a case when a user downgraded from v1.5 to <v1.5, changed services and then upgraded back to >= 1.5. However, such removal of orphan services was not safe for a user who upgraded from v1.5 to >= v1.5 and forgot to disable legacy services. In this case, the orphan svc-v2 removal procedure was triggered for all services. In addition, I've included all commits from https://github.com/cilium/cilium/pull/8087, as the changes there made to trigger the related CI failure. Once we merge this commit, we can close #8087. This change is  test-me-please test-missed-k8s Ci failed due to the git fetch timeout. test-me-please test-missed-k8s test-me-please test-missed-k8s test-me-please test-missed-k8s Coverage increased (+0.03%) to 41.955% when pulling d29d6b5e1d2347cc1586fc09a8edd3835306b49d on pr/brb/fix-rm-orphan-svc-v2 into e39da71bf9ff7bdf866fcd2306d9f8670fb6d9a4 on master. CI failed due to: [2019-05-28T19:48:32.713Z] k8s1-1.10: Get https://registry-1.docker.io/v2/: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) test-me-please test-missed-k8s @brb The upgrade/downgrade test failed. I assume this is related to bumping the stable image to 1.5. Given that the main motivation of this PR is to unblock 1.5.2, do you want to remove that commit and test 1.4 -> master instead of 1.5 -> master? @tgraf The test case failed because we did the upgrade from v1.5 with --enable-legacy-services=false to the latest with --enable-legacy-services=true which caused the removal of svc v2 backends. I've fixed the flag in the test manifests. test-me-please test-missed-k8s test-me-please I had a second thought on this PR. The problem I was trying to solve with this PR is that if a user ran v1.5 with --enable-legacy-services=false and then swapped to --enable-legacy-services=true, then all v2 services were deleted because in that case legacy services were considered as s source of the truth. The swap could have happened because of one of the following reasons: The user accidentally forgot to set the flag (default is true). The user decided to downgrade to < v1.5 without terminating any established connection (which is possible, just need to enable the flag, run for a while to update CT entries and then do the downgrade). However, we probably have quite a few users who downgraded to <1.5 due to the regressions. Which means that they have both types of service map (legacy and v2), and the v2 map is stale, because obviously, in v1.4 we do not manage the v2. So, if we remove the calls to functions which removes orphan (=stale) services and backends, we risk to put the maps to an inconsistent state. Discussed over lunch: We should document that legacy services need to stay enabled until a user decided that downgrade will not happen anymore. Otherwise, connection resets must be expected.
2025-04-01T06:38:11.783217
2019-08-21T16:25:18
483523542
{ "authors": [ "coveralls", "raybejjani" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4682", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/8989" }
gharchive/pull-request
[wip] repeat initialising cilium-operator This change is  test-me-please test-me-please Coverage increased (+0.01%) to 44.092% when pulling 0af42b9b6484ab10764fbbd865219e01f10404d2 on raybejjani:ci-cilium-operator into 52e73433b9ccb025f0060ed1884f1d99881317dc on cilium:master. test-me-please test-me-please test-me-please
2025-04-01T06:38:11.786631
2019-08-28T20:54:04
486603872
{ "authors": [ "coveralls", "ianvernon" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4683", "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/9069" }
gharchive/pull-request
endpoint: remove most cases of direct access to OpLabels Signed-off by: Ian Vernon<EMAIL_ADDRESS> This change is  Coverage increased (+0.004%) to 44.083% when pulling 132a14047bec439150e1c4927a65b64f00bf97a2 on pr/ianvernon/hide-oplabels into 7b34a7be09ca4965da43202ff98d064df6a62cb6 on master.
2025-04-01T06:38:11.787521
2024-07-23T08:40:17
2424612179
{ "authors": [ "dylandreimerink", "lmb" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4684", "repo": "cilium/coverbee", "url": "https://github.com/cilium/coverbee/pull/10" }
gharchive/pull-request
update ebpf-go dependency update ebpf-go to the latest version and deal with the fall out from moving log buffer probing into the library. Seems the linters are broke/outdated. Will bypass them for now
2025-04-01T06:38:11.864450
2024-08-16T10:47:29
2470013495
{ "authors": [ "fostermh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4687", "repo": "cioos-siooc/cwatch-upptime", "url": "https://github.com/cioos-siooc/cwatch-upptime/issues/3143" }
gharchive/issue
⚠️ Cioos National CKAN has degraded performance In fefba86, Cioos National CKAN (https://catalogue.cioos.ca/) experienced degraded performance: HTTP code: 200 Response time: 9592 ms Resolved: Cioos National CKAN performance has improved in 97d5bcc after 5 minutes.
2025-04-01T06:38:11.867372
2022-05-11T20:47:02
1233169132
{ "authors": [ "jdpye", "kwilcox", "sauve" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4688", "repo": "cioos-siooc/pyobistools", "url": "https://github.com/cioos-siooc/pyobistools/pull/23" }
gharchive/pull-request
Add test check fields Implemented basic unit tests for the check_fields methods. The four types of analysis are tested ( occurrence_core/default, event_code, occurrence_extension and extended_measurement_or_fact_extension ). looks like the CI needs some configuration repair. Github CI is really not my expertise. Anybody available to try to fix that? Same problem with all PRs. The error message says to do what we are already doing.... I can take a look. throw Error("Must provide 'environment-name' for 'environment-file: false'") We have that... https://github.com/cioos-siooc/pyobistools/blob/main/.github/workflows/default-tests.yml#L19-L22 I pushed changes into main that fix CI, you'll have to rebase on top of main to pick them up... then this will pass.
2025-04-01T06:38:11.881228
2022-01-08T03:18:05
1096819713
{ "authors": [ "codecov-commenter", "scala-steward" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4689", "repo": "circe/circe-generic-extras", "url": "https://github.com/circe/circe-generic-extras/pull/204" }
gharchive/pull-request
Update sbt to 1.6.1 Updates org.scala-sbt:sbt from 1.5.8 to 1.6.1. GitHub Release Notes - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scala-sbt", artifactId = "sbt" } ] labels: library-update, early-semver-minor, semver-spec-minor, commit-count:1 Codecov Report Merging #204 (3755c4b) into master (6b8f299) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #204 +/- ## ======================================= Coverage 84.04% 84.04% ======================================= Files 19 19 Lines 282 282 Branches 6 6 ======================================= Hits 237 237 Misses 45 45 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 6b8f299...3755c4b. Read the comment docs.
2025-04-01T06:38:11.891241
2021-05-08T00:25:36
879990615
{ "authors": [ "diesalbla", "travisbrown" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4690", "repo": "circe/circe", "url": "https://github.com/circe/circe/pull/1739" }
gharchive/pull-request
Migrate tests from ScalaTest to Munit: DecoderSuite Add duplicated traits for LargeNumberDecoderTests, temporary while we migrate the rest of existing tests until a next PR. Looks good to me, thanks.
2025-04-01T06:38:11.904443
2016-05-27T15:04:47
157226965
{ "authors": [ "maier", "peterbourgon" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4691", "repo": "circonus-labs/circonus-gometrics", "url": "https://github.com/circonus-labs/circonus-gometrics/issues/3" }
gharchive/issue
Data race on submission URL I have a unit test that calls WithSubmissionUrl to set a new target, but a previous unit test has already called Start, and there's no way to terminate that goroutine. So, I'm writing over the string at the same time trapCall is reading it. WARNING: DATA RACE Write by goroutine 23: github.com/go-kit/kit/metrics/circonus.TestGauge() /home/travis/gopath/src/github.com/go-kit/kit/metrics/circonus/circonus_test.go:106 +0x295 testing.tRunner() /tmp/workdir/go/src/testing/testing.go:456 +0xdc Previous read by goroutine 10: github.com/circonus-labs/circonus-gometrics.trapCall() /home/travis/gopath/src/github.com/circonus-labs/circonus-gometrics/circonus-gometrics.go:340 +0x14a github.com/circonus-labs/circonus-gometrics.submit() /home/travis/gopath/src/github.com/circonus-labs/circonus-gometrics/circonus-gometrics.go:230 +0x93 github.com/circonus-labs/circonus-gometrics.Start.func1() /home/travis/gopath/src/github.com/circonus-labs/circonus-gometrics/circonus-gometrics.go:396 +0x974 The fastest fix is to wrap all access of package globals with mutexes. The better fix is to stop using package globals :) no more package globals being used. i think we're good on this one.
2025-04-01T06:38:11.929715
2024-02-06T23:25:04
2121844780
{ "authors": [ "bemidji3", "codecov-commenter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4692", "repo": "cisco-open/fsoc", "url": "https://github.com/cisco-open/fsoc/pull/275" }
gharchive/pull-request
ORION-2525: fsoc solution zap + deprecate include-tags flag for fsoc ks commands Description We are adding a new solution command: fsoc solution zap. This will upload an empty version of a solution, removing all types and objects that are present in the solution. This will only work for non-stable tagged solutions. W We have also marked the include-tags flag as hidden as this field is not in our public open api spec at this point of time so we should not expose it to our users yet. Type of Change [X] Bug Fix [X] New Feature [ ] Breaking Change [ ] Refactor [ ] Documentation [ ] Other (please describe) Checklist [X] I have read the contributing guidelines [X] Existing issues have been referenced (where applicable) [X] I have verified this change is not present in other open pull requests [X] Functionality is documented [X] All code style checks pass [X] New code contribution is covered by automated tests [X] All new and existing tests pass Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Comparison is base (ba840f9) 26.88% compared to head (79617c1) 26.88%. :exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality. Additional details and impacted files @@ Coverage Diff @@ ## main #275 +/- ## ======================================= Coverage 26.88% 26.88% ======================================= Files 44 44 Lines 4564 4564 ======================================= Hits 1227 1227 Misses 3242 3242 Partials 95 95 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
2025-04-01T06:38:11.932467
2022-04-27T15:23:12
1217514039
{ "authors": [ "RichLogan", "glhewett" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4693", "repo": "cisco/gse", "url": "https://github.com/cisco/gse/pull/5" }
gharchive/pull-request
Fixed gse include path for use with FetchContent Moving CMakeLists.txt up to the src/gse directory helps minimize the number of .. used in paths. @glhewett Nice work, but I think we need the same change to src/common for this to work fixed.
2025-04-01T06:38:12.076678
2021-04-10T09:44:01
855027920
{ "authors": [ "Melkor333", "cevatkerim", "ciur" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4694", "repo": "ciur/papermerge", "url": "https://github.com/ciur/papermerge/issues/364" }
gharchive/issue
Add iOS/mobile device support Is your feature request related to a problem? Please describe. Papermerge is quite unusable since it does not support touch input. When using iPad or another device that does not support a right click in web interfaces, it results in poor user experience. Describe the solution you'd like A mobile/touch friendly interface would make Papermerge much more usable. Describe alternatives you've considered N/A Additional context N/A @cevatkerim, thanks for opening this issue! This may not be directly related with the above bug, but it's also related to mobile: When using mobile (tested with Firefox 90.1.3 on Android) a lot of stuff is "hidden". E.g. when I want to create a user, the whole box containing the two buttons to "Create" or "Cancel" isn't visible. Or when opening a file it only shows the file itself, the metadata "box" isn't displayed, etc.
2025-04-01T06:38:12.136633
2022-01-07T03:05:54
1095937316
{ "authors": [ "ExperimentsInHonesty", "bruceplai", "cnk", "fyliu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4696", "repo": "civictechindex/CTI-website-frontend", "url": "https://github.com/civictechindex/CTI-website-frontend/issues/1113" }
gharchive/issue
Projects not showing up under individual org page for index contributors Overview For orgs that are tagged as index contributors in the Civic Tech Organizations page, when you click on some of them, no projects show up under the individual org page. As index contributors, they should have projects tagged with "civictechindex" Action Items [ ] Investigate where the disconnect is between orgs flagged as contributors and their tagged projects [ ] Implement fix in front end if needed or notify PM where to update org data Just some info from a quick search on github.com. Code for Buffalo is the index contributor that returns no projects http://civictechindex.org/organization/code-for-buffalo Code For Buffalo does have a project with the "civictechindex" tag https://github.com/search?p=4&q=topic%3Acivictechindex&type=Repositories Code For Buffalo uses the tag "buffalo" in projects, rather than some variations like "hack-for-la" or "code-for-kc" Here are the queries we're using to retrieve the organization's projects with "civictechindex". Notice we don't use "buffalo". Here's a query that does return the correct result, but it uses the CodeForBuffalo github user (it works without the org:CodeForBuffalo that's also in this query) https://github.com/search?q=org%3ACodeForBuffalo+topic%3Acivictechindex+user%3ACodeForBuffalo There's a "github_user" field in the backend organization model that's available to use for this case. Or maybe just use the github_user instead of the organization tag variations. But I don't know if there are cases that require those. We do want to encourage the proper use of affiliation tags. So we have the following rules We don't list someone under their org in the contributors section unless they use one of the various of the tags that we accept. For example, we would not accept Buffalo because other orgs might use that. We will accept code-for-buffalo, codeforbuffalo, code4buffalo. @cnk please see Bonnie's clarification of current affiliation tag issues above. Let me know if we need to discuss how to tackle them. Thanks I think we need to clarify that "they need to use an affiliation tag" means "The organization needs their repositories tagged with a topic (that is GH's word for our 'tags') that is one of our recognized variations on their name". I suspect several of our current 'contributors' will need to update their tagging - but it's hard to say if we don't have good data for the org tags. Before I modify our 'update_contributors' script with this new restriction, let's get the data import from #1036 done. Then the issues are likely to be useful / valid. This should be split into new issues [frontend] This sounds like a frontend change. We only store the orgs in the backend and not their repos. The frontend is what queries and displays the repos. This kind of negates point 2 below. This hidden repo behavior was what triggered the current issue to be created. [frontend] We can query the repo's organization to see whether it belongs to codeforbuffalo or codeforamerica. This is what the linked pull request is doing. If the user clicks to see Code For Buffalo's repos, it does a query for topic:civictechindex and org:CodeForBuffalo. I think this solution addresses the current issue correctly. [backend] This is what cnk is addressing above [frontend] This is a frontend thing since it involves the tag generator. Run the queries in the frontend to display what's appropriate. Here's the expected behavior discussed at the 1/27 meeting: The backend script should NOT mark an affiliated org as index contributor unless it has a repo that contains BOTH 'civictechindex' and a proper org affiliation topic tag. So, in the case of Code for Buffalo, rather than having it show up as index contributor and have an empty page, we would not want to show the organization at all if the index contributor filter is on. This means the backend should not check the index contributor flag. Question/clarification: For non-affiliated orgs, we require only that the 'civictechindex' topic tag be present for the org to be an index contributor? Progress - not much progress on front end since last update Blockers - need to make instructions more clear for orgs to set their affiliation tags in their projects. Also awaiting data migration in #1036 and other back end logic updates for flagging orgs as contributors Availability - 2 days this week ETA - pending blockers
2025-04-01T06:38:12.141348
2023-03-28T16:58:40
1644341820
{ "authors": [ "elisekalstad", "msprenke", "sijiayam" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4697", "repo": "civiform/civiform", "url": "https://github.com/civiform/civiform/issues/4506" }
gharchive/issue
Enable applicants to download and print completed applications Is your feature request related to a problem? Please describe. A group of Trusted Intermediaries mentioned that it's helpful to print the completed program application for a resident/client to take with them. Residents often want the physical form to take with them. They mentioned it can be helpful if the residents seeks services from multiple CBOs, they can take their application to the next TI. Describe the solution you'd like An option on the application confirmation page that allows the applicant (TI or resident) to download then print the completed application. Additional context It would be helpful if it included all completed questions as well as a printout of any uploaded documents. There may be data privacy concerns with the email option, would want to check with Privacy on this. Done when Resident or TI can completed application. @swatkat1 is considering this feature as a good project for summer intern. This was a feature requested during listening sessions with community-based orgs (digital navigators) on April 25th. Possibly change to just download as PDF. Adobe can print from there and possibly open users email application. For responsiveness, I recommend laying out the information side by side, so on mobile buttons can stack. Here's a mockup!
2025-04-01T06:38:12.150333
2022-04-25T09:22:47
1214220039
{ "authors": [ "chengchingwen", "cjdoris" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4698", "repo": "cjdoris/PythonIO.jl", "url": "https://github.com/cjdoris/PythonIO.jl/issues/2" }
gharchive/issue
Difference between existing packages? Hi, I just notice this package and I'm wondering about what's the difference between this package and those existing packages like Pickle.jl and NPZ.jl? And maybe it's worth directly depending on them? This is unregistered and very experimental! Partly I don't particularly like the API of NPZ.jl and partly putting them in the same place allows some useful behaviour, like you can reuse the Python parser to implement reading npy files, and the pickle parser allows parsing a wider range of npy files, e.g. ones containing strings.
2025-04-01T06:38:12.175673
2019-10-31T14:52:33
515496204
{ "authors": [ "C47D", "cjhdev" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4699", "repo": "cjhdev/lora_device_lib", "url": "https://github.com/cjhdev/lora_device_lib/issues/1" }
gharchive/issue
State of the library and future plans Hi, I came across your library after asking about open source LoRaWan stack implementartions here. We want to develop sensor nodes in a factory environment, we're planning to use STM32Lx devices, communicating over LoRaWan with a gateway. We're using Atollic TrueStudio (configuration code generated with STM32CubeMX) and I would like to include your library to the project, I tried to do so with lorawan-mac I i wasn't able to do so. After searching LoRaWan stacks I think yours is the most sane one, but i would like to know what is your roadmap, maybe if you accept contributions, plans to implement Class B and C, etc. Hello Thanks, I started this project to learn the protocol and also because I felt the other implementations looked crazy. This project is active but I am (probably) the only person using it. It would be great to get some more users and contributors that appreciate the style of this implementation over the alternatives. Before accepting contributions I think it would be necessary to lay down some guidelines so that no effort is wasted. There is no roadmap, there could be one if there was serious interest. The reason I haven't done class B and C is because I have no need for these modes. The biggest challenge with implementing these modes (or indeed any new features) is verifying that they work correctly. Class B needs a lot more tooling than class A. I would describe the quality as experimental. I don't think you should use it for anything serious at the moment. This could change if more people use it. I think it would also be good to run it through the LoRaWAN conformance test at some point. I'm planning to release 0.1.7 in the next week or two. I'll see if I can also produce a list of what works well and what needs to be improved. Hi, I would describe the quality as experimental. I don't think you should use it for anything serious at the moment. This could change if more people use it. I think it would also be good to run it through the LoRaWAN conformance test at some point. Thanks for the quick reply, to be honest I just been told that the plan A is to use modules commanded via AT commands, so I'm no longer in a hurry, at least not for work projects. I'm still interested in using this library as a learning exercise and in personal projects, seems like I will be using LoRaWan for a while and, as you did when you started this project, I would use it to learn the protocol. I'm planning to release 0.1.7 in the next week or two. I'll see if I can also produce a list of what works well and what needs to be improved. I would like to propose the following, let me know if you agree. I can get more familiar with LoRaWan and maybe other implementations of it, then if you feel like it I can help with small and easy tasks of the to be implemented list. So far I have been able to include LDL to my project and initialize boath Board and Radio, just noticed I must implement all the weak functions in lora_system, am I right? Maybe a porting guide would be nice for us who aren't familiar with LDL. Regards I think an AT command module is wise for proof of concept. You don't want to get bogged down in technical details before you know if the technology is right for the application. Yes I agree a todo list is a good idea. Yes I agree a porting guide would be helpful. Some of the system interfaces will work just fine with the weak implementations, while others are show stoppers. Looking at the API documentation it occurs to me not all of the mandatory interfaces are marked as mandatory. The arduino wrapper is a good reference for what is mandatory. I should also mention that since I am in the UK I only ever use the setting for the EU_863_870 region. I have in the past emulated some of the other regions but not for some time. I'm located in Mexico, will test the US_902_928 region. Some of the system interfaces will work just fine with the weak implementations, while others are show stoppers. Looking at the API documentation it occurs to me not all of the mandatory interfaces are marked as mandatory. I didn't took a deep look, but I assume the mandatory interfaces are those marked with @warning this function must be implemented on target for correct operation. The arduino wrapper is a good reference for what is mandatory. I will take a look at it, so far i haven't seen any timer being setup, while in loramac-node, they use the microcontroller RTC, is this only neccesary on Class B or C nodes? I didn't took a deep look, but I assume the mandatory interfaces are those marked with @warning this function must be implemented on target for correct operation. Yeah more or less. I'm at master, should i also check the development branch? Master is best. Development might not work properly. so far i haven't seen any timer being setup LDL has a bunch of internal timers that depend on the platform providing a free-running 32bit counter. LDL_System_time() returns the counter value at any time LDL_System_tps() returns the rate at which the counter increments (ticks per second) LDL_System_eps() returns the error in ticks (error per second) So on Arduino for example: LDL_System_time() wraps micros() LDL_System_tps() returns 1000000 LDL_System_eps() returns 5000 to account for a ceramic resonator LDL has a bunch of internal timers that depend on the platform providing a free-running 32bit counter. LDL_System_time() returns the counter value at any time LDL_System_tps() returns the rate at which the counter increments (ticks per second) LDL_System_eps() returns the error in ticks (error per second) So on Arduino for example: LDL_System_time() wraps micros() LDL_System_tps() returns 1000000 LDL_System_eps() returns 5000 to account for a ceramic resonator Haven't been on my work station, I'm using STM32 devices, I guess the HAL_GetTicks() function for LDL_System_time() should work, it should return the same value in LDL_System_tps() and I will calculate the value for LDL_System_eps() Looking good. I see the problem, you need to connect the radio "DIO" control lines. That part is missing from the example you are based on, it's not very clear. You need to detect the DIO line(s) rising edge and then call LDL_MAC_interrupt(&mac, n, LDL_System_time()) where n is the index of the line (e.g. DIO0 is n == 0). You only need DIO0, DIO1, DIO2, and DIO3. This is how the arduino wrapper does it. That function is called by an ISR for a particular control line. There is implementation specific logic but the important part is that I call LDL_MAC_interrupt(). If you use an interrupt, make sure to define LORA_SYSTEM_ENTER_CRITICAL and _LEAVE_CRITICAL. This should work: #define LORA_SYSTEM_ENTER_CRITICAL(APP) volatile uint32_t primask = __get_PRIMASK();__disable_irq(); #define LORA_SYSTEM_ENTER_CRITICAL(APP) __set_PRIMASK(primask); I recommend to put this in a header file, then define LORA_TARGET_INCLUDE to be the name of that file (e.g. -DLORA_TARGET_INCLUDE='"your_header.h"'). All the other LDL build options can go there. You will need to also define: LORA_ENABLE_SX1276 LORA_ENABLE_US_902_928 LORA_DISABLE_FULL_CODEC I assume you've already done this somewhere I can't see. I mean, you can do it all from the makefile if you prefer. Once you get that sorted you will probably find that the example sends too frequently since the US doesn't have a duty cycle limit. To slow it down you can either add a delay in your app, or set LDL_MAC_setAggregatedDutyCycle() to impose a global duty cycle limit. A setting of 2 will give you a 1% duty cycle limit. Hi, Thanks for taking a look at the project, just noticed i uploaded the keys of it, so i had to make the repo private :/, let me know if you want me to give you access to it. I see the problem, you need to connect the radio "DIO" control lines. That part is missing from the example you are based on, it's not very clear. You need to detect the DIO line(s) rising edge and then call LDL_MAC_interrupt(&mac, n, LDL_System_time()) where n is the index of the line (e.g. DIO0 is n == 0). You only need DIO0, DIO1, DIO2, and DIO3. This is how the arduino wrapper does it. That function is called by an ISR for a particular control line. There is implementation specific logic but the important part is that I call LDL_MAC_interrupt(). Here's how i implemented it: void HAL_GPIO_EXTI_Callback(uint16_t GPIO_Pin) { switch (GPIO_Pin) { case DIO0_Pin: LDL_MAC_interrupt(&mac, 0, LDL_System_time()); break; case DIO1_Pin: LDL_MAC_interrupt(&mac, 1, LDL_System_time()); break; case DIO2_Pin: LDL_MAC_interrupt(&mac, 2, LDL_System_time()); break; case DIO3_Pin: LDL_MAC_interrupt(&mac, 3, LDL_System_time());; break; default: break; } } If you use an interrupt, make sure to define LORA_SYSTEM_ENTER_CRITICAL and _LEAVE_CRITICAL. This should work: #define LORA_SYSTEM_ENTER_CRITICAL(APP) volatile uint32_t primask = __get_PRIMASK();__disable_irq(); #define LORA_SYSTEM_LEAVE_CRITICAL(APP) __set_PRIMASK(primask); I recommend to put this in a header file, then define LORA_TARGET_INCLUDE to be the name of that file (e.g. -DLORA_TARGET_INCLUDE='"your_header.h"'). All the other LDL build options can go there. My custom header file is named LDL_options.h and goes as follows: #ifndef LDL_OPTIONS_H_ #define LDL_OPTIONS_H_ #include "cmsis_gcc.h" // http://stm32f4-discovery.net/2015/06/how-to-properly-enabledisable-interrupts-in-arm-cortex-m/ static volatile uint32_t primask = 0; #define LORA_SYSTEM_ENTER_CRITICAL(APP) do { primask = __get_PRIMASK(); __disable_irq(); } while (0); #define LORA_SYSTEM_LEAVE_CRITICAL(APP) __set_PRIMASK(primask); #endif /* LDL_OPTIONS_H_ */ You will need to also define: LORA_ENABLE_SX1276 LORA_ENABLE_US_902_928 LORA_DISABLE_FULL_CODEC I assume you've already done this somewhere I can't see. I mean, you can do it all from the makefile if you prefer. Yep, I define those symbols on the IDE, but now i think it's better to have them on the LDL_options.h file so others doesn't need to load the project into the IDE just to see the configuration. So the LDL_options.h file ends up like this: #ifndef LDL_OPTIONS_H_ #define LDL_OPTIONS_H_ #include "cmsis_gcc.h" // http://stm32f4-discovery.net/2015/06/how-to-properly-enabledisable-interrupts-in-arm-cortex-m/ #define LORA_ENABLE_SX1276 #define LORA_ENABLE_US_902_928 #define LORA_DISABLE_FULL_CODEC // #define LORA_TARGET_INCLUDE /* See lora_platform */ static volatile uint32_t primask = 0; #define LORA_SYSTEM_ENTER_CRITICAL(APP) do { primask = __get_PRIMASK(); __disable_irq(); } while (0); #define LORA_SYSTEM_LEAVE_CRITICAL(APP) __set_PRIMASK(primask); #endif /* LDL_OPTIONS_H_ */ Once you get that sorted you will probably find that the example sends too frequently since the US doesn't have a duty cycle limit. To slow it down you can either add a delay in your app, or set LDL_MAC_setAggregatedDutyCycle() to impose a global duty cycle limit. A setting of 7 will give you a ~1% duty cycle limit. edit: made mistake on global duty cycle Will edit the comment once i get some results later today. Thanks for the help and patience. 10km is probably too far for initial debug. Too close (i.e. sitting right next to the gateway) can also be an issue. It's often useful to print time (i.e. LDL_System_time()) with each event for double checking timing. Hi, Thanks for the tips. I had to modify the logging macros (replacing PRIu32 for %d) because of my underlying functions. Everything else seems to be working as expected. Will report back when I get a gateway.
2025-04-01T06:38:12.183351
2016-06-24T14:24:09
162158572
{ "authors": [ "mafrosis", "thicklord" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4700", "repo": "cjlucas/rtorrent-python", "url": "https://github.com/cjlucas/rtorrent-python/issues/29" }
gharchive/issue
Docs for 1.0-alpha? Are these available somewhere convenient? Are these available anywhere at all ?
2025-04-01T06:38:12.278210
2021-07-30T16:00:23
956883400
{ "authors": [ "alex60217101990", "claffin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4701", "repo": "claffin/cloudproxy", "url": "https://github.com/claffin/cloudproxy/issues/49" }
gharchive/issue
always destroy instances Expected Behavior an GCP and DO cloud providers always destroy instances: 2021-07-30 15:59:33.646 | INFO | uvicorn.protocols.http.h11_impl:send:461 - <IP_ADDRESS>:60922 - "GET /destroy HTTP/1.1" 200 2021-07-30 15:59:36.640 | INFO | uvicorn.protocols.http.h11_impl:send:461 - <IP_ADDRESS>:60922 - "GET /destroy HTTP/1.1" 200 2021-07-30 15:59:30.937 | INFO | uvicorn.protocols.http.h11_impl:send:461 - <IP_ADDRESS>:60924 - "GET /destroy HTTP/1.1" 200 The UI regularly calls the /destroy endpoint to get a list of all the proxy instances pending deletion. This get request you are seeing is just that and it's not actually deleting the proxies. If you start Cloudproxy and don't open the UI, then you'll notice those requests aren't there.
2025-04-01T06:38:12.280171
2023-09-21T13:01:08
1906923693
{ "authors": [ "keighrim" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4702", "repo": "clamsproject/app-swt-detection", "url": "https://github.com/clamsproject/app-swt-detection/issues/2" }
gharchive/issue
image preprocessing for training For the first batches of training, we plan to use video frames extracted from AAPB videos. However, there's concerns around how to "normalize" different image size and aspect ratios from videos from different decades. This thread is to discuss how we implement the normalizing strategies. We decided to go only with 4:3 videos (before HD broadcasting era, circa early 2000s) in the early rounds of annotation. Note that as long as we are using the pre-trained backbone weights, those weights in torch-vision package come with their own preprocessing code. That is, we can add some additional preprocessing based on some domain knowledge before the torch-shipped preprocessing. However, as of now we don't see a lot of needs for doing so.
2025-04-01T06:38:12.296843
2024-03-16T16:10:54
2190117065
{ "authors": [ "HighCommander4", "jcarrete5" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4703", "repo": "clangd/clangd", "url": "https://github.com/clangd/clangd/issues/1975" }
gharchive/issue
clangd "--query-driver" option doesn't follow symlinks When I run clangd with the --query-driver option set to a path containing a symlink, the argument has no effect and I still have missing system headers. In the following case, <iostream> can't be found. #include <iostream> int main(int argc, char* argv[]) { std::cout << "hi\n"; return 0; } I can work around this issue by using the complete path without any symlinks. Within the logs, the oe-workdir path item is a symlink to another location. Logs Please attach the clangd stderr log if you can. (Usually available from the editor) If possible, run with --log=verbose - note that the logs will include the contents of open files! If this is a crash, try to put llvm-symbolizer on your PATH per the troubleshooting instructions. [START][2024-03-16 12:09:47] LSP logging initiated [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'I[12:09:47.584] clangd version 17.0.6\nI[12:09:47.584] Features: linux\nI[12:09:47.584] PID: 102869\nI[12:09:47.584] Working directory: /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight\nI[12:09:47.584] argv[0]: /usr/bin/clangd\nI[12:09:47.584] argv[1]: --log=verbose\nI[12:09:47.584] argv[2]: --query-driver=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/oe-workdir/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++\nV[12:09:47.584] User config file is /home/jasonc/.config/clangd/config.yaml\nI[12:09:47.584] Starting LSP over stdin/stdout\nV[12:09:47.584] <<< {"id":1,"jsonrpc":"2.0","method":"initialize","params":{"capabilities":{"offsetEncoding":["utf-8","utf-16"],"textDocument":{"callHierarchy":{"dynamicRegistration":false},"codeAction":{"codeActionLiteralSupport":{"codeActionKind":{"valueSet":["","quickfix","refactor","refactor.extract","refactor.inline","refactor.rewrite","source","source.organizeImports"]}},"dataSupport":true,"dynamicRegistration":false,"isPreferredSupport":true,"resolveSupport":{"properties":["edit"]}},"completion":{"completionItem":{"commitCharactersSupport":true,"deprecatedSupport":true,"documentationFormat":["markdown","plaintext"],"insertReplaceSupport":true,"insertTextModeSupport":{"valueSet":[1,2]},"labelDetailsSupport":true,"preselectSupport":true,"resolveSupport":{"properties":["documentation","detail","additionalTextEdits","sortText","filterText","insertText","textEdit","insertTextFormat","insertTextMode"]},"snippetSupport":true,"tagSupport":{"valueSet":[1]}},"completionItemKind":{"valueSet":[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25]},"completionList":{"itemDefaults":["commitCharacters","editRange","insertTextFormat","insertTextMode","data"]},"contextSupport":true,"dynamicRegistration":false,"editsNearCursor":true,"insertTextMode":1},"declaration":{"linkSupport":true},"definition":{"linkSupport":true},"documentHighlight":{"dynamicRegistration":false},"documentSymbol":{"dynamicRegistration":false,"hierarchicalDocumentSymbolSupport":true,"symbolKind":{"valueSet":[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]}},"hover":{"contentFormat":["markdown","plaintext"],"dynamicRegistration":false},"implementation":{"linkSupport":true},"publishDiagnostics":{"relatedInformation":true,"tagSupport":{"valueSet":[1,2]}},"references":{"dynamicRegistration":false},"rename":{"dynamicRegistration":false,"prepareSupport":true},"semanticTokens":{"augmentsSyntaxTokens":true,"dynamicRegistration":false,"formats":["relative"],"multilineTokenSupport":false,"overlappingTokenSupport":true,"requests":{"full":{"delta":true},"range":false},"serverCancelSupport":false,"tokenModifiers":["declaration","definition","readonly","static","deprecated","abstract","async","modification","documentation","defaultLibrary"],"tokenTypes":["namespace","type","class","enum","interface","struct","typeParameter","parameter","variable","property","enumMember","event","function","method","macro","keyword","modifier","comment","string","number","regexp","operator","decorator"]},"signatureHelp":{"dynamicRegistration":false,"signatureInformation":{"activeParameterSupport":true,"documentationFormat":["markdown","plaintext"],"parameterInformation":{"labelOffsetSupport":true}}},"synchronization":{"didSave":true,"dynamicRegistration":false,"willSave":true,"willSaveWaitUntil":true},"typeDefinition":{"linkSupport":true}},"window":{"showDocument":{"support":true},"showMessage":{"messageActionItem":{"additionalPropertiesSupport":false}},"workDoneProgress":true},"workspace":{"applyEdit":true,"configuration":true,"didChangeWatchedFiles":{"dynamicRegistration":false,"relativePatternSupport":true},"semanticTokens":{"refreshSupport":true},"symbol":{"dynamicRegistration":false,"hierarchicalWorkspaceSymbolSupport":true,"symbolKind":{"valueSet":[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]}},"workspaceEdit":{"resourceOperations":["rename","create","delete"]},"workspaceFolders":true}},"clientInfo":{"name":"Neovim","version":"0.9.5"},"initializationOptions":{},"processId":102865,"rootPath":"/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight","rootUri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight","trace":"off","workspaceFolders":[{"name":"/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight","uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight"}]}}\n\nI[12:09:47.584] <-- initialize(1)\nI[12:09:47.585] --> reply:initialize(1) 0 ms\nV[12:09:47.585] >>> {"id":1,"jsonrpc":"2.0","result":{"capabilities":{"astProvider":true,"callHierarchyProvider":true,"clangdInlayHintsProvider":true,"codeActionProvider":{"codeActionKinds":["quickfix","refactor","info"]},"compilationDatabase":{"automaticReload":true},"completionProvider":{"resolveProvider":false,"triggerCharacters":[".","<",">",":","\\"","/","*"]},"declarationProvider":true,"definitionProvider":true,"documentFormattingProvider":true,"documentHighlightProvider":true,"documentLinkProvider":{"resolveProvider":false},"documentOnTypeFormattingProvider":{"firstTriggerCharacter":"\\n","moreTriggerCharacter":[]},"documentRangeFormattingProvider":true,"documentSymbolProvider":true,"executeCommandProvider":{"commands":["clangd.applyFix","clangd.applyTweak"]},"foldingRangeProvider":true,"hoverProvider":true,"implementationProvider":true,"inactiveRegionsProvider":true,"inlayHintProvider":true,"memoryUsageProvider":true,"referencesProvider":true,"renameProvider":{"prepareProvider":true},"selectionRangeProvider":true,"semanticTokensProvider":{"full":{"delta":true},"legend":{"tokenModifiers":["declaration","definition","deprecated","deduced","readonly","static","abstract","virtual","dependentName","defaultLibrary","usedAsMutableReference","usedAsMutablePointer","constructorOrDestructor","userDefined","functionScope","classScope","fileScope","globalScope"],"tokenTypes":["variable","variable","parameter","function","method","function","property","variable","class","interface","enum","enumMember","type","type","unknown","namespace","typeParameter","concept","type","macro","modifier","operator","bracket","label","comment"]},"range":false},"signatureHelpProvider":{"triggerCharacters":["(",")","{","}","<",">",","]},"standardTypeHierarchyProvider":true,"textDocumentSync":{"change":2,"openClose":true,"save":true},"typeDefinitionProvider":true,"typeHierarchyProvider":true,"workspaceSymbolProvider":true},"offsetEncoding":"utf-8","serverInfo":{"name":"clangd","version":"clangd version 17.0.6 linux x86_64-pc-linux-gnu"}}}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.699] <<< {\"jsonrpc\":\"2.0\",\"method\":\"initialized\",\"params\":{}}\n\nI[12:09:47.699] <-- initialized\nV[12:09:47.699] <<< {\"jsonrpc\":\"2.0\",\"method\":\"textDocument/didOpen\",\"params\":{\"textDocument\":{\"languageId\":\"cpp\",\"text\":\"/*\\n * SPDX-License-Identifier: GPL-3.0-or-later\\n *\\n * Copyright (C) 2023 Jason Carrete\\n *\\n * This file is part of Flight Controller.\\n *\\n * Flight Controller is free software: you can redistribute it and/or modify\\n * it under the terms of the GNU General Public License as published by\\n * the Free Software Foundation, either version 3 of the License, or\\n * (at your option) any later version.\\n *\\n * This program is distributed in the hope that it will be useful,\\n * but WITHOUT ANY WARRANTY; without even the implied warranty of\\n * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\\n * GNU General Public License for more details.\\n *\\n * You should have received a copy of the GNU General Public License\\n * along with this program. If not, see <https://www.gnu.org/licenses/>.\\n */\\n\\n#include \\\"appinfo.h\\\"\\n#include \\\"version.h\\\"\\n\\n#include <iostream>\\n\\nnamespace ffd = freeflight_daemon;\\n\\nint main(int argc, char* argv[])\\n{\\n std::cout << ffd::get_app_name() << ' ' << freeflight::get_version() << ' '\\n << freeflight::get_name() << '\\\\n';\\n return 0;\\n}\\n\",\"uri\":\"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\",\"version\":0}}}\n\nI[12:09:47.699] <-- textDocument/didOpen\nV[12:09:47.699] System include extraction: driver clang expanded to /usr/bin/clang\nV[12:09:47.699] System include extraction: not allowed driver /usr/bin/clang\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.700] <<< {"id":2,"jsonrpc":"2.0","method":"textDocument/semanticTokens/full","params":{"textDocument":{"uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp"}}}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.700] <-- textDocument/semanticTokens/full(2)\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.700] config note at /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd:1:0: Parsing config fragment\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.700] config note at /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd:1:0: Parsed 1 fragments from file\nV[12:09:47.700] Config fragment: compiling /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd:1 -> 0x0000724F50003730 (trusted=false)\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.701] --> textDocument/publishDiagnostics\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.701] >>> {"jsonrpc":"2.0","method":"textDocument/publishDiagnostics","params":{"diagnostics":[],"uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd"}}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.701] Loaded compilation database from /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/compile_commands.json\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.701] Broadcasting compilation database from /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.701] System include extraction: not allowed driver /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.701] ASTWorker building file /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0 with command \n[/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git]\n/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++ --target=aarch64-poky-linux --driver-mode=g++ --sysroot=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot -I/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd -I/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/freeflight/public -mcpu=cortex-a57 -march=armv8-a+crc -mbranch-protection=standard -fstack-protector-strong -O2 -D_FORTIFY_SOURCE=2 -Wformat -Wformat-security -Werror=format-security --sysroot=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot -O2 -pipe -g -feliminate-unused-debug-types -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native= -fvisibility-inlines-hidden -O2 -g -DNDEBUG -std=gnu++20 -o src/flightd/CMakeFiles/flightd.dir/main.cpp.o -c -resource-dir=/usr/lib/clang/17 -- /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.702] Loaded compilation database from /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git/compile_commands.json\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.702] --> window/workDoneProgress/create(0)\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.702] >>> {"id":0,"jsonrpc":"2.0","method":"window/workDoneProgress/create","params":{"token":"backgroundIndexProgress"}}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.702] Enqueueing 1 commands for indexing\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.702] Driver produced command: cc1 -cc1 -triple aarch64-poky-linux -fsyntax-only -disable-free -clear-ast-before-backend -disable-llvm-verifier -discard-value-names -main-file-name main.cpp -mrelocation-model pic -pic-level 2 -pic-is-pie -mframe-pointer=non-leaf -fmath-errno -ffp-contract=on -fno-rounding-math -mconstructor-aliases -funwind-tables=2 -target-cpu cortex-a57 -target-feature +neon -target-feature +v8a -target-feature +crc -target-abi aapcs -msign-return-address=non-leaf -msign-return-address-key=a_key -mbranch-target-enforce -debug-info-kind=constructor -dwarf-version=5 -debugger-tuning=gdb -fcoverage-compilation-dir=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git -resource-dir /usr/lib/clang/17 -I /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd -I /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/freeflight/public -D _FORTIFY_SOURCE=2 -D NDEBUG -isysroot /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot -internal-isystem /usr/lib/clang/17/include -internal-isystem /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot/usr/local/include -internal-externc-isystem /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot/include -internal-externc-isystem /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot/usr/include -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -O2 -Wformat -Wformat-security -Werror=format-security -std=gnu++20 -fdeprecated-macro -fdebug-compilation-dir=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native= -ferror-limit 19 -fvisibility-inlines-hidden -stack-protector 2 -fno-signed-char -fgnuc-version=4.2.1 -fno-implicit-modules -fcxx-exceptions -fexceptions -vectorize-loops -vectorize-slp -no-round-trip-args -target-feature -fmv -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -x c++ /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.702] Building first preamble for /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0\nV[12:09:47.702] BackgroundIndex: building version 1 after loading index from disk\nV[12:09:47.702] <<< {"id":0,"jsonrpc":"2.0","result":null}\n\nI[12:09:47.702] <-- reply(0)\nI[12:09:47.702] --> $/progress\nV[12:09:47.702] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"begin","percentage":0,"title":"indexing"}}}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.702] BackgroundIndex: serving version 1 (50980 bytes)\nI[12:09:47.702] --> $/progress\nV[12:09:47.703] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"report","message":"0/1","percentage":0}}}\n\nI[12:09:47.703] --> $/progress\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.703] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"report","message":"0/1","percentage":0}}}\n\nI[12:09:47.703] --> $/progress\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.703] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"end"}}}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.716] Built preamble of size 529320 for file /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0 in 0.01 seconds\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'I[12:09:47.717] --> workspace/semanticTokens/refresh(1)\nV[12:09:47.717] >>> {"id":1,"jsonrpc":"2.0","method":"workspace/semanticTokens/refresh","params":null}\n\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.717] <<< {"jsonrpc":"2.0","method":"$/cancelRequest","params":{"id":2}}\n\nI[12:09:47.717] <-- $/cancelRequest\nV[12:09:47.717] indexed preamble AST for /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0:\n symbol slab: 5 symbols, 5376 bytes\n ref slab: 0 symbols, 0 refs, 128 bytes\n relations slab: 0 relations, 24 bytes\nV[12:09:47.717] <<< {"id":3,"jsonrpc":"2.0","method":"textDocument/semanticTokens/full","params":{"textDocument":{"uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp"}}}\n\nI[12:09:47.717] <-- textDocument/semanticTokens/full(3)\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.717] <<< {"id":1,"jsonrpc":"2.0","result":null}\n\nI[12:09:47.717] <-- reply(1)\nV[12:09:47.717] Build dynamic index for header symbols with estimated memory usage of 22004 bytes\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.724] Trying to fix unresolved name "cout" in scopes: [std::]\n' [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.724] Dex query tree: false\nV[12:09:47.724] Dex query tree: false\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "E[12:09:47.725] IncludeCleaner: Failed to get an entry for resolved path : No such file or directory\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.726] indexed file AST for /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0:\n symbol slab: 2 symbols, 4680 bytes\n ref slab: 2 symbols, 2 refs, 4272 bytes\n relations slab: 0 relations, 24 bytes\nV[12:09:47.726] Build dynamic index for main-file symbols with estimated memory usage of 12040 bytes\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.726] --> textDocument/publishDiagnostics\n" [ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.726] >>> {\"jsonrpc\":\"2.0\",\"method\":\"textDocument/publishDiagnostics\",\"params\":{\"diagnostics\":[{\"code\":\"pp_file_not_found\",\"message\":\"'iostream' file not found\",\"range\":{\"end\":{\"character\":19,\"line\":24},\"start\":{\"character\":9,\"line\":24}},\"relatedInformation\":[],\"severity\":1,\"source\":\"clang\"},{\"code\":\"undeclared_var_use\",\"message\":\"Use of undeclared identifier 'std'\",\"range\":{\"end\":{\"character\":7,\"line\":30},\"start\":{\"character\":4,\"line\":30}},\"relatedInformation\":[],\"severity\":1,\"source\":\"clang\"},{\"code\":\"misc-unused-alias-decls\",\"codeDescription\":{\"href\":\"https://clang.llvm.org/extra/clang-tidy/checks/misc/unused-alias-decls.html\"},\"message\":\"Namespace alias decl 'ffd' is unused (fix available)\",\"range\":{\"end\":{\"character\":0,\"line\":27},\"start\":{\"character\":0,\"line\":26}},\"relatedInformation\":[],\"severity\":2,\"source\":\"clang-tidy\",\"tags\":[1]}],\"uri\":\"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\",\"version\":0}}\n\nI[12:09:47.726] --> reply:textDocument/semanticTokens/full(2) 26 ms, error: Task was cancelled.\nV[12:09:47.726] >>> {\"error\":{\"code\":-32800,\"message\":\"Request cancelled\"},\"id\":2,\"jsonrpc\":\"2.0\"}\n\nV[12:09:47.726] ASTWorker running SemanticHighlights on version 0 of /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\nI[12:09:47.726] --> reply:textDocument/semanticTokens/full(3) 8 ms\nV[12:09:47.726] >>> {\"id\":3,\"jsonrpc\":\"2.0\",\"result\":{\"data\":[26,10,3,15,65537,0,6,17,15,131072,2,4,4,3,131075,0,9,4,2,16387,0,12,4,2,16387],\"resultId\":\"1\"}}\n\n" System information Output of clangd --version: clangd version 17.0.6 Features: linux Platform: x86_64-pc-linux-gnu Editor/LSP plugin: Neovim/nvim-lspconfig Operating system: Arch Linux Kernel version: 6.7.9 Duplicate of #1605
2025-04-01T06:38:12.304715
2020-06-16T11:26:52
639588308
{ "authors": [ "HighCommander4", "kadircet", "lh123", "njames93" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4704", "repo": "clangd/clangd", "url": "https://github.com/clangd/clangd/issues/433" }
gharchive/issue
Include completion fails at the end of file clangd version: clangd version 11.0.0 (https://github.com/llvm/llvm-project.git 3badd17b6989621b5aa2732800f697dabbda034d) OS: Windows 10 I think this issue still exists. @HighCommander4 Originally posted by @lh123 in https://github.com/clangd/clangd/issues/38#issuecomment-643115928 You're right, I see the same issue (global completion instead of include completions). I think what happened when I was testing #38, is that I typed in #include "llvm/Sup manually. When I type the " character, my editor auto-inserts a matching closing quote, so the actual test case I was testing was: #include "llvm/Sup^" (not the presence of the closing quote), which works fine. I can't seem to reproduce this, neither on LLVM or a dummy project. Could you please share clangd logs ? It reproduces if the file has no code in and the include brackets are unbalanced and followed by eof It reproduces if the file has no code in and the include brackets are unbalanced and followed by eof Interesting I thought I had fixed this one... Well good thing is, this means at least I know the fix :D Sent out https://reviews.llvm.org/D95419 Interesting I thought I had fixed this one... Well good thing is, this means at least I know the fix :D Sent out https://reviews.llvm.org/D95419 Works like a charm now on my end, thanks. Works like a charm now on my end, thanks.
2025-04-01T06:38:12.317299
2020-03-23T01:36:40
585869355
{ "authors": [ "chmanie", "fannheyward", "sam-mccall", "stokhos" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4705", "repo": "clangd/coc-clangd", "url": "https://github.com/clangd/coc-clangd/issues/25" }
gharchive/issue
Clangd formatting issue? Hi: I notice that clang has formatting feature, but I don't know how to set it up. Is there any docs for this? best regards Peiyun Jin clangd is using clang-format to do formatting, https://clangd.llvm.org/features.html#formatting , you can set .clang-format for style options. Hi I have installed coc-clangd, have .clang-format file in the project folder, and set "coc.preferences.formatOnSaveFiletypes": ["cpp"], "coc.preferences.formatOnSave": true, But neither :Format nor format on save worked. Could you tell me how can fix this? Best I think the problem is caused by this line "clangd.arguments": ["-Wall", "-Werror", "-std=c++17"] in above in my coc-settings.json. After adding this line into my coc-settings.json. The language server stops working for some reason. { "coc.preferences.formatOnSaveFiletypes": ["cpp"], "coc.preferences.formatOnSave": true, "clangd.arguments": ["-Wall", "-Werror", "-std=c++17"], "languageserver": { "python": { "command": "python", "args": [ "-mpyls", "-vv", "--log-file", "/tmp/lsp_python.log" ], "trace.server": "verbose", "filetypes": [ "python" ], "settings": { "pyls": { "enable": true, "trace": { "server": "verbose" }, "commandPath": "", "configurationSources": [ "pycodestyle" ], "plugins": { "jedi_completion": { "enabled": true }, "jedi_hover": { "enabled": true }, "jedi_references": { "enabled": true }, "jedi_signature_help": { "enabled": true }, "jedi_symbols": { "enabled": true, "all_scopes": true }, "mccabe": { "enabled": true, "threshold": 15 }, "preload": { "enabled": true }, "pycodestyle": { "enabled": true }, "pydocstyle": { "enabled": false, "match": "(?!test_).*\\.py", "matchDir": "[^\\.].*" }, "pyflakes": { "enabled": true }, "rope_completion": { "enabled": true }, "yapf": { "enabled": true } } } } } } }``` Look like it's caused by server, @sam-mccall can you look into this? I think the problem is caused by this line "clangd.arguments": ["-Wall", "-Werror", "-std=c++17"] in above in my coc-settings.json. After adding this line into my coc-settings.json. The language server stops working for some reason. The reason is clangd.arguments is extra flags to pass to clangd, and those are not valid clangd arguments: $ clangd -Wall -Werror -std=c++17 clangd: Unknown command line argument '-Wall'. Try: 'clangd --help' clangd: Did you mean '--help'? clangd: Unknown command line argument '-Werror'. Try: 'clangd --help' clangd: Did you mean '--color'? clangd: Unknown command line argument '-std=c++17'. Try: 'clangd --help' clangd: Did you mean '--log=c++17'? If you want to set the flags for parsing your code, this is configured using compile_commands.json or compile_flags.txt: https://clangd.llvm.org/installation.html#project-setup @sam-mccall Hi Sam, Is there a way to enable clang-tidy and -j inside coc-settings.json? Or how can I enable them with vim? best regards Sure: "clangd.arguments": ["-j=3", "-clang-tidy-checks=bugprone-*"] - those really are clangd flags. Similar to clang-format, for finer-grained clang-tidy config you should use the standard .clang-tidy config file in your source tree, clangd should respect it. @sam-mccall Do I need to enable clang-tide first, "clangd.arguments": ["-j=3", "-clang-tide=true","-clang-tidy-checks=bugprone-*"]? What's clangd --version? It's on by default as of clangd 9. It may barely work in clangd 8, it was experiencing and off by default. I'm using fedora 32 beta. Clangd is version 10. With clangd 10 you shouldn't have to explicitly enable clang tidy. On Sat, Mar 28, 2020, 7:30 PM Stokhos<EMAIL_ADDRESS>wrote: I'm using fedora 32 beta. Clangd is version 10. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/clangd/coc-clangd/issues/25#issuecomment-605500662, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAEGBAMIBI3LZ3LECRB7SNTRJY65TANCNFSM4LRQVAMQ . Sorry to add to this but it seems related. I am very confused about what this plugin can do. Should I be able to do formatting? Issuing a coc :Format command doesn't do anything for me. Would then clang-format be used or clang-tidy with the --fix option (or both)?
2025-04-01T06:38:12.321682
2019-01-31T08:06:54
405125196
{ "authors": [ "clangen", "ibrokemypie" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4706", "repo": "clangen/musikcube", "url": "https://github.com/clangen/musikcube/issues/240" }
gharchive/issue
Ogg Opus Files Dont Show On macOS Mojave, installed via homebrew. MP3 files show fine, but the majority of my library which is in ogg opus just doesnt show in the library viewer. Oh, I see, tag parsing failed? That's possible, I don't have tons of Opus files. Is there any way you can share a file with tags that don't seem to be working? sure, try this one 01 Tree Village.ogg.zip Ah! So... taglib (our tag parsing library) sees the .ogg extension and assumes it's a vorbis file, then fails when it's not. When this happens, we should probably detect the error and then check to see if it's an opus file. Are there any other file formats that commonly have an .ogg extension that may not be vorbis? :thinking: Got it. I modified the taglib parser reader plugin so it can detect opus files in an ogg container -- your example file seems to parse fine now! I also made it such that as soon as I can get examples of other formats in ogg containers it should be trivial to add support to them as well. seems to be fixed after re-adding the library on latest head, thanks! Thanks for confirming the fix!
2025-04-01T06:38:12.346644
2021-05-21T09:52:44
897884262
{ "authors": [ "joaopaulovieira", "leticiafernandes", "pedrochamberlain", "unique1o1" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4707", "repo": "clappr/clappr", "url": "https://github.com/clappr/clappr/issues/2013" }
gharchive/issue
Browser starts to lag after playing 3-5 videos when using player.configure to update the source url Browser: Firefox 88.0.1 OS: Pop OS 20.04 LTS Clappr Version: latest from http://cdn.clappr.io/latest/clappr.min.js Steps to reproduce: play 3-5 video one after another. I was expecting to have a smooth playback but instead it started lagging(my browser) Did you try to reproduce this issue at http://cdn.clappr.io/ No I'm trying to update the video source using the following code: // [...] player.configure({ source: 'another-url', // ... }); which updates the video but after playing around 3-5 videos one after another the browser starts to lag. I also tried to reproduce this issue on multiple devices, the same issue occurred. NOTE : source URL is an HLS video encrypted with AES-128 Did you try to reproduce this issue at http://cdn.clappr.io/ No Can't you reproduce the reported behavior or just haven't tried? I also tried to reproduce this issue on multiple devices, the same issue occurred. Yes I have but not on http://cdn.clappr.io/ Yes, I understand that you have tested it on other devices. The recommendation to test on http://cdn.clappr.io/ (specifically on http://cdn.clappr.io/demo) is that it becomes a common point that we can also validate on our side. I was unable to reproduce this issue. Can you generate any evidence of this? (CPU / Memory consumption of the tab that is running Clappr, some video showing the problem visually) Hi @unique1o1, is the problem still happening? Can you generate any evidence of this? (CPU / Memory consumption of the tab that is running Clappr, some video showing the problem visually) I'm closing this issue due to inactivity. If needed, please feel free to reopen it.
2025-04-01T06:38:12.348139
2015-10-25T03:38:44
113204453
{ "authors": [ "leandromoreira", "polaris-zx" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:4708", "repo": "clappr/clappr", "url": "https://github.com/clappr/clappr/issues/609" }
gharchive/issue
no video on galaxy s5 There is no video on Galaxy s5 when a live video or a replay is playing. But the video is diaplayed after you click the progress bar. only the audio. are you still facing this issue @polaris-zx ? also, did you test with different stream sources? is it only with hls ?