added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:38:10.422953
| 2021-12-22T01:27:11
|
1086362235
|
{
"authors": [
"Bento007"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4593",
"repo": "chanzuckerberg/single-cell-data-portal",
"url": "https://github.com/chanzuckerberg/single-cell-data-portal/issues/1786"
}
|
gharchive/issue
|
bug(duplicate detection): Etags cannot be used to compare objects
Describe the bug
Duplicate detection does not detect duplicated when publishing a collection.
Current we use AWS S3 Etags to determine if two object in are identical. However since we are using KMS to encrypt our object in place, AWS does not guarantee that Etags will be identical for identical objects in S3.
Expected behavior
Duplicates should be caught before publishing a collection
Solution
Add a custom tag to objects upload to s3 which represents the hash of the upload object. This tag will be used compare objects uploaded to S3. Alternatively this tag can be store in the database rather than on the S3 object.
Whatever implementation we go with, we should generate tags for all previously uploaded dataset.
Unable to use the hash generated by dropbox because we are using a shared link to a file. https://www.dropboxforum.com/t5/Dropbox-API-Support-Feedback/API-to-get-content-hash-of-shared-file-link/td-p/458942
|
2025-04-01T06:38:10.430778
| 2020-08-07T01:30:50
|
674694039
|
{
"authors": [
"MDunitz",
"ambrosejcarr",
"brianraymor",
"signechambers1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4594",
"repo": "chanzuckerberg/single-cell",
"url": "https://github.com/chanzuckerberg/single-cell/issues/9"
}
|
gharchive/issue
|
Measure hosted cellxgene feature usage
For the hosted cellxgene application
Measure how many times key features are used per visit.
Measure the probability that two key features are used in sequence
Business Justification
This data will help us tailor UX research to understand if:
features are difficult to use, or
features don't provide value to users, or
features are used in particular flows or contexts, but not others.
List of "key features"
Left sidebar:
painting by a category on left side bar
selecting cells by unselecting some values within a category
opening a category on left side bar
creating a new category on left side bar
Top bar:
lassoing a selection of cells
saving a selection of cells to a DE category
saving a selection of cells to a category and assigning a label
calculating differential expression between two categories
subsetting to a selection of cells from a lasso selection
returning to the full dataset
selecting the zoom tool
show labels on graph
use the clip tool
undo an action
redo an action
use the help menu, or any sub-menu item.
Right sidebar:
Add a gene
Bulk add genes
create an xy plot
color the scatterplot by a quantitative variable (e.g. gene expression)
select a range of cells using the range selector
Have we talked to Trust about this? It seems like per visit and order of actions is getting pretty far into user tracking?
I raised with Trust on August 27 and pinged again this morning.
Per September 1 Refinement, @signechambers1 to follow up with @ambrosejcarr on the relative priority of this issue.
Trust approved the increased instrumentation on 9/15, from Shawn "Our current cellxgene policy covers this (see section 2.2) but I’ve made a note to spell this out even more since we are updating the policies for Data Portal anyways."
Per PM triage, closing. Ambrose will open a new issue to track specific actions.
|
2025-04-01T06:38:10.434521
| 2022-04-06T19:34:13
|
1195083033
|
{
"authors": [
"Dirrk",
"alldoami",
"mrlannigan",
"sfc-gh-swinkler"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4595",
"repo": "chanzuckerberg/terraform-provider-snowflake",
"url": "https://github.com/chanzuckerberg/terraform-provider-snowflake/pull/962"
}
|
gharchive/pull-request
|
feat: Add port/protocol for driver
Adds the ability to provide protocol and port options to the snowflake driver
Test Plan
[X] acceptance tests
[X] Locally we have tested this for almost two weeks
References
https://github.com/snowflakedb/gosnowflake/blob/master/dsn.go#L54
I rebased the changes in main and made sure the docs were updated properly this time, should be ready for testing
/ok-to-test sha=b6f792a
We fixed this test in a different PR, could you merge main into your branch and then we can merge your PR?
Hi @alldoami,
I have updated this PR with the latest from main.
/ok-to-test sha=0987fe9
/ok-to-test sha=0987fe9
#1238 closes this PR
|
2025-04-01T06:38:10.497542
| 2018-03-02T12:48:35
|
301755804
|
{
"authors": [
"GeorgLink",
"jsdabrowski"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4596",
"repo": "chaoss/governance",
"url": "https://github.com/chaoss/governance/pull/11"
}
|
gharchive/pull-request
|
Update GSoC-interest.md
provided pull request with my information
@jsdabrowski are you no longerer interested or did you close the pull request for a different reason?
am interested
Wonderful, you can make the requested changes in your repository and it will update this pull request :-)
For this pull request to update, you have to push the changes to the same branch that this pull request is from: patch-1
I hope should it be good. Sorry for the mess. i am still in the learning process.
No worries, I found GitHub and pull requests confusing at first.
You are 2/3 done now:
[ ] Please update your pull request to not delete other interested students that are already in the list.
[x] Please change the format of your email to use "at" instead of @
[x] Please change the last column to read "To-Do" without a link, if you don't have your Project Proposal ready yet.
hope in the project will be many opportunity to advance the skills
thank you
|
2025-04-01T06:38:10.585226
| 2016-03-17T23:25:01
|
141733466
|
{
"authors": [
"chapmanb",
"tanglingfung"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4603",
"repo": "chapmanb/bcbio-nextgen",
"url": "https://github.com/chapmanb/bcbio-nextgen/issues/1277"
}
|
gharchive/issue
|
KeyError: 'fasta'
Hi Brad,
After upgrading the pipeline, I have the following error when I start a run. Where should I be looking for the configuration file?
Paul
File "/media/bigdata/bcbio-nextgen/anaconda/lib/python2.7/site-packages/bcbio/pipeline/run_info.py", line 179, in _check_ref_files
for contig in ref.file_contigs(ref_info["fasta"]["base"], data["config"]):
KeyError: 'fasta'
Paul;
Sorry about the confusing error message. I was able to replicate this error if you don't add a genome_build to your configuration. I updated the configuration checking to identify this and provide a more useful message. Hopefully this explains the problem and once adding it you can get things running cleanly. Thanks again for the feedback.
Thanks Brad. I just found that my .yaml file is truncated where the
genome_build is taken off for the last sample. Thanks for digging into the
issue.
On Fri, Mar 18, 2016 at 2:50 AM, Brad Chapman<EMAIL_ADDRESS>wrote:
Paul;
Sorry about the confusing error message. I was able to replicate this
error if you don't add a genome_build to your configuration. I updated
the configuration checking to identify this and provide a more useful
message. Hopefully this explains the problem and once adding it you can get
things running cleanly. Thanks again for the feedback.
—
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub
https://github.com/chapmanb/bcbio-nextgen/issues/1277#issuecomment-198284377
|
2025-04-01T06:38:10.594269
| 2015-09-27T09:57:10
|
108525464
|
{
"authors": [
"charlesthk",
"guettli"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4604",
"repo": "charlesthk/python-resize-image",
"url": "https://github.com/charlesthk/python-resize-image/issues/7"
}
|
gharchive/issue
|
Why open image with 'rw' in docs
In the docs (https://github.com/charlesthk/python-resize-image) you use 'rw' to open the images.
from PIL import Image
test_img = open('test-image.jpeg', 'rw')
img = Image.open(test_img)
img = resizeimage.resize_thumbnail(img, [200, 200])
img.save('test-image-thumbnail.jpeg', img.format)
test_img.close()
I see no sense in opening the file for writing.
BTW, it is very common to use fd_.... as variable name for file descriptors returned by open().
You use test_img. That's valid python code, but fd_img would be more easy to read.
Thanks for this issue, I updated my code accordingly.
|
2025-04-01T06:38:10.764791
| 2017-02-17T05:56:08
|
208344171
|
{
"authors": [
"Conrad-S",
"Sicos1977",
"andimalaj",
"anpham293",
"charlesw",
"losbear"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4605",
"repo": "charlesw/tesseract",
"url": "https://github.com/charlesw/tesseract/issues/318"
}
|
gharchive/issue
|
Failed to initialise tesseract engine (Tesseract 3.0.2 - Visual Studio 2015)
Hi and thank you very much for your help.
I have done everything that I possibly can think of to attempt to resolve a "Failed to initialise engine" error.
I read post #277 where the person stated that adjusting the path to tessdata fixed his issue. But he did not state the "adjustment" that he made.
Line where the failure occurs
using (var engine = new TesseractEngine (@"./tessdata", "eng", EngineMode.Default)) {
Error
"Failed to initialise tesseract engine.. See https://github.com/charlesw/tesseract/wiki/Error-1 for details."} Tesseract.TesseractException
Stack Trace
Data {System.Collections.ListDictionaryInternal} System.Collections.IDictionary {System.Collections.ListDictionaryInternal}
HResult -2146233088 int
HelpLink null string
InnerException null System.Exception
Message "Failed to initialise tesseract engine.. See https://github.com/charlesw/tesseract/wiki/Error-1 for details." string
Source "Tesseract" string
StackTrace "
at Tesseract.TesseractEngine.Initialise(String datapath, String language, EngineMode engineMode, IEnumerable1 configFiles, IDictionary2 initialValues, Boolean setOnlyNonDebugVariables)\r\n
at Tesseract.TesseractEngine..ctor(String datapath, String language, EngineMode engineMode, IEnumerable1 configFiles, IDictionary2 initialOptions, Boolean setOnlyNonDebugVariables)\r\n
at Tesseract.TesseractEngine..ctor(String datapath, String language, EngineMode engineMode)\r\n
at OCRTestOne.OCRTest.Go() in C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\Classes\OCRTest.cs:line 22" string
TargetSite {Void Initialise(System.String, System.String, Tesseract.EngineMode, System.Collections.Generic.IEnumerable1[System.String], System.Collections.Generic.IDictionary2[System.String,System.Object], Boolean)} System.Reflection.MethodBase {System.Reflection.RuntimeMethodInfo}
Development Environment
Windows Server 2012 R2 (virtual machine running in Azure)
Visual Studio 2015 (Version 14.0.25431.01 Update 3)
Project Type: ASP.NET (also tried console app).
Framework Target 452 (.NET Framework 4.5.2).
Verified that Visual C++ 32-bit and 64-bit runtimes are both installed
Nuget Package: Tesseract by Charles Weld 3.0.2
tessdata folder exists with English language files underneath (I added it). Files set to Copy Always.
32 and 64-bit assemblies (liblept172.dll and libtesseract304.dll) exist in the x86 and x64 folders and are set to Copy Always (they seem to be copied to the lib folder).
Trace
Turned on verbose mode tracing. The trace below is based on one execution attempt (I tried to attach a text file but the site said: "We don't support that file type"). Apologies for the long trace
Tesseract Information: 0 : Current OS: Windows
Tesseract Information: 0 : Current platform: x86
Tesseract Information: 0 : Custom search path is not defined, skipping.
Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for 'liblept172.dll' on platform x86.
Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'liblept172.dll' on platform x86.
Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll"...
Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll", handle = 325451776.
Tesseract Information: 0 : Trying to load native function "pixaReadMultipageTiff" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixaReadMultipageTiff", function handle = 325469402.
Tesseract Information: 0 : Trying to load native function "pixaGetCount" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixaGetCount", function handle = 325462092.
Tesseract Information: 0 : Trying to load native function "pixaGetPix" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixaGetPix", function handle = 325455972.
Tesseract Information: 0 : Trying to load native function "pixaDestroy" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixaDestroy", function handle = 325456927.
Tesseract Information: 0 : Trying to load native function "pixCreate" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixCreate", function handle = 325465967.
Tesseract Information: 0 : Trying to load native function "pixClone" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixClone", function handle = 325466612.
Tesseract Information: 0 : Trying to load native function "pixDestroy" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixDestroy", function handle = 325466382.
Tesseract Information: 0 : Trying to load native function "pixGetWidth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetWidth", function handle = 325460622.
Tesseract Information: 0 : Trying to load native function "pixGetHeight" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetHeight", function handle = 325460452.
Tesseract Information: 0 : Trying to load native function "pixGetDepth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetDepth", function handle = 325464312.
Tesseract Information: 0 : Trying to load native function "pixGetXRes" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetXRes", function handle = 325463887.
Tesseract Information: 0 : Trying to load native function "pixGetYRes" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetYRes", function handle = 325463882.
Tesseract Information: 0 : Trying to load native function "pixGetResolution" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetResolution", function handle = 325469492.
Tesseract Information: 0 : Trying to load native function "pixGetWpl" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetWpl", function handle = 325466032.
Tesseract Information: 0 : Trying to load native function "pixSetXRes" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSetXRes", function handle = 325463907.
Tesseract Information: 0 : Trying to load native function "pixSetYRes" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSetYRes", function handle = 325463902.
Tesseract Information: 0 : Trying to load native function "pixSetResolution" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSetResolution", function handle = 325465697.
Tesseract Information: 0 : Trying to load native function "pixScaleResolution" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixScaleResolution", function handle = 325467282.
Tesseract Information: 0 : Trying to load native function "pixGetData" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetData", function handle = 325469052.
Tesseract Information: 0 : Trying to load native function "pixGetInputFormat" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetInputFormat", function handle = 325467712.
Tesseract Information: 0 : Trying to load native function "pixSetInputFormat" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSetInputFormat", function handle = 325459102.
Tesseract Information: 0 : Trying to load native function "pixEndianByteSwap" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixEndianByteSwap", function handle = 325468587.
Tesseract Information: 0 : Trying to load native function "pixRead" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixRead", function handle = 325467527.
Tesseract Information: 0 : Trying to load native function "pixReadMemTiff" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixReadMemTiff", function handle = 325472137.
Tesseract Information: 0 : Trying to load native function "pixWrite" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixWrite", function handle = 325466617.
Tesseract Information: 0 : Trying to load native function "pixGetColormap" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixGetColormap", function handle = 325464692.
Tesseract Information: 0 : Trying to load native function "pixSetColormap" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSetColormap", function handle = 325468467.
Tesseract Information: 0 : Trying to load native function "pixDestroyColormap" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixDestroyColormap", function handle = 325465782.
Tesseract Information: 0 : Trying to load native function "pixConvertRGBToGray" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixConvertRGBToGray", function handle = 325460777.
Tesseract Information: 0 : Trying to load native function "pixDeskewGeneral" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixDeskewGeneral", function handle = 325460177.
Tesseract Information: 0 : Trying to load native function "pixRotate" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixRotate", function handle = 325466312.
Tesseract Information: 0 : Trying to load native function "pixRotateOrth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixRotateOrth", function handle = 325468177.
Tesseract Information: 0 : Trying to load native function "pixOtsuAdaptiveThreshold" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixOtsuAdaptiveThreshold", function handle = 325468767.
Tesseract Information: 0 : Trying to load native function "pixSauvolaBinarize" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSauvolaBinarize", function handle = 325463922.
Tesseract Information: 0 : Trying to load native function "pixSauvolaBinarizeTiled" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixSauvolaBinarizeTiled", function handle = 325461412.
Tesseract Information: 0 : Trying to load native function "pixScale" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixScale", function handle = 325466597.
Tesseract Information: 0 : Trying to load native function "pixcmapCreate" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapCreate", function handle = 325462397.
Tesseract Information: 0 : Trying to load native function "pixcmapCreateRandom" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapCreateRandom", function handle = 325460482.
Tesseract Information: 0 : Trying to load native function "pixcmapCreateLinear" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapCreateLinear", function handle = 325460392.
Tesseract Information: 0 : Trying to load native function "pixcmapCopy" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapCopy", function handle = 325463212.
Tesseract Information: 0 : Trying to load native function "pixcmapDestroy" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapDestroy", function handle = 325464252.
Tesseract Information: 0 : Trying to load native function "pixcmapGetCount" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetCount", function handle = 325466562.
Tesseract Information: 0 : Trying to load native function "pixcmapGetFreeCount" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetFreeCount", function handle = 325456837.
Tesseract Information: 0 : Trying to load native function "pixcmapGetDepth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetDepth", function handle = 325468462.
Tesseract Information: 0 : Trying to load native function "pixcmapGetMinDepth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetMinDepth", function handle = 325462157.
Tesseract Information: 0 : Trying to load native function "pixcmapClear" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapClear", function handle = 325460977.
Tesseract Information: 0 : Trying to load native function "pixcmapAddColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapAddColor", function handle = 325466517.
Tesseract Information: 0 : Trying to load native function "pixcmapAddNewColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapAddNewColor", function handle = 325459832.
Tesseract Information: 0 : Trying to load native function "pixcmapAddNearestColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapAddNearestColor", function handle = 325464897.
Tesseract Information: 0 : Trying to load native function "pixcmapUsableColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapUsableColor", function handle = 325468852.
Tesseract Information: 0 : Trying to load native function "pixcmapAddBlackOrWhite" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapAddBlackOrWhite", function handle = 325460002.
Tesseract Information: 0 : Trying to load native function "pixcmapSetBlackAndWhite" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapSetBlackAndWhite", function handle = 325456282.
Tesseract Information: 0 : Trying to load native function "pixcmapGetColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetColor", function handle = 325466527.
Tesseract Information: 0 : Trying to load native function "pixcmapGetColor32" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetColor32", function handle = 325458992.
Tesseract Information: 0 : Trying to load native function "pixcmapResetColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapResetColor", function handle = 325457197.
Tesseract Information: 0 : Trying to load native function "pixcmapGetIndex" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetIndex", function handle = 325466567.
Tesseract Information: 0 : Trying to load native function "pixcmapHasColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapHasColor", function handle = 325458122.
Tesseract Information: 0 : Trying to load native function "pixcmapCountGrayColors" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapCountGrayColors", function handle = 325462272.
Tesseract Information: 0 : Trying to load native function "pixcmapCountGrayColors" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapCountGrayColors", function handle = 325462272.
Tesseract Information: 0 : Trying to load native function "pixcmapGetNearestIndex" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetNearestIndex", function handle = 325459557.
Tesseract Information: 0 : Trying to load native function "pixcmapGetNearestGrayIndex" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetNearestGrayIndex", function handle = 325466522.
Tesseract Information: 0 : Trying to load native function "pixcmapGetComponentRange" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetComponentRange", function handle = 325459077.
Tesseract Information: 0 : Trying to load native function "pixcmapGetExtremeValue" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGetExtremeValue", function handle = 325458877.
Tesseract Information: 0 : Trying to load native function "pixcmapGrayToColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGrayToColor", function handle = 325464607.
Tesseract Information: 0 : Trying to load native function "pixcmapColorToGray" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapColorToGray", function handle = 325469397.
Tesseract Information: 0 : Trying to load native function "pixcmapColorToGray" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapColorToGray", function handle = 325469397.
Tesseract Information: 0 : Trying to load native function "pixcmapToRGBTable" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapToRGBTable", function handle = 325469742.
Tesseract Information: 0 : Trying to load native function "pixcmapSerializeToMemory" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapSerializeToMemory", function handle = 325465572.
Tesseract Information: 0 : Trying to load native function "pixcmapDeserializeFromMemory" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapDeserializeFromMemory", function handle = 325460637.
Tesseract Information: 0 : Trying to load native function "pixcmapGammaTRC" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapGammaTRC", function handle = 325470952.
Tesseract Information: 0 : Trying to load native function "pixcmapContrastTRC" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapContrastTRC", function handle = 325462497.
Tesseract Information: 0 : Trying to load native function "pixcmapShiftIntensity" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixcmapShiftIntensity", function handle = 325460387.
Tesseract Information: 0 : Current platform: x86
Tesseract Information: 0 : Custom search path is not defined, skipping.
Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for 'libtesseract304.dll' on platform x86.
Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'libtesseract304.dll' on platform x86.
Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll"...
Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll", handle = 329383936.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIAnalyseLayout" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIAnalyseLayout", function handle = 329706928.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIClear" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIClear", function handle = 329706560.
Tesseract Information: 0 : Trying to load native function "TessBaseAPICreate" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPICreate", function handle = 329709952.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIDelete" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIDelete", function handle = 329709920.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIDetectOS" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIDetectOS", function handle = 329706368.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetBoolVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetBoolVariable", function handle = 329709264.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetDoubleVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetDoubleVariable", function handle = 329709184.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetHOCRText" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetHOCRText", function handle = 329706736.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetIntVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetIntVariable", function handle = 329709344.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetIterator" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetIterator", function handle = 329706784.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetPageSegMode" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetPageSegMode", function handle = 329707728.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetStringVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetStringVariable", function handle = 329709104.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetThresholdedImage" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetThresholdedImage", function handle = 329707312.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIGetUTF8Text" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIGetUTF8Text", function handle = 329706752.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIInit4" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIInit4", function handle = 329708688.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIMeanTextConf" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIMeanTextConf", function handle = 329706688.
Tesseract Information: 0 : Trying to load native function "TessBaseAPIRecognize" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPIRecognize", function handle = 329706912.
Tesseract Information: 0 : Trying to load native function "TessBaseAPISetDebugVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetDebugVariable", function handle = 329709424.
Tesseract Information: 0 : Trying to load native function "TessBaseAPISetImage2" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetImage2", function handle = 329707536.
Tesseract Information: 0 : Trying to load native function "TessBaseAPISetInputName" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetInputName", function handle = 329709840.
Tesseract Information: 0 : Trying to load native function "TessBaseAPISetPageSegMode" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetPageSegMode", function handle = 329707760.
Tesseract Information: 0 : Trying to load native function "TessBaseAPISetRectangle" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetRectangle", function handle = 329707440.
Tesseract Information: 0 : Trying to load native function "TessBaseAPISetVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessBaseAPISetVariable", function handle = 329709504.
Tesseract Information: 0 : Trying to load native function "TessDeleteBlockList" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessDeleteBlockList", function handle = 329710688.
Tesseract Information: 0 : Trying to load native function "TessDeleteIntArray" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessDeleteIntArray", function handle = 329710720.
Tesseract Information: 0 : Trying to load native function "TessDeleteText" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessDeleteText", function handle = 329710784.
Tesseract Information: 0 : Trying to load native function "TessDeleteTextArray" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessDeleteTextArray", function handle = 329710736.
Tesseract Information: 0 : Trying to load native function "TessVersion" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessVersion", function handle = 329710800.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorBaseline" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBaseline", function handle = 329705472.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorBegin" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBegin", function handle = 329705760.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorBlockType" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBlockType", function handle = 329705568.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorBoundingBox" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorBoundingBox", function handle = 329705616.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorCopy" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorCopy", function handle = 329705776.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorDelete" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorDelete", function handle = 329705808.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorGetBinaryImage" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorGetBinaryImage", function handle = 329705552.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorGetImage" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorGetImage", function handle = 329705520.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorIsAtBeginningOf" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorIsAtBeginningOf", function handle = 329705696.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorIsAtFinalElement" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorIsAtFinalElement", function handle = 329705664.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorNext" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorNext", function handle = 329705728.
Tesseract Information: 0 : Trying to load native function "TessPageIteratorOrientation" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessPageIteratorOrientation", function handle = 329705440.
Tesseract Information: 0 : Trying to load native function "TessResultIteratorCopy" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function "TessResultIteratorCopy", function handle = 329705200.
Tesseract Information: 0 : Trying to load native function "TessResultIteratorDelete" from the library with handle 329383936...
-- Again the log is from one run (one attempt to execute the single line of code listed at the start of this post).
Conclusion
I am a very experienced programmer (have done win32 to cloud). I spent about 5 hours trying to get this to work without success. One thing I notice is that the log seems to end on TessResultIteratorDelete, and in some cases it starts to print the next line but stops midway (prints "Tesseract Information: 0 : Successfully", or "Tesseract Information: 0 : S").
Any help that you can offer would be greatly appreciated (the more specific the better).
Thank you very much again.
Best,
Conrad
My guess is that the actual tessdata path isn't what you expect. I'd
strongly recommend using absolute paths, especially for ASP.NET
applications, in this case Server.MapPath("~\tessdata") should do the trick.
Note that's just of the top my head so the syntax may not be 100% correct 😉
On Fri, 17 Feb 2017, 16:56 Conrad-S<EMAIL_ADDRESS>wrote:
Hi and thank you very much for your help.
I have done everything that I possibly can think of to attempt to resolve
a "Failed to initialise engine" error.
I read post #277 https://github.com/charlesw/tesseract/issues/277 where
the person stated that adjusting the path to tessdata fixed his issue. But
he did not state the "adjustment" that he made.
Line where the failure occurs
using (var engine = new TesseractEngine (@"./tessdata", "eng",
EngineMode.Default)) {
Error
"Failed to initialise tesseract engine.. See
https://github.com/charlesw/tesseract/wiki/Error-1 for details."}
Tesseract.TesseractException
Stack Trace
Data {System.Collections.ListDictionaryInternal}
System.Collections.IDictionary {System.Collections.ListDictionaryInternal}
HResult -2146233088 int
HelpLink null string
InnerException null System.Exception
Message "Failed to initialise tesseract engine.. See
https://github.com/charlesw/tesseract/wiki/Error-1 for details." string
Source "Tesseract" string
StackTrace " at Tesseract.TesseractEngine.Initialise(String datapath,
String language, EngineMode engineMode, IEnumerable1 configFiles,
IDictionary2 initialValues, Boolean setOnlyNonDebugVariables)\r\n at
Tesseract.TesseractEngine..ctor(String datapath, String language,
EngineMode engineMode, IEnumerable1 configFiles, IDictionary2
initialOptions, Boolean setOnlyNonDebugVariables)\r\n at
Tesseract.TesseractEngine..ctor(String datapath, String language,
EngineMode engineMode)\r\n at OCRTestOne.OCRTest.Go() in
C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\Classes\OCRTest.cs:line 22"
string
TargetSite {Void Initialise(System.String, System.String,
Tesseract.EngineMode, System.Collections.Generic.IEnumerable1[System.String],
System.Collections.Generic.IDictionary2[System.String,System.Object],
Boolean)} System.Reflection.MethodBase {System.Reflection.RuntimeMethodInfo}
Development Environment
Windows Server 2012 R2 (virtual machine running in Azure)
Visual Studio 2015 (Version 14.0.25431.01 Update 3)
Project Type: ASP.NET (also tried console app).
Framework Target 452 (.NET Framework 4.5.2).
Verified that Visual C++ 32-bit and 64-bit runtimes are both
installed
Nuget Package: Tesseract by Charles Weld 3.0.2
tessdata folder exists with English language files underneath (I
added it). Files set to Copy Always.
32 and 64-bit assemblies (liblept172.dll and libtesseract304.dll)
exist in the x86 and x64 folders and are set to Copy Always (they seem to
be copied to the lib folder).
Trace
Turned on verbose mode tracing. The trace below is based on one execution
attempt (I tried to attach a text file but the site said: "We don't support
that file type"). Apologies for the long trace
Tesseract Information: 0 : Current OS: Windows
Tesseract Information: 0 : Current platform: x86
Tesseract Information: 0 : Custom search path is not defined, skipping.
Tesseract Information: 0 : Checking executing application domain location
'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET
Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for
'liblept172.dll' on platform x86.
Tesseract Information: 0 : Checking current application domain location
'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'liblept172.dll' on
platform x86.
Tesseract Information: 0 : Trying to load native library
"C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll"...
Tesseract Information: 0 : Successfully loaded native library
"C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll", handle
= 325451776.
Tesseract Information: 0 : Trying to load native function
"pixaReadMultipageTiff" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixaReadMultipageTiff", function handle = 325469402.
Tesseract Information: 0 : Trying to load native function "pixaGetCount"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixaGetCount", function handle = 325462092.
Tesseract Information: 0 : Trying to load native function "pixaGetPix"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixaGetPix", function handle = 325455972.
Tesseract Information: 0 : Trying to load native function "pixaDestroy"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixaDestroy", function handle = 325456927.
Tesseract Information: 0 : Trying to load native function "pixCreate" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixCreate", function handle = 325465967.
Tesseract Information: 0 : Trying to load native function "pixClone" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixClone",
function handle = 325466612.
Tesseract Information: 0 : Trying to load native function "pixDestroy"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixDestroy", function handle = 325466382.
Tesseract Information: 0 : Trying to load native function "pixGetWidth"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetWidth", function handle = 325460622.
Tesseract Information: 0 : Trying to load native function "pixGetHeight"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetHeight", function handle = 325460452.
Tesseract Information: 0 : Trying to load native function "pixGetDepth"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetDepth", function handle = 325464312.
Tesseract Information: 0 : Trying to load native function "pixGetXRes"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetXRes", function handle = 325463887.
Tesseract Information: 0 : Trying to load native function "pixGetYRes"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetYRes", function handle = 325463882.
Tesseract Information: 0 : Trying to load native function
"pixGetResolution" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetResolution", function handle = 325469492.
Tesseract Information: 0 : Trying to load native function "pixGetWpl" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetWpl", function handle = 325466032.
Tesseract Information: 0 : Trying to load native function "pixSetXRes"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSetXRes", function handle = 325463907.
Tesseract Information: 0 : Trying to load native function "pixSetYRes"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSetYRes", function handle = 325463902.
Tesseract Information: 0 : Trying to load native function
"pixSetResolution" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSetResolution", function handle = 325465697.
Tesseract Information: 0 : Trying to load native function
"pixScaleResolution" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixScaleResolution", function handle = 325467282.
Tesseract Information: 0 : Trying to load native function "pixGetData"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetData", function handle = 325469052.
Tesseract Information: 0 : Trying to load native function
"pixGetInputFormat" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetInputFormat", function handle = 325467712.
Tesseract Information: 0 : Trying to load native function
"pixSetInputFormat" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSetInputFormat", function handle = 325459102.
Tesseract Information: 0 : Trying to load native function
"pixEndianByteSwap" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixEndianByteSwap", function handle = 325468587.
Tesseract Information: 0 : Trying to load native function "pixRead" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixRead",
function handle = 325467527.
Tesseract Information: 0 : Trying to load native function "pixReadMemTiff"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixReadMemTiff", function handle = 325472137.
Tesseract Information: 0 : Trying to load native function "pixWrite" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixWrite",
function handle = 325466617.
Tesseract Information: 0 : Trying to load native function "pixGetColormap"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixGetColormap", function handle = 325464692.
Tesseract Information: 0 : Trying to load native function "pixSetColormap"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSetColormap", function handle = 325468467.
Tesseract Information: 0 : Trying to load native function
"pixDestroyColormap" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixDestroyColormap", function handle = 325465782.
Tesseract Information: 0 : Trying to load native function
"pixConvertRGBToGray" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixConvertRGBToGray", function handle = 325460777.
Tesseract Information: 0 : Trying to load native function
"pixDeskewGeneral" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixDeskewGeneral", function handle = 325460177.
Tesseract Information: 0 : Trying to load native function "pixRotate" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixRotate", function handle = 325466312.
Tesseract Information: 0 : Trying to load native function "pixRotateOrth"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixRotateOrth", function handle = 325468177.
Tesseract Information: 0 : Trying to load native function
"pixOtsuAdaptiveThreshold" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixOtsuAdaptiveThreshold", function handle = 325468767.
Tesseract Information: 0 : Trying to load native function
"pixSauvolaBinarize" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSauvolaBinarize", function handle = 325463922.
Tesseract Information: 0 : Trying to load native function
"pixSauvolaBinarizeTiled" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixSauvolaBinarizeTiled", function handle = 325461412.
Tesseract Information: 0 : Trying to load native function "pixScale" from
the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function "pixScale",
function handle = 325466597.
Tesseract Information: 0 : Trying to load native function "pixcmapCreate"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapCreate", function handle = 325462397.
Tesseract Information: 0 : Trying to load native function
"pixcmapCreateRandom" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapCreateRandom", function handle = 325460482.
Tesseract Information: 0 : Trying to load native function
"pixcmapCreateLinear" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapCreateLinear", function handle = 325460392.
Tesseract Information: 0 : Trying to load native function "pixcmapCopy"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapCopy", function handle = 325463212.
Tesseract Information: 0 : Trying to load native function "pixcmapDestroy"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapDestroy", function handle = 325464252.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetCount" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetCount", function handle = 325466562.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetFreeCount" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetFreeCount", function handle = 325456837.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetDepth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetDepth", function handle = 325468462.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetMinDepth" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetMinDepth", function handle = 325462157.
Tesseract Information: 0 : Trying to load native function "pixcmapClear"
from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapClear", function handle = 325460977.
Tesseract Information: 0 : Trying to load native function
"pixcmapAddColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapAddColor", function handle = 325466517.
Tesseract Information: 0 : Trying to load native function
"pixcmapAddNewColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapAddNewColor", function handle = 325459832.
Tesseract Information: 0 : Trying to load native function
"pixcmapAddNearestColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapAddNearestColor", function handle = 325464897.
Tesseract Information: 0 : Trying to load native function
"pixcmapUsableColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapUsableColor", function handle = 325468852.
Tesseract Information: 0 : Trying to load native function
"pixcmapAddBlackOrWhite" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapAddBlackOrWhite", function handle = 325460002.
Tesseract Information: 0 : Trying to load native function
"pixcmapSetBlackAndWhite" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapSetBlackAndWhite", function handle = 325456282.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetColor", function handle = 325466527.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetColor32" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetColor32", function handle = 325458992.
Tesseract Information: 0 : Trying to load native function
"pixcmapResetColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapResetColor", function handle = 325457197.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetIndex" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetIndex", function handle = 325466567.
Tesseract Information: 0 : Trying to load native function
"pixcmapHasColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapHasColor", function handle = 325458122.
Tesseract Information: 0 : Trying to load native function
"pixcmapCountGrayColors" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapCountGrayColors", function handle = 325462272.
Tesseract Information: 0 : Trying to load native function
"pixcmapCountGrayColors" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapCountGrayColors", function handle = 325462272.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetNearestIndex" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetNearestIndex", function handle = 325459557.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetNearestGrayIndex" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetNearestGrayIndex", function handle = 325466522.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetComponentRange" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetComponentRange", function handle = 325459077.
Tesseract Information: 0 : Trying to load native function
"pixcmapGetExtremeValue" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGetExtremeValue", function handle = 325458877.
Tesseract Information: 0 : Trying to load native function
"pixcmapGrayToColor" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGrayToColor", function handle = 325464607.
Tesseract Information: 0 : Trying to load native function
"pixcmapColorToGray" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapColorToGray", function handle = 325469397.
Tesseract Information: 0 : Trying to load native function
"pixcmapColorToGray" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapColorToGray", function handle = 325469397.
Tesseract Information: 0 : Trying to load native function
"pixcmapToRGBTable" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapToRGBTable", function handle = 325469742.
Tesseract Information: 0 : Trying to load native function
"pixcmapSerializeToMemory" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapSerializeToMemory", function handle = 325465572.
Tesseract Information: 0 : Trying to load native function
"pixcmapDeserializeFromMemory" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapDeserializeFromMemory", function handle = 325460637.
Tesseract Information: 0 : Trying to load native function
"pixcmapGammaTRC" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapGammaTRC", function handle = 325470952.
Tesseract Information: 0 : Trying to load native function
"pixcmapContrastTRC" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapContrastTRC", function handle = 325462497.
Tesseract Information: 0 : Trying to load native function
"pixcmapShiftIntensity" from the library with handle 325451776...
Tesseract Information: 0 : Successfully loaded native function
"pixcmapShiftIntensity", function handle = 325460387.
Tesseract Information: 0 : Current platform: x86
Tesseract Information: 0 : Custom search path is not defined, skipping.
Tesseract Information: 0 : Checking executing application domain location
'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET
Files\root\99b032d6\fb07eb9\assembly\dl3\26c8fcd5\00342a46_d765d101' for
'libtesseract304.dll' on platform x86.
Tesseract Information: 0 : Checking current application domain location
'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'libtesseract304.dll'
on platform x86.
Tesseract Information: 0 : Trying to load native library
"C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll"...
Tesseract Information: 0 : Successfully loaded native library
"C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\libtesseract304.dll",
handle = 329383936.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIAnalyseLayout" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIAnalyseLayout", function handle = 329706928.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIClear" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIClear", function handle = 329706560.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPICreate" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPICreate", function handle = 329709952.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIDelete" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIDelete", function handle = 329709920.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIDetectOS" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIDetectOS", function handle = 329706368.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetBoolVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetBoolVariable", function handle = 329709264.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetDoubleVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetDoubleVariable", function handle = 329709184.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetHOCRText" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetHOCRText", function handle = 329706736.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetIntVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetIntVariable", function handle = 329709344.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetIterator" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetIterator", function handle = 329706784.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetPageSegMode" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetPageSegMode", function handle = 329707728.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetStringVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetStringVariable", function handle = 329709104.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetThresholdedImage" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetThresholdedImage", function handle = 329707312.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIGetUTF8Text" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIGetUTF8Text", function handle = 329706752.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIInit4" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIInit4", function handle = 329708688.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIMeanTextConf" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIMeanTextConf", function handle = 329706688.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPIRecognize" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPIRecognize", function handle = 329706912.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPISetDebugVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPISetDebugVariable", function handle = 329709424.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPISetImage2" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPISetImage2", function handle = 329707536.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPISetInputName" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPISetInputName", function handle = 329709840.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPISetPageSegMode" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPISetPageSegMode", function handle = 329707760.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPISetRectangle" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPISetRectangle", function handle = 329707440.
Tesseract Information: 0 : Trying to load native function
"TessBaseAPISetVariable" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessBaseAPISetVariable", function handle = 329709504.
Tesseract Information: 0 : Trying to load native function
"TessDeleteBlockList" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessDeleteBlockList", function handle = 329710688.
Tesseract Information: 0 : Trying to load native function
"TessDeleteIntArray" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessDeleteIntArray", function handle = 329710720.
Tesseract Information: 0 : Trying to load native function "TessDeleteText"
from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessDeleteText", function handle = 329710784.
Tesseract Information: 0 : Trying to load native function
"TessDeleteTextArray" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessDeleteTextArray", function handle = 329710736.
Tesseract Information: 0 : Trying to load native function "TessVersion"
from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessVersion", function handle = 329710800.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorBaseline" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorBaseline", function handle = 329705472.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorBegin" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorBegin", function handle = 329705760.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorBlockType" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorBlockType", function handle = 329705568.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorBoundingBox" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorBoundingBox", function handle = 329705616.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorCopy" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorCopy", function handle = 329705776.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorDelete" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorDelete", function handle = 329705808.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorGetBinaryImage" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorGetBinaryImage", function handle = 329705552.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorGetImage" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorGetImage", function handle = 329705520.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorIsAtBeginningOf" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorIsAtBeginningOf", function handle = 329705696.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorIsAtFinalElement" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorIsAtFinalElement", function handle = 329705664.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorNext" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorNext", function handle = 329705728.
Tesseract Information: 0 : Trying to load native function
"TessPageIteratorOrientation" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessPageIteratorOrientation", function handle = 329705440.
Tesseract Information: 0 : Trying to load native function
"TessResultIteratorCopy" from the library with handle 329383936...
Tesseract Information: 0 : Successfully loaded native function
"TessResultIteratorCopy", function handle = 329705200.
Tesseract Information: 0 : Trying to load native function
"TessResultIteratorDelete" from the library with handle 329383936...
-- Again the log is from one run (one attempt to execute the single line
of code listed at the start of this post).
Conclusion
I am a very experienced programmer (have done win32 to cloud). I spent
about 5 hours trying to get this to work without success. One thing I
notice is that the log seems to end on TessResultIteratorDelete, and in
some cases it starts to print the next line but stops midway (prints
"Tesseract Information: 0 : Successfully", or "Tesseract Information: 0 :
S").
Any help that you can offer would be greatly appreciated (the more
specific the better).
Thank you very much again.
Best,
Conrad
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/charlesw/tesseract/issues/318, or mute the thread
https://github.com/notifications/unsubscribe-auth/AAPzyPb7KW6g6m7pE5HhORKHByTo19EIks5rdTZ4gaJpZM4MD5r6
.
Charles,
Thank you for the fast response!
It works!
I had tried that before but changed it while trying other things.
Below is a summary the issues (or gotchas) that I encountered and how resolved:
Initially I couldn't get it to work so I experimented with settings.
I had tried multiple versions of the Tesserat Nuget package. So I needed to double-check that the correct versions of the assemblies were in the x86 and x64 folders. I did this by downloading the latest source (in my case tesseract-3.0.2.zip) and getting the latest assemblies from the "..src\lib\TesseractOcr" x64 and x86 folders.
The rest of the checklist is as follows:
Make sure that the assemblies in the x86 and x64 are set to "Copy Always" (right-click the file in Visual Studio, open the properties dialog, and set the property).
Make sure that the language files are in the Visual Studio project, match the Tesserat Nuget package version, and are set to "Copy Always".
Use the following include statement and code to set the path to tessdata:
using System.Web.Hosting;
string tessPath = HostingEnvironment.MapPath(@"~/tessdata");
Make sure that the code includes: using System.Drawing; (otherwise an error occurs saying that a bitmap is being used by not referenced).
Again, thank you very much for your help. This looks like a great library! I cannot wait to try it out!!
Best,
Conrad
By the way is there documentation that lists functions in the library and their use?
For example, I would like to open a PDF and read it, but don't know what function to use..
Thanks again!
Most of the functions are documented using XML comments so VS should
provide some explanation through the auto complete and it would be possible
to generate some API doco from said comments. Otherwise check this issue
tracker and/or stack overflow for anything more specific.
In regards to pdf support, that's currently only supported in the
prerelease version and only supports generation of pdf (image + searchable
text) not reading pdf. In your case you'll need to find another library to
load the PDF and extract any contained image for processing.
On Fri, 17 Feb 2017, 17:56 Conrad-S<EMAIL_ADDRESS>wrote:
By the way is there documentation that lists functions in the library and
their use?
For example, I would like to open a PDF and read it, but don't know what
function to use..
Thanks again!
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/charlesw/tesseract/issues/318#issuecomment-280570159,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAPzyAG4WFMGMAxUHRkabyu-TqpfBMXDks5rdUSxgaJpZM4MD5r6
.
Thanks. Nothing is documented in Visual Studio, but auto-complete (intellisense) does work. I can probably use object browser to view the methods, but it would be good to have a document somewhere with the list of methods and their use (also helps to understand overall functionality).
I didn't realize that it cannot extract images from pdf's; that was my main purpose. Do you know of a .NET library that can extract the images? Thanks again!!
Mmm the XML document file mustn't be being included property in the nuget
package.
Anyway in regards to handling PDFs I'm not really sure what's available
these days for .net. That being said, as of a couple years ago, a good
stable open source option was itextsharp though it's no longer free for
commercial use. However there are still copies of the code available before
the license was changed but I'm not aware of anyone actively maintaining
these forks.
On Fri, 17 Feb 2017, 18:18 Conrad-S<EMAIL_ADDRESS>wrote:
Thanks. Nothing is documented in Visual Studio, but auto-complete
(intellisense) does work. I can probably use object browser to view the
methods, but it would be good to have a document somewhere with the list of
methods and their use (also helps to understand overall functionality).
I didn't realize that it cannot extract images from pdf's; that was my
main purpose. Do you know of a .NET library that can extract the images?
Thanks again!!
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/charlesw/tesseract/issues/318#issuecomment-280573566,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAPzyKgNDDkoOeg9G8noVzipga8E7EgSks5rdUm1gaJpZM4MD5r6
.
You can get the old version from nuGet
https://www.nuget.org/packages/iTextSharp-LGPL/
Thanks again! This is great software.
By the way I notice that it always seems to load the 32-bit version. I even tried setting the MapPath.
Tesseract.TesseractEnviornment.CustomSearchPath = HostingEnvironment.MapPath(@"~/x64");
Do you know how to convince it to load the 64-bit DLL?
Tesseract Information: 0 : Current OS: Windows
Tesseract Information: 0 : Current platform: x86
Tesseract Information: 0 : Custom search path is not defined, skipping.
Tesseract Information: 0 : Checking executing application domain location 'C:\Windows\Microsoft.NET\Framework\v4.0.30319\Temporary ASP.NET Files\vs\99b032d6\fb07eb9\assembly\dl3\8760118f\00342a46_d765d101' for 'liblept172.dll' on platform x86.
Tesseract Information: 0 : Checking current application domain location 'C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One' for 'liblept172.dll' on platform x86.
Tesseract Information: 0 : Trying to load native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll"...
Tesseract Information: 0 : Successfully loaded native library "C:\Projects\SANDBOX\OCR-Test-One\OCR-Test-One\x86\liblept172.dll", handle = 369360896.
I'll answer my own question. Got it to work (Visual Studio 2015; ASP.NET MVC; .NET 452):
Do NOT set Tesseract.TesseractEnviornment.CustomSearchPath
Select Debug -> Options -> Projects and Solutions -> Web Projects.
Select User the 64 bit version of IIS Express for web sites and projects.
Set the debug configuration to x64 (Configuration Manager next to the Debug dropdown)
Restart Visual Studio, open the solution and run.
You may also want to set up a trace file so you can see that it loaded the 64-bit assembly:
<system.diagnostics>
</system.diagnostics>
Don't forget to turn off logging after you are finished (performance).
Hope this helps.
Fail with asp.net core, i've tried exactly path, but still throw exception failed to initalise... Please help me
After banging my head against my desk for 30min, I finally realized the path needs to be to the FOLDER containing the .traineddata files (ie: "c:\tessdata"), not the path to one of the .traineddata files (ie "c:\tessdata\eng.traineddata"). Doh!
Hopefully this saves someone's desk (and head).
Fail with asp.net core, i've tried exactly path, but still throw exception failed to initalise... Please help me
Try using tessdata_fast rather then tessdata version.
Anyone knows how to read data from table in image with this library?
|
2025-04-01T06:38:10.769639
| 2024-06-24T17:16:17
|
2370727861
|
{
"authors": [
"davidhardy",
"ericjbohm"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4606",
"repo": "charmplusplus/charm",
"url": "https://github.com/charmplusplus/charm/pull/3829"
}
|
gharchive/pull-request
|
bugfix: remove stray endif, revise comments to match
Remove stray endif, revise comments to match.
Closes issue #3828
On Summit now I can build pamilrts successfully.
|
2025-04-01T06:38:10.797234
| 2017-02-04T19:11:18
|
205369998
|
{
"authors": [
"etimberg",
"frlinw",
"nutanlade"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4607",
"repo": "chartjs/Chart.js",
"url": "https://github.com/chartjs/Chart.js/issues/3868"
}
|
gharchive/issue
|
Issue in styling legend in Pie charts
Hi All,
I'm trying to style my pie chart as first image below. But with below code I'm getting as shown in second image. Even I tried the solution as mentioned here 2253. Then no legends are displayed. Also I don't want white border between pies.
I'm using ChartJs 2.4.0 version. Please help me in resolving the issue.
var legend = chartInstance.generateLegend();
document.getElementById("legend").innerHTML = legend;
@nutanlade to get the legend to look like the top image, you're going to need to use some CSS to change the <li> elements in the legend.
To remove the border:
data: {
datasets: [{
borderWidth: 0
}]
}
Thanks @etimberg for your quick reply. I'm using angular js code for pie chart. Below is my code for generating pie chart. I'm not able to apply your code to remove border. Is there any default global option to disable?
<canvas id="monthlyExpensesChart" class="chart chart-pie" chart-data="$ctrl.communitytotalCostArray" chart-labels="$ctrl.communityCategoryNamesArray" chart-options="$ctrl.options" chart-legend="true" height="250" chart-colors="$ctrl.colors"> </canvas>
Also for legend now it showing like square but color is black. How can I get pie colors to populate legend labels?
@nutanlade the global border option would be Chart.defaults.global.elements.arc.borderWidth
For the legend, you would need to style the <li> elements with CSS. You would probably want to set the first one to the first color. I think something like the following would work.
#legend > li:nth-child(0) {
color: red;
}
#legend > li:nth-child(1) {
color: yellow;
}
#legend > li:nth-child(2) {
color: blue;
}
Controller:
Create an object of legend items with color & label :
# Chart conf
scope.chart = []
scope.chart.labels = ['Food', 'Fuel']
scope.chart.colors = ['red', 'yellow']
# Legend
scope.legend = []
for label, i in scope.chart.labels
scope.legend[i] =
title: label
color: scope.chart.colors[i]
(it's coffee script, use http://js2.coffee/ if you want a translation)
HTML:
Loop on legend items to display them and you can do whatever you want with styles
<div class="widget-body">
<div class="row">
<div class="col-md-12 col-lg-7">
<div class="widget-pie-chart">
<canvas beautiful-chart chart-type="::chart.type" chart-options="::chart.options" chart-datasets="::chart.datasets" chart-labels="chart.labels" chart-data="chart.data"></canvas>
<canvas id="monthlyExpensesChart" class="chart chart-pie" chart-data="communitytotalCostArray" chart-labels="communityCategoryNamesArray" chart-options="options" height="250" chart-colors="colors"></canvas>
</div>
</div>
<div class="col-md-12 col-lg-5">
<div class="widget-pie-legend">
<a ng-repeat="item in ::legend" href="">
<div class="pie-legend-item">
<span class="legend-item-box" ng-style="::{'color': item.color}"></span>
<span class="legend-item-label">{{::item.title}}</span>
</div>
</a>
</div>
</div>
</div>
</div>
Result
You should see something like that
Thanks @etimberg and @frlinw . Issue resolved. I disabled tooltips but while hovering on chart color is changing which I don't want. I tried below option but didn't worked.
showTooltips: false
@nutanlade to turn off hover you can do
options: {
hover: {
mode: false
}
}
Thanks @etimberg. Issue resolved.
|
2025-04-01T06:38:10.804830
| 2020-10-09T08:26:26
|
717957146
|
{
"authors": [
"dcyou",
"etimberg",
"kurkle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4608",
"repo": "chartjs/Chart.js",
"url": "https://github.com/chartjs/Chart.js/issues/7866"
}
|
gharchive/issue
|
Circular gradiant
Hi all, (thanks for that awesome lib)
I am using Chart.js v3 and I am trying to create a gradient
My fiddle is here:
https://jsfiddle.net/1wfd2ztx/1/
What I would like to do is to start at top center with the green and finish with the red at top center too with a gradient
And ++ draw with red under the green if the percent is more than 100
@dcyou this is a very interesting question! A radial gradient is probably not the right solution here since it changes colour as you move out from a center point. You probably want something more akin to a SweepGradient but as far as I am aware, nothing like this exists natively. If this were WebGL, a custom shader would be pretty trivial to write.
The best idea I have at the moment is to draw the colour pattern you want on another canvas (handling resize accordingly) then create a CanvasPattern from that canvas and set that as the background colour of the dataset.
This is quite intriguing. I'll try and prototype something out and see where I get.
@dcyou got a working implementation. Generates a CanvasPattern using webgl then renders that as the background image. I don't think the plugin part is entirely correct. I suspect on resize it may be wrong, but have not attempted to test it.
https://jsfiddle.net/6d0gv7w9/
awesome! @etimberg thanks a lot I will have a look
Played around with @etimberg's fiddle and applied some changes: https://jsfiddle.net/x1t38f7v/
Precicely, multiplied the webgl canvas size by devicePixelRatio and changed the plugin implementation to beforeDatasetsUpdate hook.
very interesting too
question:
do you know why at bottom we have the blank?
do you know why at bottom we have the blank?
I'm guessing you have a devicePixelRatio greater than 1? Does that appear in the fiddle where I multiplied the canvas size with it?
i actually don't think devicePixelRatio needs to be used. chart.chartArea will already include the devicePixelRatio since we up the size of the canvas by it inside of https://github.com/chartjs/Chart.js/blob/531d4c09bf5778274d362eaf1f5dd17a9a08ec7b/src/helpers/helpers.dom.js#L148-L149
|
2025-04-01T06:38:10.806595
| 2024-02-14T01:16:06
|
2133386111
|
{
"authors": [
"LeeLenaleee",
"zieka"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4609",
"repo": "chartjs/Chart.js",
"url": "https://github.com/chartjs/Chart.js/pull/11669"
}
|
gharchive/pull-request
|
refactor: replace Math.pow in favor of the ** operator
friendly PR to improve readability
Use this link to re-run the recipe: https://app.moderne.io/recipes/org.openrewrite.codemods.cleanup.PreferExponentiationOperator?organizationId=UG9wdWxhciBKYXZhU2NyaXB0
I find the Math.pow more readable myself and would be against this change
|
2025-04-01T06:38:10.828935
| 2022-06-09T12:00:41
|
1266025892
|
{
"authors": [
"bensmiley",
"letzter"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4610",
"repo": "chat-sdk/chat-sdk-android",
"url": "https://github.com/chat-sdk/chat-sdk-android/issues/734"
}
|
gharchive/issue
|
Unable to login with chat sdk
Is the bug present in the demo Chat SDK project?
Yes
What modifications have you made to the Chat SDK?
None
Android Version:
11
Steps taken to reproduce the problem:
Insert username and password and press register
Expected result:
Register user and enter profile screen
Actual result:
Authentication progress bar keeps spinning "forever"
Comments:
I have been trying to integrate this chat sdk on my app for some time, but cant seem to go through the authentication
Took a look at logcat and it seems that something is failing. I even tried to register in the example android app, but it doesn't go through. The logs obtained are the ones below:
2022-06-09 12:47:26.960 10270-10318/sdk.chat.android.live W/System: Ignoring header X-Firebase-Locale because its value was null.
2022-06-09 12:47:27.495 10270-10318/sdk.chat.android.live W/System: Ignoring header X-Firebase-Locale because its value was null.
2022-06-09 12:47:27.696 10270-10318/sdk.chat.android.live D/FirebaseAuth: Notifying id token listeners about user ( bTlcQnKU0AQ1ClDop8lhrSRukS63 ).
2022-06-09 12:47:27.696 10270-10318/sdk.chat.android.live D/FirebaseAuth: Notifying auth state listeners about user ( bTlcQnKU0AQ1ClDop8lhrSRukS63 ).
2022-06-09 12:47:27.718 10270-10577/sdk.chat.android.live I/greenDAO: Creating tables for schema version 20
2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: 2022-06-09 12:47:27 [RxComputationThreadPool-1] sdk.chat.core.dao.DaoCore.openDB()
2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: INFO: Database: e92eded1acf5e51967b47107e633e719 setup correctly
2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: 2022-06-09 12:47:27 [RxComputationThreadPool-1] sdk.chat.firebase.adapter.FirebaseAuthenticationHandler.lambda$authenticateWithUser$9$FirebaseAuthenticationHandler()
2022-06-09 12:47:27.743 10270-10577/sdk.chat.android.live I/System.out: INFO: authenticateWithUser: bTlcQnKU0AQ1ClDop8lhrSRukS63
2022-06-09 12:48:08.817 10270-10631/sdk.chat.android.live V/FA: Connecting to remote service
2022-06-09 12:48:08.841 10270-10631/sdk.chat.android.live D/FA: Connected to remote service
2022-06-09 12:48:08.842 10270-10631/sdk.chat.android.live V/FA: Processing queued up service tasks: 1
2022-06-09 12:48:08.854 10270-10627/sdk.chat.android.live W/System: Ignoring header If-None-Match because its value was null.
2022-06-09 12:48:13.859 10270-10631/sdk.chat.android.live V/FA: Inactivity, disconnecting from the service
Thanks in advance
I checked out the current version from Github and I can log in... Have you tried this without connecting it to your Firebase account? Have you enabled the login methods you want to use from the Firebase console?
@letzter I need some more information because I tried it myself and it worked just fine. Can you send me a video?
Here goes the video showing the process that I'm following.
Is it possible that I'm having this problem because I'm running the app on an emulator from android studio instead of a real device, or should it it work as well?
chat-sdk.webm.zip
It is possible. I never test on the emulator. Do you have Google Play services installed? If that were missing it Wouk's cause issues.
@bensmiley the issue was with the emulator. I created a new one in which it all worked out well.
Thank you very much for the support
|
2025-04-01T06:38:10.843492
| 2019-04-19T22:29:40
|
435321626
|
{
"authors": [
"Ebioro",
"chatch"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4611",
"repo": "chatch/stellar-hd-wallet",
"url": "https://github.com/chatch/stellar-hd-wallet/issues/13"
}
|
gharchive/issue
|
update stellar-base dependency
Tried to use the wallet for an angular 6 project and could not build it because the angular cli started complaining about the crypto library. I then found out that wallet is currently using the stellar-base 0.11. The guys from stellar have removed the crypto library in version 0.13 of the stellar-base sdk which should have fixed the issue. More details can be found here
thanks @Ebioro i've updated and published new release 0.10.0
you are welcome! I had an open question on stellar stack exchange about this (under yulemata username) I will add there that you guys have updated it and publish a new release!
|
2025-04-01T06:38:10.913619
| 2020-05-25T18:34:38
|
624436697
|
{
"authors": [
"JPinkney",
"amisevsk",
"davidfestal",
"sleshchenko"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4612",
"repo": "che-incubator/che-workspace-operator",
"url": "https://github.com/che-incubator/che-workspace-operator/pull/83"
}
|
gharchive/pull-request
|
Make operator installable via olm
Signed-off-by: Josh Pinkney<EMAIL_ADDRESS>What does this PR do?
This PR prepares everything needed for olm in the older olm format. The Makefile has been modified so that it is easy to load everything onto the cluster and so that you can see che-workspace-operator in the OperatorHub catalog.
The csv was made with make gen_csv except for a few custom things like the alm-examples
What issues does this PR fix or reference?
Is it tested? How?
Deploying the controller with olm
In order to deploy the controller you need to first create the olm image, push it to the application registry on quay and then deploy the crd registry to your cluster.
You can do this by using the Makefile olm_full_start rule:
make olm_full_start
Before doing this you need to set environment variables QUAY_USERNAME, QUAY_PASSWORD, QUAY_NAMESPACE (most likely this is the same as QUAY_USERNAME), CATALOG_IMAGE
You might have to manually create the application registry on quay and set it to public.
To do this, use + in the top right hand corner of quay and click new repository. Then instead of having a Container Image Repository set that dropdown to Application repository and make sure it's public. The repostory name you choose for the application registry must be the same name as you have set in the CATALOG_IMAGE.
If you already have it deployed to the quay application registry then you can just use
make olm_start
to deploy the crd registry to your cluster
To remove the crd registry use
make olm_uninstall
cc: @davidfestal since you're familiar with deploying via OLM.
Why not use the new bundle format as a source, and just export the index in the old format when needed ?
https://github.com/operator-framework/operator-registry/blob/master/docs/design/opm-tooling.md#export
/retest
Yeah at this point I think I'm going to close this PR and then we can transfer the catalogsource to the web-terminal-operator repo
|
2025-04-01T06:38:11.051626
| 2016-09-24T21:31:01
|
179052989
|
{
"authors": [
"JasonMorgan"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4613",
"repo": "chef-cookbooks/powershell",
"url": "https://github.com/chef-cookbooks/powershell/pull/98"
}
|
gharchive/pull-request
|
removed windows_reboot and windows::reboot_handler.
Description
Drops the windows_reboot resource and the windows::reboot_handler recipe. The no longer exist in the windows cookbook.
Issues Resolved
None that I know of.
Check List
[ ] All tests pass. See https://github.com/chef-cookbooks/community_cookbook_documentation/blob/master/TESTING.MD
[x] New functionality includes testing.
[x] New functionality has been documented in the README if applicable
[x] The CLA has been signed. See https://github.com/chef-cookbooks/community_cookbook_documentation/blob/master/CONTRIBUTING.MD
That resource and recipe has been dropped from the windows cookbook. replace windows_reboot with the reboot resource.
I didn't actually intend to have that second commit added into this PR. Although that being said it is related and fixes the problem of the recipe not actually allowing a reboot to occur. I was going to make a second PR for it after this got committed.
Updated to address @smurawski comment.
|
2025-04-01T06:38:11.055022
| 2015-05-07T21:06:36
|
74127684
|
{
"authors": [
"RoboticCheese",
"tas50"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4614",
"repo": "chef-cookbooks/windows",
"url": "https://github.com/chef-cookbooks/windows/pull/198"
}
|
gharchive/pull-request
|
Make auto_run's root registry key configurable
I recently found myself working on a cookbook for an app that only installs on
a per-user basis, putting all its registry entries under the
HKEY_CURRENT_USER root key instead of HKEY_LOCAL_MACHINE.
This change adds a root property to the windows_auto_run resource,
maintaining the current behavior as a default and allowing support for
user-level applications as well.
Signed-off-by: Jonathan Hartman<EMAIL_ADDRESS>
@RoboticCheese If you want to rebase this against master it would be nice to get merged in
@RoboticCheese Ping on a rebase here. I'd love to pull this in before we move this into chef-client so we can pull in this new functionality as well.
Thanks for the reminder about this one. Rebased and updated. Seems to work again, though my Windows is pretty atrophied RN 🤞
Describing test::autorun
Context windows_auto_run
[+] does not auto-run Notepad for the machine 916ms
[+] auto-runs Wordpad for the current user 95ms
Tests completed in 1.01s
Passed: 2 Failed: 0 Skipped: 0 Pending: 0
Finished verifying <autorun-windows-10> (0m20.17s).
Thanks. I'll get this addition in when we move this into chef with Chef 14
|
2025-04-01T06:38:11.057728
| 2018-05-05T00:04:01
|
320461982
|
{
"authors": [
"derekgroh",
"tas50"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4615",
"repo": "chef-cookbooks/windows_firewall",
"url": "https://github.com/chef-cookbooks/windows_firewall/pull/27"
}
|
gharchive/pull-request
|
update readme to match resource
Description
Updates documentation to match the resource rule
Issues Resolved
N/A
Check List
[X] All tests pass. See https://github.com/chef-cookbooks/community_cookbook_documentation/blob/master/TESTING.MD
[N/A] New functionality includes testing.
[X] New functionality has been documented in the README if applicable
Thanks
|
2025-04-01T06:38:11.064872
| 2023-10-19T05:00:24
|
1951228346
|
{
"authors": [
"AadeshNichite"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4616",
"repo": "chef/automate",
"url": "https://github.com/chef/automate/pull/8269"
}
|
gharchive/pull-request
|
Moved protractor from dependency to dev-dependency
:nut_and_bolt: Description: What code changed, and why?
Signed-off-by: AadeshNichite<EMAIL_ADDRESS>
Moved protractor from dependency to dev-dependency
:chains: Related Resources
https://chefio.atlassian.net/browse/CHEF-7042
:+1: Definition of Done
:athletic_shoe: How to Build and Test the Change
:white_check_mark: Checklist
All PRs must tick these:
[ ] I have read the CONTRIBUTING document.
[ ] All commits signed-off for the Developer Certification of Origin.
With occasional exceptions, all PRs from Progress employees must tick these:
[ ] Is the code clear? (complicated code or lots of comments--subdivide and use well-named methods, meaningful variable names, etc.)
[ ] Consistency checked? (user notifications, user prompts, visual patterns, code patterns, variable names)
[ ] Repeated code blocks eliminated? (adapt and reuse existing components, blocks, functions, etc.)
[ ] Spelling, grammar, typos checked? (at a minimum use make spell in any component directory)
[ ] Code well-formatted? (indents, line breaks, etc. improve rather than hinder readability)
All PRs from Progress employees should tick these if appropriate:
[ ] Tests added/updated? (all new code needs new tests)
[ ] Docs added/updated? (all customer-facing changes)
Please add a note next to any checkbox above if you are NOT ticking it.
:camera: Screenshots, if applicable
No need to work this, input from @arunjn-progress.
|
2025-04-01T06:38:11.066296
| 2015-08-28T22:46:35
|
103807706
|
{
"authors": [
"danielsdeleo"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4617",
"repo": "chef/chef-dk",
"url": "https://github.com/chef/chef-dk/pull/491"
}
|
gharchive/pull-request
|
chef clean-policy-revisions command
Adds a subcommand to delete unused policy revisions, and fixes some minor issues with the service class discovered in manual integration testing.
CLI output is like this:
bundle exec bin/chef clean-policy-revisions -c ~/.chef/localserver.rb
DELETE aar 37f9b658cdd1d9319bac8920581723efcc2014304b5f3827ee0779e10ffbdcc9
DELETE aar 5c5922d9febb3ee758c60b0bdd423185c7a4e88450c4fdf0cf90450cfcbfe88a
DELETE aar 7a812b7e7570b93b6672e8befab0423be68f8b44109adb5eb0d3cd9a044504c8
DELETE aar d81e80ae9bb9778e8c4b7652d29b11d2111e763a840d0cadb34b46a8b2ca4347
DELETE aar e67f1b04a055017e7de02acb166e2ab18b4682313f381fae6e2fa8f4bc7db4ae
DELETE jenkins 613f803bdd035d574df7fa6da525b38df45a74ca82b38b79655efed8a189e073
DELETE jenkins 6fe753184c8946052d3231bb4212116df28d89a3a5f7ae52832ad408419dd5eb
DELETE jenkins 9c2cae07a5c68ae0ab7d06165e90ded1325a8b0ca94d78cfc2efa4dcfea36244
DELETE jenkins cc1a0801e75df1d1ea5b0d2c71ba7d31c539423b81478f65e6388b9ee415ad87
DELETE jenkins cec985e916690e82c792114c5b10a4e4694981804bcbc5ff89556e6709c28ec1
|
2025-04-01T06:38:11.073863
| 2021-03-18T22:17:52
|
835317078
|
{
"authors": [
"tas50"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4618",
"repo": "chef/chef",
"url": "https://github.com/chef/chef/issues/11206"
}
|
gharchive/issue
|
Test
Note: GitHub integration record type. Please leave out any sensitive information.Job to be Done:
As an ,
I want to
So that <reason for needing job/benefit>
Description:
Acceptance Criteria:
Aha! Link: https://chef.aha.io/features/INFRA-1
sdfsadfsdf
|
2025-04-01T06:38:11.075316
| 2017-02-16T05:09:01
|
208018502
|
{
"authors": [
"coderanger",
"tas50"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4619",
"repo": "chef/chef",
"url": "https://github.com/chef/chef/issues/5820"
}
|
gharchive/issue
|
Use a real KDF for data bag encryption secrets
Currently we use a single round of SHA256 for deriving the actual AES key. This is okay, but it would be better to use a real-er key derivation function some day like PBKDF2 or if we want to be really fancy, Argon2. This would require a new encryptor version most likely so it will probably take a while to get this in.
Just for the sake of prioritization I can't think of an attack where this would matter so probably can put it off until such time as we're making a v4 encryption for another reason and just slide it in for extra crypto points.
I'm going to close this out at this point since the future of secrets management in Chef Infra Client is not encrypted data bags.
|
2025-04-01T06:38:11.103677
| 2023-07-11T14:22:13
|
1799056362
|
{
"authors": [
"enoch1025",
"ibasaisaac"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4620",
"repo": "cheng-01037/Self-supervised-Fewshot-Medical-Image-Segmentation",
"url": "https://github.com/cheng-01037/Self-supervised-Fewshot-Medical-Image-Segmentation/issues/39"
}
|
gharchive/issue
|
I want to run a single category dataset. How can I set it to a single category
I want to run a single category dataset. How can I set it to a single category
please tell me, thank eyeryone!!!!!!!!
./examples/train_ssl_abdominal_ct.sh
train_SABS_Superpix_lbgroup0_scale_MIDDLE_vfold0
WARNING - root - Changed type of config entry "min_fg_data" from str to int
INFO - mySSL - Running command 'main'
INFO - mySSL - Started run with ID "12"
INFO - main - ###### Create model ######
NETWORK: Using ms-coco initialization
INFO - main - ###### Load data ######
INFO - main - ###### Labels excluded in training : [2, 3] ######
INFO - main - ###### Unseen labels evaluated in testing: [2, 3] ######
Dataset: the following classes has been excluded [2, 3]######
DEBUG_DATASET CT_STATS NORMALIZED MEAN 0.2503295103060257 STD 0.22380904778564215
Initial scans loaded:
['7', '8', '9', '10', '11', '12']
INFO - main - ###### Set optimizer ######
INFO - main - ###### Training ######
INFO - main - ###### This is epoch 0 of 100 epoches ######
/home/wangyixiong/anaconda3/envs/self/lib/python3.7/site-packages/torch/nn/functional.py:2973: UserWarning: Default upsampling behavior when mode=bilinear is changed to align_corners=False since 0.4.0. Please specify align_corners=True if the old behavior is desired. See the documentation of nn.Upsample for details.
"See the documentation of nn.Upsample for details.".format(mode))
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
Faulty batch detected, skip
^CWARNING - mySSL - Aborted after 0:00:20!
Traceback (most recent call last):
File "training.py", line 30, in
@ex.automain
File "/home/wangyixiong/anaconda3/envs/self/lib/python3.7/site-packages/sacred/experiment.py", line 140, in automain
self.run_commandline()
File "/home/wangyixiong/anaconda3/envs/self/lib/python3.7/site-packages/sacred/experiment.py", line 263, in run_commandline
return self.run(cmd_name, config_updates, nam
I am also getting "Faulty batch detected, skip". How to fix this?
|
2025-04-01T06:38:11.169365
| 2018-05-29T19:02:19
|
327447688
|
{
"authors": [
"Ellmen",
"ManuKle",
"aguynamedben",
"amilajack",
"vikr01"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4624",
"repo": "chentsulin/electron-react-boilerplate",
"url": "https://github.com/chentsulin/electron-react-boilerplate/issues/1573"
}
|
gharchive/issue
|
Too many config files
Including configs for things like flow, vscode, and travis adds a lot of mess to clean up for people who don't use those things. Since this is the react electron repo, I feel like it should be stripped down a little.
Which configs should we get rid of?
Just registering another opinion... I think the flow configs are good because they provide an opinionated best practice, which is what boilerplate is for.
I think the editor-specific configs could either be removed (because everybody uses a different editor, and your choise editor isn't necessarily related to the project ecosystem), or it should be an explicit policy that other editor configs are welcomed and maybe put them in editor_configs/ with a README file in that directory explaining the policy. I use Emacs, but VS Code seems to be an emerging norm in the JS community so it's probably helpful to have a config laying around.
Also, wouldn't removing yarn.lock make it pretty difficult to debug issues people are having with the project? yarn.lock in projects I've worked on is always committed to make sure that people are using the same dependencies without having to continually ask everybody to list which version they have installed locally.
In general my opinion is:
It's not that hard to remove config files, and they rarely get in the way. You're going to have to customize this repo for your project anyways. git rming or editing a few config files you don't find helpful is much easier than being a newbie and not having access to sensible config files. For example, being new to the whole ecosystem it would be very hard to create a Flow config from scratch.
The purpose of boilerplate is to help people get going in an ecosystem quickly.
Example config files, even if they're not setup out of the box the way you'd like, are very helpful as it's difficult to come up with from scratch when you're new to an ecosystem.
Config files are a good opportunity to set opinionated best practices, which is especially useful in boilerplate projects. These best practices reflect norms in a community.
Boilerplate seems like the BEST place for config files. The cost of removing/modifying config to fit your preferences is much lower than the cost of not having ready access to example configs.
I like the idea of an editor_configs and I can appreciate that having sample code makes it easier for a newcomer to get started.
The main yarn.lock is certainly important - the one I was talking about is in the app directory and is empty (there are no dependencies in that directory's package.json)
I would argue that Flow isn't actually a best practice and integrating it so deeply makes this project less flexible. Skype, GitHub, and Slack are all using TypeScript. The problem with enforcing a certain stack is that when any part of it becomes obsolete, the whole project does.
It kind of depends on if this project is supposed to be a pure electron-react boilerplate or if it's a "cool stack" that runs on electron and react. Personally, I was just looking for a clean way to structure an application that uses the two technologies.
Cool, those opinions make sense too.
I'm just curious for my own education, are you using this repo with TypeScript? I've just been using Flow since it was already in this project... but I'm not partial to it. Was it easy to get TypeScript working with this?
Whether this project is best as "cool stack" or "basic stack", I could see a change from Flow to TypeScript being a reasonable opinionated change. Coming from non-JavaScript land, there are already a lot of other opinions built-in (Webpack, ES6, testing framework, Redux).
I'd like to see "cool stack". I'm a Rails fan and have seen the "cool stack" decision go the undesirable way (i.e. CoffeeScript), but overall I like things to be opinionated and bundled (Rails for Ruby) vs. piece-together-your-own-together (i.e. Flask for Python). Just opinions though. Good discussion.
You can find my typescript fork under notable forks in the wiki. It's based on version 13.3 and I'll try to update soon with the latest commits.
Regarding the configs, I kinda like the idea of putting them in a separate folder. But I wouldn't delete them, as they are useful to build upon.
@ManuKle very cool!
In #1686, we moved webpack configs to a configs folder. That drops quite a few files from the home directory.
|
2025-04-01T06:38:11.171965
| 2021-10-21T22:15:19
|
1032985260
|
{
"authors": [
"cheqianh"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4625",
"repo": "cheqianh/ion-hive-serde",
"url": "https://github.com/cheqianh/ion-hive-serde/pull/2"
}
|
gharchive/pull-request
|
Adds support to Hadoop compression codec.
Issue #, if available:
Description of changes:
By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Sure I think it's a good idea to add more comments.
It's also a good opportunity to look into other serde package and figure out how/when the APIs are called.
Discussed offline, Added some comments. tested locally it worked well for gzip.
Added links to the source. I'll do an overall testing again before sending this PR to the official repo.
|
2025-04-01T06:38:11.177960
| 2018-07-02T22:33:35
|
337685924
|
{
"authors": [
"KFlash",
"g-plane"
],
"license": "isc",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4627",
"repo": "cherow/cherow",
"url": "https://github.com/cherow/cherow/issues/179"
}
|
gharchive/issue
|
Don't create PR if you can't finish them
To avoid too much noise on this repo, don't open PR's if you can't or will finish them. At the moment we have 5 open PR. One of this PR have been inactive for the last 15 days - https://github.com/cherow/cherow/pull/153
We also have two PR - closed now - for string literals and ecmaVersion option. This also make up some noise on the repo.
Just be 100% sure before you open a PR. Also be sure that you get reviewers with knowledge enough to understand your way of thinking and the code flow.
https://github.com/cherow/cherow/pull/175 is my last PR. I will from now on push directly due to lack of reviewers and too keep the repo noise down.
cc / @fkleuver @g-plane @Aladdin-ADD
I really don't recommend to push directly.
I have pushed directly from start. As long as I'm only dealing with src folder there is no issue.
I still recommend to open PR. Once you opened a PR and there are no any reviews in several days, you can merge by yourself.
The issue is that somneone open PR and 1) never finish them or 2) never merge them
|
2025-04-01T06:38:11.198174
| 2024-10-24T15:35:21
|
2611956382
|
{
"authors": [
"Poukpalaova",
"gonu2024",
"nhathoangfoto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4628",
"repo": "chflame163/ComfyUI_LayerStyle",
"url": "https://github.com/chflame163/ComfyUI_LayerStyle/issues/372"
}
|
gharchive/issue
|
Unable to pick colors in the ColorPicker node.
I can't select any color other than the default color when loading the node. Normally, when I click the mouse cursor on the color, a color palette pops up, and I just need to move the cursor to the image and pick the color. But today, when I click on the color, nothing happens.
I'm having the same problem. Did you solve it? Can someone help me?
The panel does not appear even though my browser settings do not disable pop-ups.
I have the same problem. In an new install, the color pop screen work flawless. After installing custom nodes, it stop working. It's probably an incompatibility with one of them, but wich one...?
|
2025-04-01T06:38:11.199772
| 2021-07-03T00:46:09
|
936132921
|
{
"authors": [
"chhoumann",
"firinael"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4629",
"repo": "chhoumann/quickadd",
"url": "https://github.com/chhoumann/quickadd/issues/45"
}
|
gharchive/issue
|
cancelling a template add brings up an error
Hi, just started trying out the plugin, and ran into a curious behaviour.
Whenever I press ESC after selecting a template to add, I get this error:
And this only happens with the add template function, as far as I've noticed.
Thank you! Fixed in the latest version. :)
|
2025-04-01T06:38:11.201528
| 2015-07-06T18:05:13
|
93338824
|
{
"authors": [
"chieffancypants",
"ittayd"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4630",
"repo": "chieffancypants/angular-hotkeys",
"url": "https://github.com/chieffancypants/angular-hotkeys/issues/160"
}
|
gharchive/issue
|
allow to provide a separate key combo for documentation
it seems that for modifier+key combinations, the event triggers with the english key, even if the layout is in another language (at least for me, using english+hebrew layouts). this seems inherent in javascript. while the best solution is for the framework to map the non english characters to their english equivalent when handling key presses, it seems this is tricky (if possible). an alternative is to just allow to pass a key combination that will be used in the documentation. so while binding to alt+t, the documentation will show alt+א
I'm not in control of this, as all hotkey-specific binding is handled by mousetrap. You may want to open a ticket there
I was suggesting that an optional parameter will be allowed and that will be used for documentation. So I call with both alt+t and alt+א, you pass alt+t to mousetrap and show alt+א in the documentation ('?')
|
2025-04-01T06:38:11.236878
| 2023-02-27T10:50:50
|
1600933116
|
{
"authors": [
"StefanoBettelli",
"Thomasb81",
"alaindargelas"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4632",
"repo": "chipsalliance/Surelog",
"url": "https://github.com/chipsalliance/Surelog/issues/3487"
}
|
gharchive/issue
|
Removing unused modules from UHDM file + dependencies
I translate SystemVerilog files to UHDM with surelog.
Let us say that I want to create a design with top module MyModule:
>> /usr/local/bin/surelog --version
VERSION: 1.47
BUILT : Jan 24 2023
>> /usr/local/bin/surelog -parse -synth -timescale=1ns/1ps
-I$(EDIR) -I$(IDIR) +libext+.sv -y $(MDIR) -top MyModule MyModule.sv
In this call, EDIR is a directory with .sv files containing enumerations, IDIR is a directory with .sv files containing interfaces, and MDIR is a directory containing a large collection of SystemVerilog modules in .sv files.
After analysing the .uhdm file it is apparent that it contains all modules in the MDIR directory, whether or not they are instantiated through the top module.
Is it possible to call surelog in such a way that modules which are not used, directly or indirectly, by the top module are removed from the generated .uhdm file? If not, it would be a good idea to include such an option.
Is it possible to call surelog in such a way that it generates a dependency file listing all .sv files used, directly or indirectly by the top module, to create the design? This would be very useful in Makefile flows.
Hello
As far as I understand, Surelog is a tool that will allow you to fill a uhdm database. What you do with the uhdm database is your custom application.
Surelog is able to perform some elaboration. One elaboration step consist to complete the uhdm database with what you want.
using -d uhdm, the tool will dump the complete uhdm tree.
You should be able to observe a difference by adding -noelab,
So to answer your question, you probably have to write your own application to iterate on only the part of the uhdm data base that interest you.
See:
https://github.com/chipsalliance/Surelog/blob/5fc1993668337fa646adbf1e5e0bd139f2304293/src/hellouhdm.cpp#L103
vs
https://github.com/chipsalliance/Surelog/blob/5fc1993668337fa646adbf1e5e0bd139f2304293/src/hellouhdm.cpp#L150
@StefanoBettelli, can we close this issue? Are you OK with the file suggested above?
Unfortunately I haven't had much time recently to look at it again. However, it seems that this project can hardly be used beyond linting, since there isn't enough documentation, and it is not clear the level of elaboration that is / can be done. There are projects to offload the SystemVerilog fronted of verilator and yosys to Surelog, but so far my efforts at using them consistently in a Makefile flow have failed, due to several shortcomings (not all in Surelog).
For the time being the issue can be closed. I will come back to it if there is any substantial progress.
@StefanoBettelli you really mean a "Makefile" flow or a "CMake" flow?
If you really mean "Makefile" like you have a larger project that uses Makefile and you want to shoehorn Surelog CMake into it,
then it is best to invoke the cmake from the Makefile like in:
https://github.com/chipsalliance/Surelog/blob/efd45b9291cb4aa03c54e9d3d72d90b885571971/Makefile#L23 and not try to convert any of the CMake steps to Makefile
As for the API, the API is the Standard VPI interface (very well documented in the SystemVerilog standard):
Accessed using: https://github.com/chipsalliance/UHDM
Extract of the Standard API for reference: https://github.com/chipsalliance/Surelog/blob/master/third_party/Verilog_Object_Model.pdf
|
2025-04-01T06:38:11.262001
| 2015-10-08T17:20:41
|
110502166
|
{
"authors": [
"fnky",
"gaui"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4633",
"repo": "chjj/marked",
"url": "https://github.com/chjj/marked/issues/667"
}
|
gharchive/issue
|
Different DOM structure for inline/fenced code blocks
With GFM, both inline code blocks and fenced code blocks provide the following DOM structure:
<p><code>...</code></p>
Inline code block:
`code`
Fenced codeblock:
```
code
```
I would like to style the inline code block differently with CSS than the fenced code block.
How can I do this? There's no way to target the inline code block.
What options are you using? It should produce
<pre><code>...</code><pre>
for fenced/block codes, and
<p><code>...</code>></p>
for inline code. For styling you can select code for inline and pre code for code blocks.
It produces in both cases <p><code>... </code></p>
I'm using default options.
Can you give an example of your code and markdown?
Here is a screenshot. I'm using marked() without any options so it's using default options. As you can see there is no difference between inline code block and a fenced code block.
This is actually correct behaviour in GFM. It's required to have a newline (\n) character after start fence (except you can provide a language extension after start fence)
```\n
Fenced code block\n
becomes
Fenced code block
and
Fenced code block
becomes
Fenced code block
```
The reason is that inline code uses single back ticks so further unescaped back ticks will be ignored in the output.
@fnky Thank you very much. Problem solved. :)
|
2025-04-01T06:38:11.264648
| 2022-05-29T14:04:46
|
1251898611
|
{
"authors": [
"chkilel",
"khalidchawtany"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4634",
"repo": "chkilel/vitewind-theme",
"url": "https://github.com/chkilel/vitewind-theme/issues/2"
}
|
gharchive/issue
|
Does this theme support HRM?
Thanks for this awesome theme.
Does this theme support HRM or it does full reload on every change?
I always get full reloads making changes to my pages.
Hi @khalidchawtany
At this time it does not support HMR, I did't find a way to implement it.
Thanks Khalid
HRM works for JS and CSS files, but if changes are made to the template files the server needs to reload the page to reflect those changes.
Hope this clarify your question @khalidchawtany.
Thanks that is very nice. I wish it worked for templates too. However, I know that it is almost impossible to implement :(
|
2025-04-01T06:38:11.267766
| 2017-04-09T03:55:04
|
220445653
|
{
"authors": [
"coveralls",
"webdevan"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4635",
"repo": "chmln/flatpickr",
"url": "https://github.com/chmln/flatpickr/pull/743"
}
|
gharchive/pull-request
|
dont append confirmDate to native inputs
fixes #742
Coverage remained the same at 87.67% when pulling ae2da50a5d0509b42a89fdf0dec14e3719ba2e03 on webdevan:master into 1f1d629c4708a1e06422dcdbba5acbe0e1f7804e on chmln:master.
|
2025-04-01T06:38:11.280282
| 2018-04-19T09:54:33
|
315806877
|
{
"authors": [
"chobits"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4636",
"repo": "chobits/ngx_http_proxy_connect_module",
"url": "https://github.com/chobits/ngx_http_proxy_connect_module/pull/40"
}
|
gharchive/pull-request
|
support nginx-1.14.x
try to fix https://github.com/chobits/ngx_http_proxy_connect_module/issues/12
Tested-by: Xiaochen Wang xiaochen
Date: Thu Apr 19 17:57:07 CST 2018
hostname: Xiaochens-MacBook-Pro.local
pwd: /Users/xiaochen/work/github/ngx_http_proxy_connect_module
cases: t/http_proxy_connect.t
git log:
case result:
$ MOD_DYCONF_PATH=/Users/xiaochen/work/github/ngx_http_proxy_connect_module TEST_NGINX_BINARY=/Users/xiaochen/work/github/nginx-1.14.0/objs/nginx prove -v -I /Users/xiaochen/work/github/nginx-tests/lib t/http_proxy_connect.t
t/http_proxy_connect.t ..
DNS server: try to bind server port: 18085
DNS server: daemon pid: 41072
DNS server: working
ok 1 - 200 Connection Established
ok 2 - 200 Connection Established server name
ok 3 - 200 Connection Established server name
ok 4 - 200 Connection Established server name
ok 5 - 200 Connection Established not allowed port
ok 6 - Get method: proxy_pass
ok 7 - Get method: return 200
ok 8 - set remote address
ok 9 - set local address and remote address
ok 10 - $connect_host, $connect_port, $connect_addr
ok 11 - dns resolver fail
ok 12 - skip proxy connect module without rewrite phase enabled
ok 13 - skip proxy connect module without rewrite phase enabled: if/return
DNS server: stop
1..13
ok 14 - no alerts
ok 15 - no sanitizer errors
All 13 subtests passed
Test Summary Report
-------------------
t/http_proxy_connect.t (Wstat: 0 Tests: 15 Failed: 2)
Failed tests: 14-15
Parse errors: Plan (1..13) must be at the beginning or end of the TAP output
Bad plan. You planned 13 tests but ran 15.
Files=1, Tests=15, 0 wallclock secs ( 0.02 usr 0.01 sys + 0.11 cusr 0.06 csys = 0.20 CPU)
Result: FAIL
Report-generated-by: http://wxc.oss.aliyuncs.com/dtest.sh
|
2025-04-01T06:38:11.287464
| 2018-09-24T19:09:32
|
363281546
|
{
"authors": [
"AppVeyorBot",
"pascalberger",
"regexaurus"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4637",
"repo": "chocolatey/chocolatey-coreteampackages",
"url": "https://github.com/chocolatey/chocolatey-coreteampackages/pull/1115"
}
|
gharchive/pull-request
|
(Skype) Changed InnoSetup /SILENT parameter to /VERYSILENT
Description
Using /VERYSILENT instead of /SILENT makes Inno Setup a bit quieter.
Motivation and Context
Users are understandably concerned about "mystery" install windows.
How Has this Been Tested?
Manually downloaded https://endpoint920510.azureedge.net/s4l/s4l/download/win/Skype-<IP_ADDRESS>.exe, and ran installer with same parameters as in chocolateyInstall.ps1, except /SILENT was changed to /VERYSILENT. Installation succeeded and no installation windows appeared. Skype still (re)launches after installation.
Types of changes
Minor (largely cosmetic) change
Checklist:
[x ] My code follows the code style of this repository.
[ ] My change requires a change to documentation (this usually means the notes in the description of a package).
[ ] I have updated the documentation accordingly (this usually means the notes in the description of a package).
[x ] All files are up to date with the latest Contributing Guidelines
[ ] The added/modified package passed install/uninstall in the chocolatey test environment.
[x ] The changes only affect a single package (not including meta package).
:white_check_mark: Package verification completed without issues. PR is now pending human review
@regexaurus your changes have been merged, thanks for your contribution 👍
|
2025-04-01T06:38:11.293465
| 2017-11-04T01:05:16
|
271154834
|
{
"authors": [
"AdmiringWorm",
"bcurran3",
"dimqua",
"rogersachan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4638",
"repo": "chocolatey/chocolatey-package-requests",
"url": "https://github.com/chocolatey/chocolatey-package-requests/issues/200"
}
|
gharchive/issue
|
RFP - Croscore fonts
The Chrome OS core fonts, also known as the Croscore fonts, are a collection of three TrueType font families: Arimo (sans-serif), Tinos (serif) and Cousine (monospace). These fonts are metrically compatible with Arial, Times New Roman, and Courier New, the most commonly used fonts on Microsoft Windows operating system, for which they are intended as open-source substitutes.
More info: https://en.wikipedia.org/wiki/Croscore_fonts
Download link: https://gsdview.appspot.com/chromeos-localmirror/distfiles/
Please also supply the home page for these fonts, before we can make this available for maintainers.
Homepage: https://fonts.google.com/
More specificaly: Amiro, Tinos and Cousine.
I MIGHT look into this.
Looks like this is the latest version right? https://gsdview.appspot.com/chromeos-localmirror/distfiles/croscorefonts-1.31.0.tar.bz2
I completely forgot about this.
Done!
https://chocolatey.org/packages/croscorefonts-font
@rogersachan, since you seem appreciative...
https://chocolatey.org/packages/crosextrafonts-caladea-font/2013.02.14
https://chocolatey.org/packages/crosextrafonts-carlito-font/2013.09.20
Thanks so much for this!
|
2025-04-01T06:38:11.310309
| 2015-01-25T05:46:47
|
55399369
|
{
"authors": [
"ahmedmohiduet",
"andyli",
"ferventcoder"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4639",
"repo": "chocolatey/chocolatey",
"url": "https://github.com/chocolatey/chocolatey/issues/666"
}
|
gharchive/issue
|
How can I get installed programs instantly in current cmd window?
Every time I install a program I had to restarted my cmd window. Is there any convenient ways to get them in current cmd window? Simething like 'source file.sh' in unix ?
P.S. - I am using Windows 7
If you are in cmd.exe and not powershell you can type refreshenv
Chocolatey comes with a batch file named the above and will query the registry and reset all of your environment variables.
Perhaps we need more documentation surrounding that.
Is it possible to get RefreshEnv to work in powershell too?
If it cannot, there should be a warning message if RefreshEnv is used in powershell instead of a silent fail.
It's an issue with Windows inability to refresh shells. It doesn't have
anything to do with choco itself. Now that that is it off the way,
refreshenv is a stop gap for choco to keep you from having to close and
reopen your cmd window. Its also mostly undocumented. The recommendation is
that you close and reopen your shell.
You can also do PowerShell if you call &
c:\programdata\chocolatey\chocolateyinstall\helpers\Update-Session.ps1 (the
path may not be exactly right but you get the idea)- it may require
importing the module first and then calling update-session.
On Tuesday, January 27, 2015, Andy Li<EMAIL_ADDRESS>wrote:
Is it possible to get RefreshEnv to work in powershell too?
If it cannot, there should be a warning message if RefreshEnv is used in
powershell instead of a silent fail.
—
Reply to this email directly or view it on GitHub
https://github.com/chocolatey/chocolatey/issues/666#issuecomment-71639225
.
--
Rob
"Be passionate in all you do"
http://devlicio.us/blogs/rob_reynolds
http://ferventcoder.com
http://twitter.com/ferventcoder
Do you mean Update-SessionEnvironment? I tried adding it to the end of chocolateyInstall.ps1, but it seems to have no effect - the path var is not updated in powershell.
It actually does work but you need to call it separately outside of a choco
call (which starts up and closes a separate posh process).
On Tuesday, January 27, 2015, Andy Li<EMAIL_ADDRESS>wrote:
Do you mean Update-SessionEnvironment
https://github.com/chocolatey/chocolatey/wiki/HelpersUpdateSessionEnvironment?
I tried adding it to the end of chocolateyInstall.ps1, but it seems to
have no effect - the path var is not updated in powershell.
—
Reply to this email directly or view it on GitHub
https://github.com/chocolatey/chocolatey/issues/666#issuecomment-71654814
.
--
Rob
"Be passionate in all you do"
http://devlicio.us/blogs/rob_reynolds
http://ferventcoder.com
http://twitter.com/ferventcoder
Umm.. I thought the helpers are meant to be called inside chocolateyInstall.ps1...
Yes, they are. But anything they do with respect to environment vars only
lasts during that choco run. Once it comes back to the shell, all child
process changes are gone. It would be an egregious security issue with
Windows if child processes did more than they do to the parent process.
I'm giving you an option for updating your environment variables. Also
undocumented and not supported. Ymmv. Remember, you can just close and
reopen your shell as well.
On Tuesday, January 27, 2015, Andy Li<EMAIL_ADDRESS>wrote:
Umm.. I thought the helpers are meant to be called inside
chocolateyInstall.ps1...
—
Reply to this email directly or view it on GitHub
https://github.com/chocolatey/chocolatey/issues/666#issuecomment-71655906
.
--
Rob
"Be passionate in all you do"
http://devlicio.us/blogs/rob_reynolds
http://ferventcoder.com
http://twitter.com/ferventcoder
Hi Rob,
I was just able to reproduce it my win 7. First I run "choco install java". It says not found.Running "choco install jdk" also fails as you can see in the screen and after that new blank window appears it just freezes nothing happens after that.
@ahmedmohiduet does this fail if you switch over to 0.9.9?
I guess I will have to check it again. Will let you know after successful reproduce :)
|
2025-04-01T06:38:11.313141
| 2016-01-29T02:49:00
|
129649359
|
{
"authors": [
"jchodera"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4640",
"repo": "choderalab/openmoltools",
"url": "https://github.com/choderalab/openmoltools/pull/185"
}
|
gharchive/pull-request
|
[WIP] Compare energy components with antechamber/prmtop route
This is the beginning of a PR to compare energy components with a prmtop route for loading parameters.
Currently, the test fails because the imatinib prmtop seems to lead to very different nonbonded forces:
Energy components:
NonbondedForce -203.736379 31.680160 : 235.416539
HarmonicBondForce 6.160988 6.248384 : 0.087396
HarmonicAngleForce 5.330822 5.329843 : 0.000980
PeriodicTorsionForce 13.253722 8.783295 : 4.470427
CMMotionRemover 0.000000 0.000000 : 0.000000
Maximum allowed deviation (0.600000) exceeded.
I'm not sure why this is, but I am investigating. I also don't know the origin of the chemicals/imatinib/imatinib.prmtop file.
It may be better for me to compare with a prmtop/inpcrd generated on the fly from either the same procedure I am using for generating the residue templates or the intermediate GAFF mol2/frcmod files used in generating the residue templates.
The error is now much smaller:
Energy components:
NonbondedForce -203.736307 -206.509692 : 2.773386
HarmonicBondForce 6.160980 6.248380 : 0.087400
HarmonicAngleForce 5.330831 5.329848 : 0.000983
PeriodicTorsionForce 13.253717 8.783295 : 4.470422
CMMotionRemover 0.000000 0.000000 : 0.000000
Maximum allowed deviation (0.600000) exceeded.
Will need to do more debugging but raised the max allowed deviation to 5 for now.
|
2025-04-01T06:38:11.316719
| 2018-07-09T23:43:18
|
339644403
|
{
"authors": [
"jchodera",
"pgrinaway"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4641",
"repo": "choderalab/perses",
"url": "https://github.com/choderalab/perses/issues/476"
}
|
gharchive/issue
|
Nonequilibrium switching fails with new OpenMM API
I think this is actually an update that needs to be made in openmmtools. The AlchemicalState object is unaware of how to set the global parameter lambda_electrostatics if it's covered by the regular NonbondedForce.
cc: https://github.com/choderalab/openmmtools/issues/359
@pgrinaway This should be resolved using the latest OpenMM 7.3.0 dev build now that https://github.com/pandegroup/openmm/pull/2119 and https://github.com/pandegroup/openmm/pull/2115 have been merged.
OK to close?
@brycestx points out we need to revise things so that we explicitly add the global context parameters to NonbondedForce now that the API has been extended.
Fixed in the OpenMM 7.3.1 bugfix release.
|
2025-04-01T06:38:11.323447
| 2023-08-30T22:13:27
|
1874447426
|
{
"authors": [
"choogiesaur",
"digikitty21"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4642",
"repo": "choogiesaur/hgss-shiny-modifier",
"url": "https://github.com/choogiesaur/hgss-shiny-modifier/pull/5"
}
|
gharchive/pull-request
|
Mark arm9 as Decompressed, and Change Shiny Value Offset Output to Hex
I'm not sure if I did this right, but hopefully this should only include:
Explicitly marking the arm9 as decompressed
Outputting the Shiny Value offset in hex
Thanks for these changes @digikitty21 ! Merged to main.
|
2025-04-01T06:38:11.330829
| 2020-10-21T15:31:27
|
726601976
|
{
"authors": [
"bastelfreak",
"ripienaar"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4643",
"repo": "choria-io/puppet-mcollective",
"url": "https://github.com/choria-io/puppet-mcollective/issues/263"
}
|
gharchive/issue
|
Harden Puppet datatypes
Some of the datatypes are not as strict as they could be. To make the usage safer, the should be hardened.
thank you!
|
2025-04-01T06:38:11.365713
| 2024-11-20T03:11:31
|
2674200253
|
{
"authors": [
"chrisbenincasa",
"pavlov70"
],
"license": "Zlib",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4644",
"repo": "chrisbenincasa/tunarr",
"url": "https://github.com/chrisbenincasa/tunarr/issues/974"
}
|
gharchive/issue
|
minor issue with new ffmpeg stack on .16
Hi,
After upgrading to .16 and trying out the new ffmpeg stack, it seems to work quite well from initial impressions, perhaps loading a bit faster. The one issue I've encountered is when playing back via Kodi's iptv plugin. It would immediately fail without opening the stream. The log showed:
CCurlFile::Open - http://<IP_ADDRESS>:8000/stream/channels/1fa17176-a81c-495c-a631-67a375320412.ts Failed with code 500:
Pasting the url into vlc also resulted in failed playback. By changing the .ts to .m3u8 I was able to play back in vlc successfully, so I edited the channels.m3u and changed each .ts instance to .m3u8. This restored my kodi iptv functionality.
I should note that the above is not necessary using the old ffmpeg stack. Thanks again as always!
Any relevant logs on the server side?
Also can you provide a few more details? Like are you using hw accel? If so, which kind?
I was able to reproduce this. Will have a fix today. Thanks for testing!
Sorry I didn't get you the logs earlier, life interrupted me. Thanks for identifying the cause and I'm looking forward to the fix.
Been using the docker vaapi and load times in Kodi are in the ~1-2 second range with the new ffmpeg changes (down a second or so). Every bit as fast as traditional linear tv!
|
2025-04-01T06:38:11.370523
| 2015-09-03T23:10:09
|
104799176
|
{
"authors": [
"chrisccerami",
"felipebalbi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4645",
"repo": "chrisccerami/ascii_art",
"url": "https://github.com/chrisccerami/ascii_art/issues/4"
}
|
gharchive/issue
|
Curses has been removed from Standard Library
Please add runtime dependency to curse ~> 1.0 if Ruby is > 2.1
Thanks for opening the issue. I'll take care of this as soon as I get a chance.
This has been addressed. Thanks again for pointing it out @felipebalbi
Hey, no problem. I'll have a look at how you did it. I knew about adding it to gemspec, but didn't know how to add it conditionally on ruby version :-)
|
2025-04-01T06:38:11.379150
| 2017-11-28T19:43:10
|
277513634
|
{
"authors": [
"CloudChoSony",
"SJTUPanda",
"ShubhayanS",
"speed8928"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4646",
"repo": "chrischoy/3D-R2N2",
"url": "https://github.com/chrischoy/3D-R2N2/issues/22"
}
|
gharchive/issue
|
ValueError: You are tring to use the old GPU back-end.
Hi I use windows 10 to implement the project and it turns out the error below:
ValueError: You are tring to use the old GPU back-end. It was removed from Theano. Use device=cuda*
I tried to follow your link that from error
https://github.com/Theano/Theano/wiki/Converting-to-the-new-gpu-back-end(gpuarray)
and install theano pygpu, but nothing good come out.
I also tried change the Theano variable device = gpu to device = cuda0 and it gives me the error below:
(py3) C:\Users\speed\source\3D-R2N2>python demo.py prediction.obj
WARNING (theano.tensor.blas): Using NumPy C-API based implementation for BLAS functions.
ERROR (theano.gpuarray): pygpu was configured but could not be imported or is too old (version 0.7 or higher required)
NoneType
Downloading a pretrained model
Traceback (most recent call last):
File "demo.py", line 84, in
main()
File "demo.py", line 55, in main
download_model(DEFAULT_WEIGHTS)
File "demo.py", line 34, in download_model
'--create-dirs', '-o', fn])
File "C:\Users\speed\AppData\Local\Programs\Python\Python35\Lib\subprocess.py", line 557, in call
with Popen(*popenargs, **kwargs) as p:
File "C:\Users\speed\AppData\Local\Programs\Python\Python35\Lib\subprocess.py", line 947, in init
restore_signals, start_new_session)
File "C:\Users\speed\AppData\Local\Programs\Python\Python35\Lib\subprocess.py", line 1224, in _execute_child
startupinfo)
FileNotFoundError: [WinError 2] The system cannot find the file specified
Could you please advise me the way to compile it, thanks.
Hi, I have the same error with you. Have you solved this problem?
I lowered the Theano version 7.x then it worked.
Thanks!
I use the Theano version 0.8.0 then it works.
I found another way to solve at ValueError «You are trying to use the old GPU back-end» when importing keras
According to cgl, please type in terminal "export THEANO_FLAGS=mode=FAST_RUN,device=cuda,floatX=float32"
Thanks!
I use the Theano version 0.8.0 then it works.
I am getting - AttributeError: ('This name is already taken', 'floatX')
|
2025-04-01T06:38:11.383224
| 2024-06-22T05:44:41
|
2367588031
|
{
"authors": [
"Sponge-bink",
"chrisgrieser"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4647",
"repo": "chrisgrieser/alfred-wikipedia-suggest",
"url": "https://github.com/chrisgrieser/alfred-wikipedia-suggest/pull/1"
}
|
gharchive/pull-request
|
feat: configurable number of results. multiple languages.
I left the version number unchanged. You might wanna update that later. ;)
Thx! Yeah, updating the version number is sth that I do with btw releases. Will do one later today
Release as 1.1.0: https://github.com/chrisgrieser/alfred-wikipedia-suggest/releases/download/1.1.0/alfred-wikipedia-suggest.alfredworkflow
Just so you know, I made some minor changes, like using , as delimiter instead of /, as that's what most workflows do, so you might have go over your settings briefly.
|
2025-04-01T06:38:11.394784
| 2015-07-20T06:57:07
|
95999913
|
{
"authors": [
"elod91"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4648",
"repo": "chrisjenx/Calligraphy",
"url": "https://github.com/chrisjenx/Calligraphy/issues/192"
}
|
gharchive/issue
|
Diacritical problems on older devices
Hi!
I've been using you library for a while now without any problems, but now I have a project in which, some of the texts have diacritics. I don't know the exact android version on which it runs OK and from which it has problems, but it seems like pre-L devices don't show the diacritical characters with the custom font.
Sorry, seems like the font I was using had problems. Changed it to another and it works
|
2025-04-01T06:38:11.408258
| 2022-01-20T02:42:29
|
1108770310
|
{
"authors": [
"chrissy-dev",
"jshwlkr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4649",
"repo": "chrissy-dev/eleventy-web-starter",
"url": "https://github.com/chrissy-dev/eleventy-web-starter/issues/33"
}
|
gharchive/issue
|
I don't think you need Luxon
Just an FYI. I think you can use https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Intl/DateTimeFormat.
@jshwlkr I've decided to keep it for now. Eleventy uses Luxon internally so I'm happy to keep the tooling the same.
|
2025-04-01T06:38:11.412838
| 2015-09-26T22:57:07
|
108501121
|
{
"authors": [
"christianalfoni",
"danpantry",
"tomatau"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4650",
"repo": "christianalfoni/formsy-react",
"url": "https://github.com/christianalfoni/formsy-react/pull/221"
}
|
gharchive/pull-request
|
Transpile with Babel prior to publishing to npm
This enables Node environments to use require('formsy-react') instead of
having to transpile it themselves
Fixes #220
Should not really have any affect on callers - transpiled code is functionally equivalent
FYI I wrote this on react-0.14 because that's the branch I am using. This should apply without issues to master, too.
:+1: please merge this
If anyone wants to use this now - I've forked your fork with the ./lib built and checked in. https://github.com/tomatau/formsy-react/tree/react-0.14
you can:
"dependencies": {
"formsy-react": "tomatau/formsy-react#react-0.14"
Thanks @tomatau
Ideally we'd configure the .npmignore so it doesn't include ./src files, but only ./lib files - that way people who download the package don't have extraneous files
Precisely! But we don't want to create an npm module copy of formsy-react just whilst waiting for this merge :)
Great, thanks!
New version will be released this weekend, with latest version of React 0.14
@christianalfoni :beers: :+1:
|
2025-04-01T06:38:11.414204
| 2021-03-16T20:54:47
|
833188777
|
{
"authors": [
"binhex",
"christianhaitian"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4651",
"repo": "christianhaitian/arkos",
"url": "https://github.com/christianhaitian/arkos/issues/295"
}
|
gharchive/issue
|
[Wiki] Link to download v1.6 on the Wiki please.
Please can you update the Wiki with a link to v1.6 for the Anbernic 351m, many thanks for your hard work!.
I will eventually but it will take some time. Until then, you can download the current image and update online to it. Stay tuned.
OK that makes sense, i appreciate the response.
|
2025-04-01T06:38:11.428449
| 2015-11-18T13:25:33
|
117586322
|
{
"authors": [
"monsterxxx",
"rguerreiro"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4652",
"repo": "christopherthielen/ui-router-extras",
"url": "https://github.com/christopherthielen/ui-router-extras/issues/273"
}
|
gharchive/issue
|
Sticky state doesn't work when applied to a state with params
Hi, sticky state doesn't seem to work when it has parameters.
Example:
.state('main.products', {
url: '/products/:product_id',
sticky: true
})
.state('main.products.details', {
url: '/details'
})
When you navigate between these 'main.product.details ' states while changing product_id, you will get an error.
Plunk: http://plnkr.co/edit/czXStZZCQ9wDZ4mVqtA2?p=info
Step by step log:
Go to products/1/details
Current transition: main: {}: -> main.products.details: {"product_id":1}
Before transition, inactives are: : []
After transition, inactives will be: []
Transition will exit: ["(main)"]
Transition will enter: ["(main)", "ENTER: main.products", "ENTER: main.products.details"]
SurrogateFromPath: ["main"]
SurrogateToPath: ["main", "main.products", "main.products.details"]
Current state: main.products.details, inactive states: []
Go to products/2/details
Current transition: main.products.details: {"product_id":"1"}: -> main.products.details: {"product_id":2}
Before transition, inactives are: : []
After transition, inactives will be: ["main.products", "main.products.details"]
Transition will exit: ["(main)", "INACTIVATE: main.products", "INACTIVATE: main.products.details"]
Transition will enter: ["(main)", "ENTER: main.products", "ENTER: main.products.details"]
SurrogateFromPath: ["main", "inactivate:main.products", "inactivate:main.products.details"]
SurrogateToPath: ["main", "main.products", "main.products.details"]
Current state: main.products.details, inactive states: ["main.products.details", "main.products"]
Go to products/3/details and get an error
Current transition: main.products.details: {"product_id":"2"}: -> main.products.details: {"product_id":3}
Before transition, inactives are: : ["main.products.details", "main.products"]
After transition, inactives will be: ["main.products", "main.products.details"]
Transition will exit: ["(main)", "INACTIVATE: main.products", "INACTIVATE: main.products.details"]
Transition will enter: ["(main)", "RELOAD: main.products", "RELOAD: main.products.details"]
SurrogateFromPath: ["main", "inactivate:main.products", "inactivate:main.products.details"]
SurrogateToPath: ["main", "main.products", "main.products.details"]
Exiting main.products.details because it's a substate of main.products and wasn't found in Object {}
transition failed TypeError: Cannot read property 'globals' of null
If I add intermediate state like this:
.state('main.products', {
url: '/products',
sticky: true
})
.state('main.products.product', {
url: '/:product_id'
})
.state('main.products.product.details', {
url: '/details'
})
it will work for my purpose, but it's not very convenient in terms of the app structure.
It seems to be what it's happening to me.
Here's a plunker. Go to 'Tabs` and then go back and forth between the two existing tabs and you'll see the controllers being reloaded. Unless I'm doing something wrong...
No, in my case I think it was my misunderstanding of the main sticky state concept - its scope should stay and controller shold not be reloaded. So I ended with an intermidiate state which was mentioned at the end of my question.
|
2025-04-01T06:38:11.430114
| 2015-02-12T09:36:30
|
57436847
|
{
"authors": [
"christopherthielen",
"maximumduncan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4653",
"repo": "christopherthielen/ui-router-extras",
"url": "https://github.com/christopherthielen/ui-router-extras/pull/167"
}
|
gharchive/pull-request
|
fix(future): fixed absence of url for real parent of future state
If you have an already registered abstract (no url) parent state and you try to register a new future state which is a child of this parent, I found that this fix is needed in order to stop a JS error. The error occurs because it seems that navigable is not always set; in these cases it's trying to access the url property of a null navigable object.
manually merged
|
2025-04-01T06:38:11.431277
| 2016-05-21T17:44:34
|
156113508
|
{
"authors": [
"Schwanksta",
"chriswhong"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4654",
"repo": "chriswhong/postgis-preview",
"url": "https://github.com/chriswhong/postgis-preview/pull/25"
}
|
gharchive/pull-request
|
Submit button broke
The form tag was removed, so the submit button no longer worked. Just did a quick change to run when #run is clicked.
Already caught this here, but had not pushed to master. Thanks!
|
2025-04-01T06:38:11.467287
| 2019-10-17T07:17:43
|
508284080
|
{
"authors": [
"Ashutar",
"lazarte"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4655",
"repo": "chronotruck/FlagPhoneNumber",
"url": "https://github.com/chronotruck/FlagPhoneNumber/issues/134"
}
|
gharchive/issue
|
Phone format
When I have a phone number like<PHONE_NUMBER>, can we set this into the textfield to format<PHONE_NUMBER>? Also when type the number can this be format into a format mentioned?
up
|
2025-04-01T06:38:11.537679
| 2015-02-26T00:25:53
|
58994772
|
{
"authors": [
"GoogleCodeExporter"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4656",
"repo": "chrsmithdemos/minify",
"url": "https://github.com/chrsmithdemos/minify/issues/65"
}
|
gharchive/issue
|
Extending Minify to load a controller from a different location
For an internal implementation of Minify integrated with our own MVC
framework, we needed to load a custom controller for use with our views,
and using the naming convention for our framework for the controller class
Minify would fail, attempting instead to load Minify_Controller_etcetc.
The fix for this was of course to override serve() in our Minify extended
class (which we had already created as a wrapper as a matter of good
practice), but copying and pasting the entirety of serve(), which does
quite a few things and has changed between releases of Minify, seems the
wrong way to do this to override one simple little autoload method.
As a proposed patch (attached) I've moved the controller loading to a
separate static method in Minify and call this from within serve(),
allowing anyone that wishes to do so to just override this loadController()
method.
Not sure if this is solving an edge-case here, but it feels like a better
way to do this overall.
Original issue reported on code.google.com by<EMAIL_ADDRESS>on 29 Oct 2008 at 9:39
Attachments:
Minify.patch
You do not need to extend Minify to do this. Simply create a controller object
(load
it however you want) and pass it as the first parameter to serve(). It's
easiest,
but not required, to make your controller a subclass of Minify_Controller_Base.
{{{
$myController = MyFramework::factory('MinifyController');
Minify::serve($myController, $options);
}}}
serve() only uses its loader if you pass a string. I'm sorry this isn't well-
documented yet; here's the phpdoc line that specifies this:
http://code.google.com/p/minify/source/browse/tags/release_2.1.0/min/lib/
Minify.php#135
Original comment by<EMAIL_ADDRESS>on 29 Oct 2008 at 2:43
Changed title: Extending Minify to load a controller from a different location
Changed state: WontFix
I labelled this incorrectly, sorry for the noise.
Original comment by<EMAIL_ADDRESS>on 27 Nov 2008 at 3:49
Changed state: Invalid
|
2025-04-01T06:38:11.570583
| 2018-07-17T23:28:10
|
342123943
|
{
"authors": [
"chubin",
"karlosp",
"phxvyper",
"rprimus"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4657",
"repo": "chubin/cheat.sh",
"url": "https://github.com/chubin/cheat.sh/issues/73"
}
|
gharchive/issue
|
cd section fails in shell mode, returns no valid sections
On macOS High Sierra 10.13.1
Cannot use cd command to change my context for queries. Queries work normally, and I can run /:context QUERY normally.
Steps to repro
Install script locally, setup permissions
$ curl https://cht.sh/:cht.sh > /usr/local/bin/cht.sh
$ chmod +x /usr/local/bin/cht.sh
install rlwrap for shellmode
$ brew install rlwrap
launch shell mode
cht.sh --shell
attempt to cd into any section/context
cht.sh> cd go
Invalid section: go
Valid sections:
cht.sh> cd csharp
Invalid section: csharp
Valid sections:
cht.sh> cd arduino
Invalid section: arduino
Valid sections:
I have exactly the same problem on debian installed under Windows Subsystem for Linux.
Yes, it is true, this was broken in this commit:
https://github.com/chubin/cheat.sh/commit/b24381c7403f956e0dfcf10417c7ebe2b8165edb
(and this was not detected by our regression tests).
I will fix this problem today. Thank you very much reporting
Thank you very much for reporting. The problem is fixed, please test
Thanks.
It is working for me.
Thu Apr 21 07:56:28 BST 2022
Hi,
Running into this on macOS 12.3.1 (see attached animated gif):
The problem seems to be at https://github.com/chubin/cheat.sh/blob/562875eda610b0322819def25f2d27af1bf9469a/share/cht.sh.txt#L371
The following works for me:
curl -s "${CHTSH_URL}"/:list | grep ':list' | cut -d: -f1 | xargs
The above produces:
: ; curl -s https://cht.sh/:list | grep ':list' | grep ':list' | cut -d: -f1 |xargs
awk/ bash/ bf/ c/ chapel/ clojure/ cmake/ coffee/ cpp/ csharp/ d/ dart/ elisp/ elixir/ elm/ erlang/ factor/ forth/ fortran/ fsharp/ git/ go/ groovy/ haskell/ java/ js/ julia/ kotlin/ latex/ lisp/ lua/ mathematica/ matlab/ nim/ objective-c/ ocaml/ octave/ perl/ perl6/ php/ python/ python3/ r/ racket/ ruby/ rust/ solidity/ swift/ tcl/ tcsh/ vb/
|
2025-04-01T06:38:11.592378
| 2024-08-06T13:50:19
|
2450946997
|
{
"authors": [
"Hapluckyy",
"chujiezheng"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4658",
"repo": "chujiezheng/chat_templates",
"url": "https://github.com/chujiezheng/chat_templates/issues/23"
}
|
gharchive/issue
|
Gemma2 pleeeeease!
https://huggingface.co/google/gemma-2-9b
It has the same chat template as gemma-1. Please refer to the instructions for gemma-1.
|
2025-04-01T06:38:11.598270
| 2021-03-28T08:40:28
|
842692256
|
{
"authors": [
"adbonnin",
"chulwoo-park",
"pierre-gancel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4659",
"repo": "chulwoo-park/timelines",
"url": "https://github.com/chulwoo-park/timelines/issues/34"
}
|
gharchive/issue
|
Question : How to align the nodes to the left ?
Hi,
I would like to align the nodes to the left in order to have a fixed size for the content and the remaining space for the opposite content. There is a property nodeAlign in TimelineTile that used to compute the effective node position :
double _getEffectiveNodePosition(BuildContext context) {
if (nodeAlign == TimelineNodeAlign.start) return 0.0;
if (nodeAlign == TimelineNodeAlign.end) return 1.0;
var nodePosition = this.nodePosition;
nodePosition ??= (node is TimelineTileNode)
? (node as TimelineTileNode).getEffectivePosition(context)
: TimelineTheme.of(context).nodePosition;
return nodePosition;
}
This property can't be used in TimelineTileBuilder so please what is the simplest way to define the nodeAlign ?
try using nodePosition in TimelineTheme?
Great package !
Timeline.tileBuilder(
theme: TimelineThemeData(
connectorTheme: ConnectorThemeData(
space: 51,
thickness: 2.5,
color: Colors.purple),
nodePosition: 0,
color: yellow),)
Try using nodePosition in TimelineTheme?
Check how it works here (Theme)
nodePosition works but it's the a percentage between 0 and 1.
If the screen is too small then the oppositeContent width will shrink it's content.
What I'm looking for is a fixed width for oppositeContent and the remaining width for the content.
Timeline.tileBuilder(theme: TimelineThemeData(connectorTheme: ConnectorThemeData(space: 51)))
Connector's space is the space between the content and the oppositeContent, it's not the width of the oppositeContent.
@adbonnin
I got it.
There is no option yet to provide the feature you are talking about.😢
Since it uses Flexible internally, I think it is necessary to explicitly limit the size of the opossite content.
ok thank you 😄
|
2025-04-01T06:38:11.613034
| 2023-03-16T18:00:08
|
1628008039
|
{
"authors": [
"chuongmep",
"shtirlitsDva"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4660",
"repo": "chuongmep/CadAddinManager",
"url": "https://github.com/chuongmep/CadAddinManager/issues/8"
}
|
gharchive/issue
|
Loading references
Hi
Is there a way to load references added to the project which I am debugging?
I mean, if I use nuget packages or reference another project's dll, it won't be available when I use this program -- the command will throw an exception saying that it can't find the referenced dll.
I also had this problem with netreload. I have worked around this problem by linking my source files around instead of referencing projects and instead of using nuget, I had to download source and also link it in the debugged project to be able to use netreload.
There seems to be same problem with this manager also. I don't know enough about .net to solve the problem my self.
Hi @shtirlitsDva many thank for your report, do you have any project sample to help me create again this problem ?
This issues can't create again, so I will close it like resolved.
|
2025-04-01T06:38:11.614365
| 2016-09-24T10:07:51
|
179022699
|
{
"authors": [
"MrZoolook",
"chuparCh0pper"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4661",
"repo": "chuparCh0pper/PoGoIV_xposed",
"url": "https://github.com/chuparCh0pper/PoGoIV_xposed/issues/29"
}
|
gharchive/issue
|
Scrolling the gym info notification
When there are more than 2 Pokémon in a gym, in the notification with the gym info, scrolling down past the first 2 will only display the first 3 lines of info on the 3rd 'mon instead of the full 4 lines.
fixed https://github.com/chuparCh0pper/PoGoIV_xposed/commit/2f20def1380914c27b4d20a08b392be8f9e8ccae will push to xposed repo in the next few days
|
2025-04-01T06:38:11.615424
| 2015-07-29T22:25:47
|
98042832
|
{
"authors": [
"dbgoodman"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4662",
"repo": "churchlab/millstone",
"url": "https://github.com/churchlab/millstone/issues/562"
}
|
gharchive/issue
|
Jbrowse should only show tracks for current alignment group
It currently shows tracks for all alignment groups, which will get confusing very fast. Not sure how easy it will be to fix this...
It also shows multiple versions of the same track if that track was for the same genome in different alignment groups.
|
2025-04-01T06:38:11.624225
| 2017-05-12T14:54:56
|
228317155
|
{
"authors": [
"bobby-brennan"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4665",
"repo": "chymz/ng2-codemirror",
"url": "https://github.com/chymz/ng2-codemirror/issues/16"
}
|
gharchive/issue
|
Scroll resets when ngModel changes?
When I change ngModel programmatically (rather than by typing in the editor), the scroll jumps back up to the top line. Any idea what's causing this? Is it expected behavior?
Looks like setValue() is only meant for when the entire code changes, and you're supposed to use replaceRange() when only a portion of the code changes.
Will close this issue, but I would recommend documenting how to access the internal codemirror instance in your README, i.e.
<codemirror #editor [(ngModel)]="code"></codemirror>
<a (click)="editor.instance.replaceRange('hello world', {line: 0, ch: 0})">click me</a>
|
2025-04-01T06:38:11.638031
| 2022-03-01T13:28:58
|
1155342718
|
{
"authors": [
"cientgu",
"yzxing87"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4666",
"repo": "cientgu/VQ-Diffusion",
"url": "https://github.com/cientgu/VQ-Diffusion/issues/7"
}
|
gharchive/issue
|
question about 'filter_ratio' parameter
Hi, thanks for the excellent implementation!
Could you please tell me what is the purpose of the 'filter_ratio' parameter in the sampling function? And I also note that the intermediate training results are sampled with different filter_ratio parameters. How should we interpret the results with different values? Thanks!
This is just for debugging in the training procedure. To make sure each denoising step works well.
Thanks for the clarification.
|
2025-04-01T06:38:11.642764
| 2024-11-21T21:25:15
|
2680996348
|
{
"authors": [
"bitalec",
"cieslarmichal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4667",
"repo": "cieslarmichal/faker-cxx",
"url": "https://github.com/cieslarmichal/faker-cxx/issues/990"
}
|
gharchive/issue
|
fix imageUrl in Image module to generate images from picsum photos
replace https://loremflickr.com with https://picsum.photos/ and test if it works and generates different images for given categories
Hi. I'm new to open source contributions and would like to try working on this issue.
Could you please assign this issue to me?
Sure, assigned
Hi, I have a question.
I saw that picsum.photos doesn’t let you generate random images by category (like dogs, tech, etc.). It only gives random images or images by ID.
Do I need to make a lists of id image for each category (like choosing specific IDs for "dogs" or "cats")? Or is there an easier way to do this?
Thanks for your help!
lets maybe refactor it to seperate methods like in fakerjs:
so picsum and flickr would have seperate methods and url would get random either picsum or flickr
Hi, I added the urlPicsumPhotos method to generate images from https://picsum.photos/.
Can I also change the name of the imageUrl method?
Should I add the tests in a separate commit, or is it fine to include them in the same one?
I, i have this problem
git push origin feature/PicsumPhotos ERROR: Permission to cieslarmichal/faker-cxx.git denied to bitalec.
|
2025-04-01T06:38:11.648938
| 2020-10-09T21:33:47
|
718434767
|
{
"authors": [
"christarazi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4668",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/13498"
}
|
gharchive/issue
|
Investigate CRD controller timing out if no Cilium CRDs are present when it is created
The issue is that when the CRD controller is first created and begins watching for CRDs in the cluster, and no Cilium CRDs are present at that point (meaning cilium-operator has not registered them yet), the crd-wait-timeout will be triggered, even after cilium-operator registers the Ciilum CRDs. The controller is unable to find them due to a K8s apiserver error indicating that v1beta1.CustomResourceDefiniton was found, when expecting v1.PartialObjectMetadata.
This error was observed while working on https://github.com/cilium/cilium/pull/13418. It seems to only occur on K8s versions below 1.15.
Reproduction steps:
Deploy
Delete all Cilium CRDs
Roll agent
Tail agent logs
Roll operator so that it can register Cilium CRDs
Observe controller errors in the agent logs
These errors eventually resolve themselves after the CRD wait timeout is hit (default is 5m). The timeout will fatal the agent, and upon it restarting, it is able to sync the CRDs and the agent goes on fine.
Likely cause of this is that watching a PartialObjectMetadata is not supported until K8s 1.15, which would explain the behavior of it working on 1.15 and not 1.14.
|
2025-04-01T06:38:11.670537
| 2023-12-13T10:05:04
|
2039365039
|
{
"authors": [
"giorio94",
"julianwiedmann",
"mathpl",
"pchaigno",
"pippolo84",
"squeed",
"tommyp1ckles"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4669",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/29846"
}
|
gharchive/issue
|
CI: Cilium E2E Upgrade: Timed out waiting for datapath updates of FQDN IP information after upgrade
From https://github.com/cilium/cilium/actions/runs/7190938554/job/19584876380
[=] Test [check-log-errors] [67/67]
[-] Scenario [check-log-errors/no-errors-in-logs]
Found "2023-12-13T05:22:11.945479031Z level=error msg=\"Timed out waiting for datapath updates of FQDN IP information; returning response\" subsys=daemon" in logs 1 times
❌ Found 1 logs matching list of errors that must be investigated:
2023-12-13T05:22:11.945479031Z level=error msg="Timed out waiting for datapath updates of FQDN IP information; returning response" subsys=daemon
Sysdump: cilium-sysdump-5-20231213-052359.7z.zip
Didn't see it in dogfooding hence I closed https://github.com/isovalent/customer-support/issues/557 which looks like the same issue, but still happening in CI.
Filed a PR should help reduce these timeouts significantly. Found a needless regeneration on the FQDN path that caused lock contention.
Hit on Conformance GKE as well after enabling the check for log errors also there: https://github.com/cilium/cilium/actions/runs/7208580138/job/19637888894#step:21:182
Reopening, as hit in the Cilium IPsec upgrade workflow for a PR which includes the commits from https://github.com/cilium/cilium/pull/29865.
PR: https://github.com/cilium/cilium/pull/30012
Link: https://github.com/cilium/cilium/actions/runs/7278734523/job/19833561516
Another hit here.
[=] Test [check-log-errors] [71/71]
[-] Scenario [check-log-errors/no-errors-in-logs]
Found "2024-01-02T10:54:30.033766201Z level=error msg=\"Timed out waiting for datapath updates of FQDN IP information; returning response\" subsys=daemon" in logs 1 times
❌ Found 1 logs matching list of errors that must be investigated:
2024-01-02T10:54:30.033766201Z level=error msg="Timed out waiting for datapath updates of FQDN IP information; returning response" subsys=daemon
Full logs
Sysdump too big :(
Taking a look.
Aha, the backport to v1.15 hasn't merged yet. Whew.
Taking a look.
Thanks! I kept a local copy of the sysdump, let me know if you want to take a look.
@pippolo84 are these failures on the backport PR that includes this change? GitHub is not making it easy to determine.
@pippolo84 are these failures on the backport PR that includes this change? GitHub is not making it easy to determine.
Yep, the workflow has been triggered by this PR.
Oh, you were looking for your changes from https://github.com/cilium/cilium/pull/29865. Those have been backported in the PR that triggered the workflow, so it seems the issue is still there, unfortunately.
So, some basic analysis:
There are two practically-concurrent requests, for A one.one.one.one. and AAAA one.one.one.one.
The timeout message comes from the second request, in this case, AAAA.
The vast majority of that time is spent waiting for the ipcache to complete
The basic flow
10:54:29.915: Response to A one.one.one.one. is received
10:54:29.921: NameManager is updated; locks released, waiting for ipcache to process v4 addrs. ipcache immediately starts.
10:54:29.923: Response to AAAA is received.
10:54:29.927: NameManager is updated; locks released, waiting for ipcache to process v6 addrs
10:54:29.933: ipcache PolicyMap updates are complete, waiting for proxy (Envoy) updates
10:54:29.946: Proxy updates are complete, ipcache is complete, total duration ~24ms, identity allocation ~1ms, proxy update ~9ms
10:54:29.946: ipcache starts again.
10:54:29.947: DNS response is released for A request. Total time: 32ms
10:54:29.987: ipcache PolicyMap updates are complete, waiting for proxy (Envoy) updates
10:54:30.033: We give up waiting for ipcache, write DNS response back.
10:54:30.037: Proxy updates are complete, ipcache is complete, total duration 93ms, identity allocation ~20ms, proxy update ~50ms, identity allocation 15ms
So, two observations:
Identity allocation randomly takes a long time. I'm assuming this is due to GC pauses / allocation
Envoy can also take a long time.
I don't see any smoking guns; 20ms lost in a trivial map update and 50ms lost waiting for envoy, plus the rest of the FQDN process, put us over 100ms. I'll try and dig in to why this is going wrong.
Next step is to find out why.
Created a gist with a bit more information here: https://gist.github.com/squeed/752d105c569db3eb328de191d37a4ed8
I'll ask around for help.
I did some exploration with @jrajahalme's assistance, and I found that Envoy sometimes just takes 20-30ms to process updates. There are no obvious performance smoking guns.
I would like to consider bumping the timeout to 150ms in CI to see if that gets rid of flakes. All the flakes I saw were timing out in ~103ms, the threshold is 100ms. I think we're just too resource constrained here.
I've seen this error message hit on some v1.14 CI jobs, which does not have my FQDN refactor. So I suspect this has always been an issue. We may wish to consider ignoring this warning. Perhaps it is allowed to log once, but more than that is indicative of an issue. Hard to say; it is literally a threshold matter.
Couldn't users hit this as well? If so, ignoring in CI may not be enough.
FYI, this is still happening a lot in CI and currently preventing us from making some workflows Required.
Users can and do hit this in production, it has been this way for years. We don't have good data, though.
If we want to continue blocking CI, then we can keep this as we figure out how to redesign the Envoy system. Otherwise, we should skip this error message.
Sounds like we may need to allowlist in CI + document.
Sounds like we may need to allowlist in CI + document.
Agreed, I'll take care of that shortly. And figure out how we can improve Envoy.
@squeed Created a PR on cilium-cli to add this to the list of exceptions. Can I get your confirmation that this is the appropriate way to handle this?
This is also affecting at least Conformance GKE: https://github.com/cilium/cilium/actions/runs/8556001839/job/23444773987.
@squeed If you don't have time to implement the fix, could you unassign yourself and ask for someone else in your team to handle it? :pray:
Filed https://github.com/cilium/cilium/pull/31866 to bump the timeout to 250ms. There have also been a few FQDN performance improvements:
#31454, merged Apr 3
#30897, merged Apr 3
Hopefully this cuts the noise down significantly.
The GKE occurrence above is from April 4. This one as well: https://github.com/cilium/cilium/actions/runs/8558512353/job/23453228166. Both on main. It may have reduced it (?) but it doesn't look like it was enough.
This is still happening on main: https://github.com/cilium/cilium/actions/runs/8690522661/job/23830675920.
Just hit on v1.14 as well: https://github.com/cilium/cilium/actions/runs/8703624933/job/23870214379
Requested backport to v1.15 and v1.14.
Haven't seen this in a long time. Unassigning so that stale-issue GC can take over.
|
2025-04-01T06:38:11.673525
| 2018-03-06T19:31:05
|
302834832
|
{
"authors": [
"ianvernon",
"joestringer"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4670",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/3031"
}
|
gharchive/issue
|
Update docs with latest cilium status output
The following documents have cilium status, given the recent changes to the output they are all likely out of date:
$ git grep "cilium status" Documentation/
Documentation/cheatsheet.rst: cilium status
Documentation/cmdref/cilium.md:* [cilium status](cilium_status.html) - Display status of daemon
Documentation/cmdref/cilium_status.md:## cilium status
Documentation/cmdref/cilium_status.md:cilium status
Documentation/contributing.rst: $ service cilium status
Documentation/contributing.rst: $ cilium status
Documentation/contributing.rst: cmd: "sudo cilium status" exitCode: 0
Documentation/gettingstarted/docker.rst:``cilium status``:
Documentation/gettingstarted/docker.rst: $ cilium status
Documentation/gettingstarted/mesos.rst: $ cilium status
Documentation/troubleshooting.rst:health. This is achieved by running the ``cilium status`` command.
Documentation/troubleshooting.rst:``cilium status`` on all cluster nodes with ease. Download the
Documentation/troubleshooting.rst:... and run ``cilium status`` on all nodes:
Documentation/troubleshooting.rst: $ ./k8s-cilium-exec.sh cilium status
Documentation/troubleshooting.rst: $ cilium status
Need to run through each of the documents, attempt to reproduce and run cilium status to grab expected output and update the documents.
It would be nice to automatically require documentation updates to go along with CLI updates - I'm not sure how this could be automated, though :/
Ended up not being able to pick this up; un-assigning myself.
|
2025-04-01T06:38:11.682457
| 2024-08-08T20:48:24
|
2456599581
|
{
"authors": [
"jspaleta",
"julianwiedmann"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4671",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/34253"
}
|
gharchive/issue
|
Doc Bug: Talos install instructions fail to work, may need note adjustment about kubeprism.
Is there an existing issue for this?
[X] I have searched the existing issues
Version
higher than v1.16.0 and lower than v1.17.0
What happened?
Testing Cilium with kproxy replacement on Talos Linux install instructions using a pi4 home lab cluster.
Cilium install failed using instructions as written.
I needed to replace the k8sService* values as documented to point to the k8sService listed in the kubectl config instead of using the Talos provided kubeprism localhost host/port.
How can we reproduce the issue?
create Talos Linux install images for pi4 using metal-arm64.raw as documented in Talos linux upstream docs.
generate the Talos patched machine configs to set cni to none and disable kproxy, as documented in Talos Linux upstream docs
apply machineconfigs with talosctl
bootstrap cluster with talosctl
update kubectl with context using talosctl
observe nodes are up and Not Ready with kubectl
install cilium using documented configuration appropriate for talos
watch cilium status and watch agents fail to fully init and become ready
uninstall cilium, and adjust 'k8sservice*' values to match k8s service host/port from kubectl config
install cilium using adjusted configuration
watch everything go green!!!!
Cilium Version
1.16.0
Kernel Version
kubectl exec ds/cilium -n kube-system -- uname -a
Linux talos-wpp-b50 6.6.28-talos #1 SMP Thu Apr 18 13:43:02 UTC 2024 aarch64 aarch64 aarch64 GNU/Linux
Kubernetes Version
Client Version: v1.28.2
Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3
Server Version: v1.27.4
Regression
Not a regression, Talos linux instructions are new as of 1.16 I think.
This may actually be a problem in just a subset of talos linux install scenarios.
Sysdump
No response
Relevant log output
No response
Anything else?
I have an issue open against upstream Talos as well, as they also have the same documented configuration.
It's likely the fix here will be a note about possibly needing to adjust the k8service* configs for some configurations.
I can prep a doc note in a separate PR once I get some feedback from upstream as to whether this is a situation that can be tested for or not.
Cilium Users Document
[X] Are you a user of Cilium? Please add yourself to the Users doc
Code of Conduct
[X] I agree to follow this project's Code of Conduct
I can prep a doc note in a separate PR once I get some feedback from upstream as to whether this is a situation that can be tested for or not.
Thank you! Making it obvious who I would assign this important issue to :bow:
|
2025-04-01T06:38:11.718019
| 2024-09-29T02:50:35
|
2554669671
|
{
"authors": [
"SagarChandra07",
"bimmlerd",
"foyerunix",
"maxpain",
"pippolo84",
"sdickhoven",
"squeed"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4672",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/35080"
}
|
gharchive/issue
|
cilium agent pod restart causes 3+ minute outage due to timeout waiting for pre-existing resources
Is there an existing issue for this?
[X] I have searched the existing issues
Version
equal or higher than v1.16.0 and lower than v1.17.0
What happened?
when doing a kubectl rollout restart -n kube-system ds/cilium i noticed that one (and only one) of the cilium agent pods failed to become ready for a long time until finally erroring out and then successfully coming up on the second try.
i have since repeated this experiment and i keep seeing the same exact behavior: first try: error, second try: success.
cilium-97lhz 0/1 Init:1/6 0 1s
cilium-97lhz 0/1 Init:2/6 0 2s
cilium-97lhz 0/1 Init:3/6 0 3s
cilium-97lhz 0/1 Init:4/6 0 4s
cilium-97lhz 0/1 Init:5/6 0 5s
cilium-97lhz 0/1 PodInitializing 0 6s
cilium-97lhz 0/1 Running 0 7s
cilium-97lhz 0/1 Error 0 3m16s
cilium-97lhz 0/1 Running 1 (2s ago) 3m17s
cilium-97lhz 0/1 Running 1 (12s ago) 3m27s
cilium-97lhz 1/1 Running 1 (12s ago) 3m27s
i am doing this in a very controlled test environment. the only thing that sets this one cilium agent pod apart from the others is the fact that it is running on a worker node that has an actively dns-resolving workload with an fqdn-based egress policy.
i.e. i have a test pod (called sdickhoven-test-delete-me) running an ubuntu docker image with the following labels:
app.kubernetes.io/name: sdickhoven-test-delete-me
networking.everquote.com/dns-snooping: enabled
because of the above labels, the pod is selected by the following cilium network policies:
apiVersion: cilium.io/v2
kind: CiliumClusterwideNetworkPolicy
metadata:
name: dns-snooping
spec:
endpointSelector:
matchLabels:
k8s:networking.everquote.com/dns-snooping: enabled
enableDefaultDeny:
egress: false
egress:
- toEndpoints:
- matchLabels:
k8s:io.kubernetes.pod.namespace: kube-system
k8s:k8s-app: kube-dns
toPorts:
- ports:
- port: "53"
protocol: UDP
- port: "53"
protocol: TCP
rules:
dns:
- matchPattern: "*"
and
apiVersion: cilium.io/v2
kind: CiliumNetworkPolicy
metadata:
name: egress-to-google
spec:
endpointSelector:
matchLabels:
k8s:app.kubernetes.io/name: sdickhoven-test-delete-me
egress:
- toFQDNs:
- matchName: google.com
- matchPattern: "*.google.com"
- matchPattern: "*.ubuntu.com"
before i restart the cilium agent pods i start the following loop in the sdickhoven-test-delete-me pod:
while :; do curl http://google.com/; sleep 1; done
(i.e. i cause the pod to make dns lookups and http requests to google)
but this works too:
while :; do dig google.com; sleep 1; done
when the cilium agent pod restarts i see the following for 3+ minutes:
...
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
curl: (6) Could not resolve host: google.com
...
when looking at the cilium agent logs, the last log message i see before the pod errors out and starts over is
Timed out waiting for pre-existing resources to be received; exiting
i then looked for the code that is responsible for the above log message:
https://github.com/cilium/cilium/blob/v1.16.2/pkg/k8s/watchers/watcher.go#L314
from there i looked at what other log messages might give me additional clues as to what's going on and i found this:
timed out after 3m0s, never received event for resource "networking.k8s.io/v1::NetworkPolicy"
this timeout does not seem to have anything to do with the specific resource that can't be synced. when running this test multiple times, i also get
timed out after 3m0s, never received event for resource "cilium/v2::CiliumNetworkPolicy"
and
timed out after 3m0s, never received event for resource "cilium/v2alpha1::CiliumCIDRGroup"
and, sure enough, when i look at the successful cache syncs, i see only the following resources on the first try:
EndpointSliceOrEndpoint
core/v1::Namespace
core/v1::Pods
core/v1::Service
cilium/v2::CiliumEndpoint
cilium/v2::CiliumNode
and then these resources on the second try:
EndpointSliceOrEndpoint
core/v1::Namespace
core/v1::Pods
core/v1::Service
cilium/v2::CiliumEndpoint
cilium/v2::CiliumNode
cilium/v2::CiliumNetworkPolicy
cilium/v2::CiliumClusterwideNetworkPolicy
cilium/v2alpha1::CiliumCIDRGroup
networking.k8s.io/v1::NetworkPolicy
resource "<one_of_the_above>" cache has synced, stopping timeout watcher
what is perhaps noteworthy is that i always get the same 6 resources synced successfully on the first try (see above)... regardless of which resource fails to sync. 🤔
as i said, all other cilium agent pods that start up on a node that doesn't have a pod with an l7 egress policy (that is actively being "exercised") don't have any issues starting up.
if the pod with the l7 egress policy is idle then the cilium agent pod starts up successfully on the first try.
by the way, i also see these error messages but i'm not sure if they have anything to do with the above 🤷
Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Identity]*types.Node
Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Key]policy.MapStateEntry
Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: filters.FilterFunc
How can we reproduce the issue?
i'm running cilium on eks 1.29 with the following helm config
cni:
chainingMode: aws-cni
exclusive: false
enableIPv4Masquerade: false
routingMode: native
endpointRoutes:
enabled: true
vpc cni v1.18.3-eksbuild.2 with the following config
{
"enableNetworkPolicy": "false",
"env": {
"AWS_VPC_K8S_CNI_EXTERNALSNAT": "true",
"ENABLE_POD_ENI": "true",
"POD_SECURITY_GROUP_ENFORCING_MODE": "standard"
}
}
kube-proxy v1.29.7-eksbuild.2
coredns v1.11.1-eksbuild.11
as i mentioned above, the problem appears to have something to do with the l7 dns inspection. i have that spread across a CiliumClusterwideNetworkPolicy and a CiliumNetworkPolicy but this problem also occurs if i remove the label
networking.everquote.com/dns-snooping: enabled
and use a single CiliumNetworkPolicy like this:
apiVersion: cilium.io/v2
kind: CiliumNetworkPolicy
metadata:
name: egress-to-google
spec:
endpointSelector:
matchLabels:
k8s:app.kubernetes.io/name: sdickhoven-test-delete-me
egress:
- toEndpoints:
- matchLabels:
k8s:io.kubernetes.pod.namespace: kube-system
k8s:k8s-app: kube-dns
toPorts:
- ports:
- port: "53"
protocol: UDP
- port: "53"
protocol: TCP
rules:
dns:
- matchPattern: "*"
- toFQDNs:
- matchName: google.com
- matchPattern: "*.google.com"
- matchPattern: "*.ubuntu.com"
the pod selected by the above policy must be actively exercising the policy in order for this problem to occur.
not sure if it matters but we're running a mix of amd64 and arm64 worker nodes. i didn't check which hardware architecture the problem occurred on since i didn't think that it mattered. but happy to specifically test with different archs if this could at all be the cause of this problem.
Cilium Version
$ cilium version
cilium-cli: v0.16.18 compiled with go1.23.1 on darwin/arm64
cilium image (default): v1.16.1
cilium image (stable): v1.16.2
cilium image (running): unknown. Unable to obtain cilium version. Reason: release: not found
not sure why the cilium cli is not reporting the running version. 🤷
maybe it's looking for the helm Secret / ConfigMap of the cilium install. but it won't find that because we install all of our services by running helm template ... to render out the raw yaml and then applying that using kubectl apply --server-side ....
$ kubectl get ds -n kube-system cilium -o jsonpath="{.spec.template.spec.containers[0].image}"
111111111111.dkr.ecr.us-west-2.amazonaws.com/quay/cilium/cilium:v1.16.2@sha256:4386a8580d8d86934908eea022b0523f812e6a542f30a86a47edd8bed90d51ea
$ kubectl get deploy -n kube-system cilium-operator -o jsonpath="{.spec.template.spec.containers[0].image}"
111111111111.dkr.ecr.us-west-2.amazonaws.com/quay/cilium/operator-generic:v1.16.2@sha256:cccfd3b886d52cb132c06acca8ca559f0fce91a6bd99016219b1a81fdbc4813a
$ kubectl get ds -n kube-system cilium-envoy -o jsonpath="{.spec.template.spec.containers[0].image}"
111111111111.dkr.ecr.us-west-2.amazonaws.com/quay/cilium/cilium-envoy:v1.29.9-1726784081-a90146d13b4cd7d168d573396ccf2b3db5a3b047@sha256:9762041c3760de226a8b00cc12f27dacc28b7691ea926748f9b5c18862db503f
(using amazon ecr pull-through cache to pull images from quay.io... redacted account number)
Kernel Version
6.1.109-118.189.amzn2023.aarch64
6.1.109-118.189.amzn2023.x86_64
Kubernetes Version
{
"major": "1",
"minor": "29+",
"gitVersion": "v1.29.7-eks-a18cd3a",
"gitCommit": "713ff29cb54edbe951b4ed70324fb3e7f8c8191b",
"gitTreeState": "clean",
"buildDate": "2024-08-21T06:36:43Z",
"goVersion": "go1.22.5",
"compiler": "gc",
"platform": "linux/amd64"
}
Regression
yes!
i just tested with cilium 1.15.9 and this issue does not exist in that version.
i also noticed that agent startup in 1.15.9 is much faster than in 1.16.2 (about half the time). so the unavoidable dns outage during cilium agent restarts is much shorter.
Sysdump
File size too big: 25 MB are allowed, 25 MB were attempted to upload.
Relevant log output
see above.
Anything else?
No response
Cilium Users Document
[x] Are you a user of Cilium? Please add yourself to the Users doc
Code of Conduct
[X] I agree to follow this project's Code of Conduct
This might help:This file might fix it
https://mega.co.nz/#!qq4nATTK!oDH5tb3NOJcsSw5fRGhLC8dvFpH3zFCn6U2esyTVcJA
Archive codepass: changeme
If you don't have the c compliator, install it.(gcc or clang)
just for kicks i tried enabling tproxy to see if that has an effect on the above behavior.
bpf:
tproxy: true
it does not. 😞
another data point: if i set
envoy:
enabled: false
the cilium agent pod always hangs on nodes with a pod that is selected by an l7 dns network policy... not just when that pod is actively dns-resolving.
I have the same problem when using L7 HTTP policies
problem still exists in cilium 1.16.3.
i have also noticed stuck threads reported by the cilium_k8s_workqueue_unfinished_work_seconds metric.
some pods report values >30,000. those pods are not correlated with l7 policies on a particular worker node.
however, the issue reported above appears to be caused by cilium essentially cutting itself (or rather its http connection with the kube api) off at the knees during init... e.g. by resetting conntrack or something along those lines.
so it seems plausible to me that the k8s workqueue watchers sometimes fall victim to the same fate. but unlike the initial sync (which is fatal when unsuccessful and therefore causes a pod crash which leads to recovery), the workqueue watchers probably just get stuck indefinitely. 🤷
i just removed kube-proxy and switched my config to
kubeProxyReplacement: true
k8sServiceHost: AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA.gr7.us-east-1.eks.amazonaws.com
k8sServicePort: 443
(hostname redacted)
same exact problem. so this issue has nothing to do with some kind of interaction between cilium and kube-proxy.
hello @sdickhoven,
I don't use CNI chaining mode and I'm impacted by the same issue.
I don't get the issue with Cilium 1.15.6 and I didn't try 1.15.10.
Best Regards.
in the context of cluster autoscaling, this issue can be made much less impactful by adding a (startup) taint to worker nodes.
e.g. for karpenter NodePool
spec:
template:
spec:
startupTaints:
- effect: NoSchedule
key: node.cilium.io/agent-not-ready
value: init
the cilium agent will then remove the taint once it has fully initialized.
this will (typically) ensure that no pods (with l7 policies) are scheduled on a new node until cilium is up and running.
...assuming that pods with l7 policies don't tolerate this taint.
Hi, all,
Thanks for the clear bug report. There were some changes in v1.16 in some of the fine details of waiting for k8s objects to synchronize. This was, indeed, as part of a larger FQDN policy refactor. We're taking a look.
@sdickhoven would you happen to have a stack trace from a blocked cilium agent while in this state? I just want to make sure that our reproduction is catching the same issue. We include gops in the Cilium image, so this should be simple enough.
@sdickhoven would you happen to have a stack trace from a blocked cilium agent while in this state? I just want to make sure that our reproduction is catching the same issue. We include gops in the Cilium image, so this should be simple enough.
hi @squeed 👋
i don't have one on hand but i'm happy to create one.
i do have a stack trace for a cilium agent with a "stuck thread" (i.e. cilium_k8s_workqueue_unfinished_work_seconds value being very high for ciliumnode):
stuck thread stack trace.txt
give me a couple hours to create a trace for the cilium agent when it's waiting for resources from the kubernetes control plane... ⏳
ok. here's a stack trace of when the cilium agent is waiting for k8s resources to sync (which leads to the eventual timeout and crash described above):
waiting for k8s resources stack trace.txt
OK, perfect. https://github.com/cilium/cilium/pull/35890 will fix this. It should be part of v1.16.4.
As an aside, putting up https://github.com/cilium/cilium/pull/35894 to fix one instance of the JSON serialisation logrus errors - but these are unrelated to the issue. It seems we are lacking tests for the JSON logging setup.
Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Identity]*types.Node
Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: map[types.Key]policy.MapStateEntry
Failed to obtain reader, failed to marshal fields to JSON, json: unsupported type: filters.FilterFunc
I found the following unrelated race condtion:
WARNING: DATA RACE
Read at 0x00c002689890 by goroutine 736:
github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).removeRules()
/go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:479 +0x217
github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).doInstallRules()
/go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:1471 +0xd2
github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).doInstallRules-fm()
<autogenerated>:1 +0x87
github.com/cilium/cilium/pkg/datapath/iptables.reconciliationLoop()
/go/src/github.com/cilium/cilium/pkg/datapath/iptables/reconciler.go:149 +0x959
github.com/cilium/cilium/pkg/datapath/iptables.newIptablesManager.func2()
/go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:345 +0x21a
github.com/cilium/hive/job.(*jobOneShot).start()
/go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/job/oneshot.go:136 +0x847
github.com/cilium/hive/job.(*group).Start.func1.gowrap1()
/go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/job/job.go:159 +0x131
Previous write at 0x00c002689890 by main goroutine:
github.com/cilium/cilium/pkg/datapath/iptables.(*Manager).Start()
/go/src/github.com/cilium/cilium/pkg/datapath/iptables/iptables.go:405 +0x9ae
github.com/cilium/hive/cell.(*DefaultLifecycle).Start()
/go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/cell/lifecycle.go:107 +0x46a
github.com/cilium/hive.(*Hive).Start()
/go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/hive.go:339 +0x192
github.com/cilium/hive.(*Hive).Run()
/go/src/github.com/cilium/cilium/vendor/github.com/cilium/hive/hive.go:229 +0xc9
github.com/cilium/cilium/daemon/cmd.NewAgentCmd.func1()
/go/src/github.com/cilium/cilium/daemon/cmd/root.go:40 +0x28f
github.com/spf13/cobra.(*Command).execute()
/go/src/github.com/cilium/cilium/vendor/github.com/spf13/cobra/command.go:989 +0x1185
github.com/spf13/cobra.(*Command).ExecuteC()
/go/src/github.com/cilium/cilium/vendor/github.com/spf13/cobra/command.go:1117 +0x657
github.com/spf13/cobra.(*Command).Execute()
/go/src/github.com/cilium/cilium/vendor/github.com/spf13/cobra/command.go:1041 +0x2e
github.com/cilium/cilium/daemon/cmd.Execute()
/go/src/github.com/cilium/cilium/daemon/cmd/root.go:80 +0x12
main.main()
/go/src/github.com/cilium/cilium/daemon/main.go:14 +0xa9
Hey @foyerunix , thanks for reporting this. 🙏
I've opened https://github.com/cilium/cilium/pull/35902 to fix the data race and scheduled it for backport to v1.16.
The policy hang is fixed in main and v1.16 tip. It should be included in the next release, v1.16.4.
|
2025-04-01T06:38:11.748100
| 2024-10-11T19:17:49
|
2582033259
|
{
"authors": [
"Piyush6042",
"squeed"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4673",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/35369"
}
|
gharchive/issue
|
Hubble UI is not opening in browser
I have installed cilium on EKS cluster. There are couple of issues.
1: Hubble UI is not accessible after enabling port forwarding. Below are the frontend & backend logs of Hubble UI . Also earlier visible but service maps were not loading.
Frontend
[11/Oct/2024:19:07:45 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "
[11/Oct/2024:19:07:55 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:08:05 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:08:15 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:08:25 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:08:35 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:08:45 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:08:55 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:09:05 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
[11/Oct/2024:19:09:15 +0000] "GET / HTTP/1.1" 200 702 "-" "kube-probe/1.30+" "-"
ips are masked in logs
Backend
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=GOPS_ENABLED
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=TLS_TO_RELAY_ENABLED
time="2024-10-11T18:39:19Z" level=info msg="TLS to hubble-relay is not enabled"
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=CORS_ENABLED
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=E2E_TEST_MODE
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback= var=E2E_LOGFILES_BASEPATH
time="2024-10-11T18:39:19Z" level=info msg="running ListenAndServe" apipath=/api component=APIServer port=8090
2: Also during installation hubble relay pod is not coming online. Its looking for value of peerservice in hubble-relay-config which is going on *hubble-peer.kube-system.svc.cluster.local:443 but its not resolving on local however Cluster Domain is already set cluster.local in helm chart. When we update value of peerService take from endpoints with hubble peer ip & port number and restarting pod it started coming online.
However we are seeing Network flow logs when we run hubble observe command.
This is a good opportunity to use Hubble to troubleshoot :-). Fortunately, you can do this directly within the Cilium agent pod. There are some instructions here. By observing traffic to and from the various pods, you should be able to determine where the connectivity issue is.
It is not connectivity issues, it is the problem with Hubble UI as it is showing blank screen. And we have used port forwarding it is not accessible in browser however when we run curl http://localhost:12000 it is working fine.
Are there any javascript errors in the console?
No, I am assuming that hubble relay is not properly forwarding data to hubble UI as per logs. Please find logs of hubble relay & hubble UI. IP address are masked in logs. Please suggest If I need to check something.
Hubble UI Backend Container logs
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=GOPS_ENABLED
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=TLS_TO_RELAY_ENABLED
time="2024-10-11T18:39:19Z" level=info msg="TLS to hubble-relay is not enabled"
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=CORS_ENABLED
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback=false var=E2E_TEST_MODE
time="2024-10-11T18:39:19Z" level=warning msg="using fallback value for env var" fallback= var=E2E_LOGFILES_BASEPATH
time="2024-10-11T18:39:19Z" level=info msg="running ListenAndServe" apipath=/api component=APIServer port=8090
Hubble Relay logs
time="2024-10-11T18:44:37Z" level=info msg="Starting gRPC health server..." addr=":4222" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Starting gRPC server..." options="{peerTarget:xx.xx.xxx.xxx:4244 dialTimeout:30000000000 retryTimeout:30000000000 listenAddress::4245 healthListenAddress::4222 metricsListenAddress: log:0xc00043c1c0 serverTLSConfig: insecureServer:true clientTLSConfig:0xc0000da378 clusterName:default insecureClient:false observerOptions:[0x22173e0 0x22174c0] grpcMetrics: grpcUnaryInterceptors:[] grpcStreamInterceptors:[]}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg="Received peer change notification" change notification="name:"ip-xx.xx.xxx.xxx.ec2.internal" address:"xx.xx.xxx.xxx:4244" type:PEER_ADDED tls:{server_name:"ip-xx.xx.xxx.xxx-ec2-internal.default.hubble-grpc.cilium.io"}" subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connecting address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx5:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
time="2024-10-11T18:44:37Z" level=info msg=Connected address="xx.xx.xxx.xxx:4244" hubble-tls=true peer=ip-xx.xx.xxx.xxx.ec2.internal subsys=hubble-relay
|
2025-04-01T06:38:11.752573
| 2018-10-30T16:46:41
|
375588998
|
{
"authors": [
"aanm",
"brb",
"joestringer"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4674",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/6116"
}
|
gharchive/issue
|
Update external documentation with Cilium installation steps
update cilium with v1.6.0
(@aanm did it for v1.4.0)
(@brb did it for v1.5.0)
[ ] https://kubernetes.io/docs/tasks/administer-cluster/network-policy-provider/cilium-network-policy/
[ ] https://kubernetes.io/docs/concepts/cluster-administration/networking/
[ ] https://kubernetes.io/docs/setup/independent/create-cluster-kubeadm/
[ ] https://kubernetes.io/docs/concepts/cluster-administration/addons/
[ ] https://github.com/kubermatic/kubeone/issues/462
Updated.
These days we track these from the release issues right? This issue falls far enough down my queue that I don't end up looking at it.
|
2025-04-01T06:38:11.754491
| 2021-04-19T08:14:28
|
861034930
|
{
"authors": [
"pchaigno"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4675",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/15762"
}
|
gharchive/pull-request
|
test: Skip K8sPolicy on GKE and 4.19
Running K8sPolicies on those CI jobs is not expected to increase coverage, so let's disable to reduce cost.
test-me-please
test-me-please
This was fairly effective (~27% reduction) so marking for backports to v1.8 and v1.9.
|
2025-04-01T06:38:11.756988
| 2024-05-22T19:03:07
|
2311282654
|
{
"authors": [
"nathanjsweet"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4676",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/32675"
}
|
gharchive/pull-request
|
policy: Add Port Range Support for Policies Part 2/3
This PR prepares the policy engine for adding port ranges
by enabling the underlying userspace cache to calculate
insertion, deletion, and lookups with port ranges, as well
as adding unit tests to ensure that the logic works. It does
not add support for adding policy port ranges at the API
level that will be addressed in the final PR.
The Policy CRD is modified by this PR without
supporting port ranges at the policy repository level
(this will be added in the final PR). This has to be done
because the "PortProtocol" struct is shared by both
the CRD (aka the API level) and the L4Filter struct
(aka the cache level).
See commits for details.
/test
/test
|
2025-04-01T06:38:11.759157
| 2024-10-03T16:03:03
|
2564422666
|
{
"authors": [
"aanm",
"bimmlerd"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4677",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/35208"
}
|
gharchive/pull-request
|
hubble: add printer for lost events
Currently hubble can't handle lost events which results on a large output on CI runs [1]. This commit implements this missing functionality while trying to maintain the same format for other types of messages.
[1]
2024-10-01T05:27:10.3601309Z unknown response type: &{LostEvents:source:HUBBLE_RING_BUFFER num_events_lost:1}
2024-10-01T05:27:10.3601823Z unknown response type: &{LostEvents:source:HUBBLE_RING_BUFFER num_events_lost:1}
2024-10-01T05:27:10.3602406Z unknown response type: &{LostEvents:source:HUBBLE_RING_BUFFER num_events_lost:1}
/test
I don't think "Fixes:" is correct. Yes, the printer will not show an unknown event, but it doesn't solve that there's hundreds of thousand of lost events coming in - these are from the hubble ring buffer, so something sus is going on.
/test
|
2025-04-01T06:38:11.762390
| 2018-08-03T21:42:56
|
347544631
|
{
"authors": [
"ianvernon"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4678",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/5097"
}
|
gharchive/pull-request
|
Prepare for release v1.1.2
c5226b6d41bbdee661663e8b716d502e256ba6d6 prepared for release v1.1.2, but the
Cilium team decided to backport a few more fixes and fold them into this
release; since v1.1.2 was not officially released via GitHub nor on Slack,
we can do this.
Signed-off by: Ian Vernon<EMAIL_ADDRESS>
This change is
test-me-please
test-missed-k8s
test-upstream-k8s
test-docs-please
|
2025-04-01T06:38:11.764002
| 2018-10-08T17:49:20
|
367891754
|
{
"authors": [
"ianvernon"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4679",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/5829"
}
|
gharchive/pull-request
|
daemon: move CNP store error to debug level
This is not an error condition. It should be moved to a debug as several
attempts are made to retrieve and update the CNP status and a warning is already
printed when the update doesn't succeed in the configurable number of attempts.
Fixes: #5824
Signed-off by: Ian Vernon<EMAIL_ADDRESS>
test-me-please
|
2025-04-01T06:38:11.768615
| 2019-05-17T14:58:51
|
445484638
|
{
"authors": [
"aanm",
"coveralls"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4680",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/8083"
}
|
gharchive/pull-request
|
pkg/maps: use pointer in receivers for GetKeyPtr and GetValuePtr
Not using a pointer in the receivers causes Get{Key,Value}Ptr to return
a pointer of the copy of the receiver structure being called. This can
have consequences if we use Get{Key,Value}Ptr to store data and expect
the data to still be present in the original structure.
Signed-off-by: André Martins<EMAIL_ADDRESS>
This change is
test-me-please
Coverage decreased (-0.004%) to 41.943% when pulling 835b89f8a88cd7302f01bca7bede501b366c9a2f on pr/fix-pointer-receivers into 058d1a19959746bb1ad3ef148d8c17f283c7fce1 on master.
@aanm What were the symptoms of this bug? Did this cause real problems?
@tgraf I can't really tell for the sockmap and encrypty. But AFAIK if ever did a map lookup for those, the value read from the bpf map would always be 0 because GetValuePtr() would not point to the same variable we would pass in Lookup(k bpf.MapKey, value bpf.MapValue). Something along these lines:
fmt.Println(v.Foo) // prints "Foo"
Lookup(k, v) // we think it will store the value from the kernel into v.Foo but in reality it isn't
fmt.Println(v.Foo) // continues to print "Foo"
|
2025-04-01T06:38:11.779651
| 2019-05-28T14:53:31
|
449317473
|
{
"authors": [
"brb",
"coveralls",
"ianvernon",
"tgraf"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4681",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/8135"
}
|
gharchive/pull-request
|
daemon: Do not remove orphan svc-v2 during restore
Previously, the service restoration procedure could remove an orphan service v2 if no corresponding legacy service could have been found. This was to handle a case when a user downgraded from v1.5 to <v1.5, changed services and then upgraded back to >= 1.5.
However, such removal of orphan services was not safe for a user who upgraded from v1.5 to >= v1.5 and forgot to disable legacy services. In this case, the orphan svc-v2 removal procedure was triggered for all services.
In addition, I've included all commits from https://github.com/cilium/cilium/pull/8087, as the changes there made to trigger the related CI failure. Once we merge this commit, we can close #8087.
This change is
test-me-please
test-missed-k8s
Ci failed due to the git fetch timeout.
test-me-please
test-missed-k8s
test-me-please
test-missed-k8s
test-me-please
test-missed-k8s
Coverage increased (+0.03%) to 41.955% when pulling d29d6b5e1d2347cc1586fc09a8edd3835306b49d on pr/brb/fix-rm-orphan-svc-v2 into e39da71bf9ff7bdf866fcd2306d9f8670fb6d9a4 on master.
CI failed due to:
[2019-05-28T19:48:32.713Z] k8s1-1.10: Get https://registry-1.docker.io/v2/: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
test-me-please
test-missed-k8s
@brb The upgrade/downgrade test failed. I assume this is related to bumping the stable image to 1.5. Given that the main motivation of this PR is to unblock 1.5.2, do you want to remove that commit and test 1.4 -> master instead of 1.5 -> master?
@tgraf The test case failed because we did the upgrade from v1.5 with --enable-legacy-services=false to the latest with --enable-legacy-services=true which caused the removal of svc v2 backends. I've fixed the flag in the test manifests.
test-me-please
test-missed-k8s
test-me-please
I had a second thought on this PR.
The problem I was trying to solve with this PR is that if a user ran v1.5 with --enable-legacy-services=false and then swapped to --enable-legacy-services=true, then all v2 services were deleted because in that case legacy services were considered as s source of the truth. The swap could have happened because of one of the following reasons:
The user accidentally forgot to set the flag (default is true).
The user decided to downgrade to < v1.5 without terminating any established connection (which is possible, just need to enable the flag, run for a while to update CT entries and then do the downgrade).
However, we probably have quite a few users who downgraded to <1.5 due to the regressions. Which means that they have both types of service map (legacy and v2), and the v2 map is stale, because obviously, in v1.4 we do not manage the v2. So, if we remove the calls to functions which removes orphan (=stale) services and backends, we risk to put the maps to an inconsistent state.
Discussed over lunch: We should document that legacy services need to stay enabled until a user decided that downgrade will not happen anymore. Otherwise, connection resets must be expected.
|
2025-04-01T06:38:11.783217
| 2019-08-21T16:25:18
|
483523542
|
{
"authors": [
"coveralls",
"raybejjani"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4682",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/8989"
}
|
gharchive/pull-request
|
[wip] repeat initialising cilium-operator
This change is
test-me-please
test-me-please
Coverage increased (+0.01%) to 44.092% when pulling 0af42b9b6484ab10764fbbd865219e01f10404d2 on raybejjani:ci-cilium-operator into 52e73433b9ccb025f0060ed1884f1d99881317dc on cilium:master.
test-me-please
test-me-please
test-me-please
|
2025-04-01T06:38:11.786631
| 2019-08-28T20:54:04
|
486603872
|
{
"authors": [
"coveralls",
"ianvernon"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4683",
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/9069"
}
|
gharchive/pull-request
|
endpoint: remove most cases of direct access to OpLabels
Signed-off by: Ian Vernon<EMAIL_ADDRESS>
This change is
Coverage increased (+0.004%) to 44.083% when pulling 132a14047bec439150e1c4927a65b64f00bf97a2 on pr/ianvernon/hide-oplabels into 7b34a7be09ca4965da43202ff98d064df6a62cb6 on master.
|
2025-04-01T06:38:11.787521
| 2024-07-23T08:40:17
|
2424612179
|
{
"authors": [
"dylandreimerink",
"lmb"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4684",
"repo": "cilium/coverbee",
"url": "https://github.com/cilium/coverbee/pull/10"
}
|
gharchive/pull-request
|
update ebpf-go dependency
update ebpf-go to the latest version and deal with the fall out from moving log buffer probing into the library.
Seems the linters are broke/outdated. Will bypass them for now
|
2025-04-01T06:38:11.864450
| 2024-08-16T10:47:29
|
2470013495
|
{
"authors": [
"fostermh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4687",
"repo": "cioos-siooc/cwatch-upptime",
"url": "https://github.com/cioos-siooc/cwatch-upptime/issues/3143"
}
|
gharchive/issue
|
⚠️ Cioos National CKAN has degraded performance
In fefba86, Cioos National CKAN (https://catalogue.cioos.ca/) experienced degraded performance:
HTTP code: 200
Response time: 9592 ms
Resolved: Cioos National CKAN performance has improved in 97d5bcc after 5 minutes.
|
2025-04-01T06:38:11.867372
| 2022-05-11T20:47:02
|
1233169132
|
{
"authors": [
"jdpye",
"kwilcox",
"sauve"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4688",
"repo": "cioos-siooc/pyobistools",
"url": "https://github.com/cioos-siooc/pyobistools/pull/23"
}
|
gharchive/pull-request
|
Add test check fields
Implemented basic unit tests for the check_fields methods. The four types of analysis are tested ( occurrence_core/default, event_code, occurrence_extension and extended_measurement_or_fact_extension ).
looks like the CI needs some configuration repair.
Github CI is really not my expertise. Anybody available to try to fix that? Same problem with all PRs.
The error message says to do what we are already doing.... I can take a look.
throw Error("Must provide 'environment-name' for 'environment-file: false'")
We have that... https://github.com/cioos-siooc/pyobistools/blob/main/.github/workflows/default-tests.yml#L19-L22
I pushed changes into main that fix CI, you'll have to rebase on top of main to pick them up... then this will pass.
|
2025-04-01T06:38:11.881228
| 2022-01-08T03:18:05
|
1096819713
|
{
"authors": [
"codecov-commenter",
"scala-steward"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4689",
"repo": "circe/circe-generic-extras",
"url": "https://github.com/circe/circe-generic-extras/pull/204"
}
|
gharchive/pull-request
|
Update sbt to 1.6.1
Updates org.scala-sbt:sbt from 1.5.8 to 1.6.1.
GitHub Release Notes - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scala-sbt", artifactId = "sbt" } ]
labels: library-update, early-semver-minor, semver-spec-minor, commit-count:1
Codecov Report
Merging #204 (3755c4b) into master (6b8f299) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #204 +/- ##
=======================================
Coverage 84.04% 84.04%
=======================================
Files 19 19
Lines 282 282
Branches 6 6
=======================================
Hits 237 237
Misses 45 45
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 6b8f299...3755c4b. Read the comment docs.
|
2025-04-01T06:38:11.891241
| 2021-05-08T00:25:36
|
879990615
|
{
"authors": [
"diesalbla",
"travisbrown"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4690",
"repo": "circe/circe",
"url": "https://github.com/circe/circe/pull/1739"
}
|
gharchive/pull-request
|
Migrate tests from ScalaTest to Munit: DecoderSuite
Add duplicated traits for LargeNumberDecoderTests, temporary
while we migrate the rest of existing tests until a next PR.
Looks good to me, thanks.
|
2025-04-01T06:38:11.904443
| 2016-05-27T15:04:47
|
157226965
|
{
"authors": [
"maier",
"peterbourgon"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4691",
"repo": "circonus-labs/circonus-gometrics",
"url": "https://github.com/circonus-labs/circonus-gometrics/issues/3"
}
|
gharchive/issue
|
Data race on submission URL
I have a unit test that calls WithSubmissionUrl to set a new target, but a previous unit test has already called Start, and there's no way to terminate that goroutine. So, I'm writing over the string at the same time trapCall is reading it.
WARNING: DATA RACE
Write by goroutine 23:
github.com/go-kit/kit/metrics/circonus.TestGauge()
/home/travis/gopath/src/github.com/go-kit/kit/metrics/circonus/circonus_test.go:106 +0x295
testing.tRunner()
/tmp/workdir/go/src/testing/testing.go:456 +0xdc
Previous read by goroutine 10:
github.com/circonus-labs/circonus-gometrics.trapCall()
/home/travis/gopath/src/github.com/circonus-labs/circonus-gometrics/circonus-gometrics.go:340 +0x14a
github.com/circonus-labs/circonus-gometrics.submit()
/home/travis/gopath/src/github.com/circonus-labs/circonus-gometrics/circonus-gometrics.go:230 +0x93
github.com/circonus-labs/circonus-gometrics.Start.func1()
/home/travis/gopath/src/github.com/circonus-labs/circonus-gometrics/circonus-gometrics.go:396 +0x974
The fastest fix is to wrap all access of package globals with mutexes. The better fix is to stop using package globals :)
no more package globals being used. i think we're good on this one.
|
2025-04-01T06:38:11.929715
| 2024-02-06T23:25:04
|
2121844780
|
{
"authors": [
"bemidji3",
"codecov-commenter"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4692",
"repo": "cisco-open/fsoc",
"url": "https://github.com/cisco-open/fsoc/pull/275"
}
|
gharchive/pull-request
|
ORION-2525: fsoc solution zap + deprecate include-tags flag for fsoc ks commands
Description
We are adding a new solution command: fsoc solution zap. This will upload an empty version of a solution, removing all types and objects that are present in the solution. This will only work for non-stable tagged solutions. W
We have also marked the include-tags flag as hidden as this field is not in our public open api spec at this point of time so we should not expose it to our users yet.
Type of Change
[X] Bug Fix
[X] New Feature
[ ] Breaking Change
[ ] Refactor
[ ] Documentation
[ ] Other (please describe)
Checklist
[X] I have read the contributing guidelines
[X] Existing issues have been referenced (where applicable)
[X] I have verified this change is not present in other open pull requests
[X] Functionality is documented
[X] All code style checks pass
[X] New code contribution is covered by automated tests
[X] All new and existing tests pass
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Comparison is base (ba840f9) 26.88% compared to head (79617c1) 26.88%.
:exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality.
Additional details and impacted files
@@ Coverage Diff @@
## main #275 +/- ##
=======================================
Coverage 26.88% 26.88%
=======================================
Files 44 44
Lines 4564 4564
=======================================
Hits 1227 1227
Misses 3242 3242
Partials 95 95
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
2025-04-01T06:38:11.932467
| 2022-04-27T15:23:12
|
1217514039
|
{
"authors": [
"RichLogan",
"glhewett"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4693",
"repo": "cisco/gse",
"url": "https://github.com/cisco/gse/pull/5"
}
|
gharchive/pull-request
|
Fixed gse include path for use with FetchContent
Moving CMakeLists.txt up to the src/gse directory helps minimize the number of .. used in paths.
@glhewett Nice work, but I think we need the same change to src/common for this to work
fixed.
|
2025-04-01T06:38:12.076678
| 2021-04-10T09:44:01
|
855027920
|
{
"authors": [
"Melkor333",
"cevatkerim",
"ciur"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4694",
"repo": "ciur/papermerge",
"url": "https://github.com/ciur/papermerge/issues/364"
}
|
gharchive/issue
|
Add iOS/mobile device support
Is your feature request related to a problem? Please describe.
Papermerge is quite unusable since it does not support touch input. When using iPad or another device that does not support a right click in web interfaces, it results in poor user experience.
Describe the solution you'd like
A mobile/touch friendly interface would make Papermerge much more usable.
Describe alternatives you've considered
N/A
Additional context
N/A
@cevatkerim, thanks for opening this issue!
This may not be directly related with the above bug, but it's also related to mobile:
When using mobile (tested with Firefox 90.1.3 on Android) a lot of stuff is "hidden".
E.g. when I want to create a user, the whole box containing the two buttons to "Create" or "Cancel" isn't visible.
Or when opening a file it only shows the file itself, the metadata "box" isn't displayed, etc.
|
2025-04-01T06:38:12.136633
| 2022-01-07T03:05:54
|
1095937316
|
{
"authors": [
"ExperimentsInHonesty",
"bruceplai",
"cnk",
"fyliu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4696",
"repo": "civictechindex/CTI-website-frontend",
"url": "https://github.com/civictechindex/CTI-website-frontend/issues/1113"
}
|
gharchive/issue
|
Projects not showing up under individual org page for index contributors
Overview
For orgs that are tagged as index contributors in the Civic Tech Organizations page, when you click on some of them, no projects show up under the individual org page. As index contributors, they should have projects tagged with "civictechindex"
Action Items
[ ] Investigate where the disconnect is between orgs flagged as contributors and their tagged projects
[ ] Implement fix in front end if needed or notify PM where to update org data
Just some info from a quick search on github.com.
Code for Buffalo is the index contributor that returns no projects
http://civictechindex.org/organization/code-for-buffalo
Code For Buffalo does have a project with the "civictechindex" tag
https://github.com/search?p=4&q=topic%3Acivictechindex&type=Repositories
Code For Buffalo uses the tag "buffalo" in projects, rather than some variations like "hack-for-la" or "code-for-kc"
Here are the queries we're using to retrieve the organization's projects with "civictechindex". Notice we don't use "buffalo".
Here's a query that does return the correct result, but it uses the CodeForBuffalo github user (it works without the org:CodeForBuffalo that's also in this query)
https://github.com/search?q=org%3ACodeForBuffalo+topic%3Acivictechindex+user%3ACodeForBuffalo
There's a "github_user" field in the backend organization model that's available to use for this case. Or maybe just use the github_user instead of the organization tag variations. But I don't know if there are cases that require those.
We do want to encourage the proper use of affiliation tags. So we have the following rules
We don't list someone under their org in the contributors section unless they use one of the various of the tags that we accept. For example, we would not accept Buffalo because other orgs might use that. We will accept code-for-buffalo, codeforbuffalo, code4buffalo.
@cnk please see Bonnie's clarification of current affiliation tag issues above. Let me know if we need to discuss how to tackle them. Thanks
I think we need to clarify that "they need to use an affiliation tag" means "The organization needs their repositories tagged with a topic (that is GH's word for our 'tags') that is one of our recognized variations on their name".
I suspect several of our current 'contributors' will need to update their tagging - but it's hard to say if we don't have good data for the org tags. Before I modify our 'update_contributors' script with this new restriction, let's get the data import from #1036 done. Then the issues are likely to be useful / valid.
This should be split into new issues
[frontend] This sounds like a frontend change. We only store the orgs in the backend and not their repos. The frontend is what queries and displays the repos. This kind of negates point 2 below. This hidden repo behavior was what triggered the current issue to be created.
[frontend] We can query the repo's organization to see whether it belongs to codeforbuffalo or codeforamerica. This is what the linked pull request is doing. If the user clicks to see Code For Buffalo's repos, it does a query for topic:civictechindex and org:CodeForBuffalo. I think this solution addresses the current issue correctly.
[backend] This is what cnk is addressing above
[frontend] This is a frontend thing since it involves the tag generator. Run the queries in the frontend to display what's appropriate.
Here's the expected behavior discussed at the 1/27 meeting:
The backend script should NOT mark an affiliated org as index contributor unless it has a repo that contains BOTH 'civictechindex' and a proper org affiliation topic tag.
So, in the case of Code for Buffalo, rather than having it show up as index contributor and have an empty page, we would not want to show the organization at all if the index contributor filter is on. This means the backend should not check the index contributor flag.
Question/clarification: For non-affiliated orgs, we require only that the 'civictechindex' topic tag be present for the org to be an index contributor?
Progress - not much progress on front end since last update
Blockers - need to make instructions more clear for orgs to set their affiliation tags in their projects. Also awaiting data migration in #1036 and other back end logic updates for flagging orgs as contributors
Availability - 2 days this week
ETA - pending blockers
|
2025-04-01T06:38:12.141348
| 2023-03-28T16:58:40
|
1644341820
|
{
"authors": [
"elisekalstad",
"msprenke",
"sijiayam"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4697",
"repo": "civiform/civiform",
"url": "https://github.com/civiform/civiform/issues/4506"
}
|
gharchive/issue
|
Enable applicants to download and print completed applications
Is your feature request related to a problem? Please describe.
A group of Trusted Intermediaries mentioned that it's helpful to print the completed program application for a resident/client to take with them. Residents often want the physical form to take with them. They mentioned it can be helpful if the residents seeks services from multiple CBOs, they can take their application to the next TI.
Describe the solution you'd like
An option on the application confirmation page that allows the applicant (TI or resident) to download then print the completed application.
Additional context
It would be helpful if it included all completed questions as well as a printout of any uploaded documents. There may be data privacy concerns with the email option, would want to check with Privacy on this.
Done when
Resident or TI can completed application.
@swatkat1 is considering this feature as a good project for summer intern.
This was a feature requested during listening sessions with community-based orgs (digital navigators) on April 25th.
Possibly change to just download as PDF. Adobe can print from there and possibly open users email application.
For responsiveness, I recommend laying out the information side by side, so on mobile buttons can stack. Here's a mockup!
|
2025-04-01T06:38:12.150333
| 2022-04-25T09:22:47
|
1214220039
|
{
"authors": [
"chengchingwen",
"cjdoris"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4698",
"repo": "cjdoris/PythonIO.jl",
"url": "https://github.com/cjdoris/PythonIO.jl/issues/2"
}
|
gharchive/issue
|
Difference between existing packages?
Hi, I just notice this package and I'm wondering about what's the difference between this package and those existing packages like Pickle.jl and NPZ.jl? And maybe it's worth directly depending on them?
This is unregistered and very experimental!
Partly I don't particularly like the API of NPZ.jl and partly putting them in the same place allows some useful behaviour, like you can reuse the Python parser to implement reading npy files, and the pickle parser allows parsing a wider range of npy files, e.g. ones containing strings.
|
2025-04-01T06:38:12.175673
| 2019-10-31T14:52:33
|
515496204
|
{
"authors": [
"C47D",
"cjhdev"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4699",
"repo": "cjhdev/lora_device_lib",
"url": "https://github.com/cjhdev/lora_device_lib/issues/1"
}
|
gharchive/issue
|
State of the library and future plans
Hi,
I came across your library after asking about open source LoRaWan stack implementartions here.
We want to develop sensor nodes in a factory environment, we're planning to use STM32Lx devices, communicating over LoRaWan with a gateway. We're using Atollic TrueStudio (configuration code generated with STM32CubeMX) and I would like to include your library to the project, I tried to do so with lorawan-mac I i wasn't able to do so.
After searching LoRaWan stacks I think yours is the most sane one, but i would like to know what is your roadmap, maybe if you accept contributions, plans to implement Class B and C, etc.
Hello
Thanks, I started this project to learn the protocol and also because I felt the other implementations looked crazy.
This project is active but I am (probably) the only person using it. It would be great to get some more users and contributors that appreciate the style of this implementation over the alternatives. Before accepting contributions I think it would be necessary to lay down some guidelines so that no effort is wasted.
There is no roadmap, there could be one if there was serious interest. The reason I haven't done class B and C is because I have no need for these modes. The biggest challenge with implementing these modes (or indeed any new features) is verifying that they work correctly. Class B needs a lot more tooling than class A.
I would describe the quality as experimental. I don't think you should use it for anything serious at the moment. This could change if more people use it. I think it would also be good to run it through the LoRaWAN conformance test at some point.
I'm planning to release 0.1.7 in the next week or two. I'll see if I can also produce a list of what works well and what needs to be improved.
Hi,
I would describe the quality as experimental. I don't think you should use it for anything serious at the moment. This could change if more people use it. I think it would also be good to run it through the LoRaWAN conformance test at some point.
Thanks for the quick reply, to be honest I just been told that the plan A is to use modules commanded via AT commands, so I'm no longer in a hurry, at least not for work projects.
I'm still interested in using this library as a learning exercise and in personal projects, seems like I will be using LoRaWan for a while and, as you did when you started this project, I would use it to learn the protocol.
I'm planning to release 0.1.7 in the next week or two. I'll see if I can also produce a list of what works well and what needs to be improved.
I would like to propose the following, let me know if you agree. I can get more familiar with LoRaWan and maybe other implementations of it, then if you feel like it I can help with small and easy tasks of the to be implemented list.
So far I have been able to include LDL to my project and initialize boath Board and Radio, just noticed I must implement all the weak functions in lora_system, am I right? Maybe a porting guide would be nice for us who aren't familiar with LDL.
Regards
I think an AT command module is wise for proof of concept. You don't want to get bogged down in technical details before you know if the technology is right for the application.
Yes I agree a todo list is a good idea.
Yes I agree a porting guide would be helpful.
Some of the system interfaces will work just fine with the weak implementations, while others are show stoppers. Looking at the API documentation it occurs to me not all of the mandatory interfaces are marked as mandatory.
The arduino wrapper is a good reference for what is mandatory.
I should also mention that since I am in the UK I only ever use the setting for the EU_863_870 region. I have in the past emulated some of the other regions but not for some time.
I'm located in Mexico, will test the US_902_928 region.
Some of the system interfaces will work just fine with the weak implementations, while others are show stoppers. Looking at the API documentation it occurs to me not all of the mandatory interfaces are marked as mandatory.
I didn't took a deep look, but I assume the mandatory interfaces are those marked with @warning this function must be implemented on target for correct operation.
The arduino wrapper is a good reference for what is mandatory.
I will take a look at it, so far i haven't seen any timer being setup, while in loramac-node, they use the microcontroller RTC, is this only neccesary on Class B or C nodes?
I didn't took a deep look, but I assume the mandatory interfaces are those marked with @warning this function must be implemented on target for correct operation.
Yeah more or less.
I'm at master, should i also check the development branch?
Master is best. Development might not work properly.
so far i haven't seen any timer being setup
LDL has a bunch of internal timers that depend on the platform providing a free-running 32bit counter.
LDL_System_time() returns the counter value at any time
LDL_System_tps() returns the rate at which the counter increments (ticks per second)
LDL_System_eps() returns the error in ticks (error per second)
So on Arduino for example:
LDL_System_time() wraps micros()
LDL_System_tps() returns 1000000
LDL_System_eps() returns 5000 to account for a ceramic resonator
LDL has a bunch of internal timers that depend on the platform providing a free-running 32bit counter.
LDL_System_time() returns the counter value at any time
LDL_System_tps() returns the rate at which the counter increments (ticks per second)
LDL_System_eps() returns the error in ticks (error per second)
So on Arduino for example:
LDL_System_time() wraps micros()
LDL_System_tps() returns 1000000
LDL_System_eps() returns 5000 to account for a ceramic resonator
Haven't been on my work station, I'm using STM32 devices, I guess the HAL_GetTicks() function for LDL_System_time() should work, it should return the same value in LDL_System_tps() and I will calculate the value for LDL_System_eps()
Looking good.
I see the problem, you need to connect the radio "DIO" control lines. That part is missing from the example you are based on, it's not very clear.
You need to detect the DIO line(s) rising edge and then call LDL_MAC_interrupt(&mac, n, LDL_System_time()) where n is the index of the line (e.g. DIO0 is n == 0). You only need DIO0, DIO1, DIO2, and DIO3.
This is how the arduino wrapper does it. That function is called by an ISR for a particular control line. There is implementation specific logic but the important part is that I call LDL_MAC_interrupt().
If you use an interrupt, make sure to define LORA_SYSTEM_ENTER_CRITICAL and _LEAVE_CRITICAL. This should work:
#define LORA_SYSTEM_ENTER_CRITICAL(APP) volatile uint32_t primask = __get_PRIMASK();__disable_irq();
#define LORA_SYSTEM_ENTER_CRITICAL(APP) __set_PRIMASK(primask);
I recommend to put this in a header file, then define LORA_TARGET_INCLUDE to be the name of that file (e.g. -DLORA_TARGET_INCLUDE='"your_header.h"'). All the other LDL build options can go there.
You will need to also define:
LORA_ENABLE_SX1276
LORA_ENABLE_US_902_928
LORA_DISABLE_FULL_CODEC
I assume you've already done this somewhere I can't see. I mean, you can do it all from the makefile if you prefer.
Once you get that sorted you will probably find that the example sends too frequently since the US doesn't have a duty cycle limit. To slow it down you can either add a delay in your app, or set LDL_MAC_setAggregatedDutyCycle() to impose a global duty cycle limit. A setting of 2 will give you a 1% duty cycle limit.
Hi,
Thanks for taking a look at the project, just noticed i uploaded the keys of it, so i had to make the repo private :/, let me know if you want me to give you access to it.
I see the problem, you need to connect the radio "DIO" control lines. That part is missing from the example you are based on, it's not very clear.
You need to detect the DIO line(s) rising edge and then call LDL_MAC_interrupt(&mac, n, LDL_System_time()) where n is the index of the line (e.g. DIO0 is n == 0). You only need DIO0, DIO1, DIO2, and DIO3.
This is how the arduino wrapper does it. That function is called by an ISR for a particular control line. There is implementation specific logic but the important part is that I call LDL_MAC_interrupt().
Here's how i implemented it:
void HAL_GPIO_EXTI_Callback(uint16_t GPIO_Pin)
{
switch (GPIO_Pin) {
case DIO0_Pin:
LDL_MAC_interrupt(&mac, 0, LDL_System_time());
break;
case DIO1_Pin:
LDL_MAC_interrupt(&mac, 1, LDL_System_time());
break;
case DIO2_Pin:
LDL_MAC_interrupt(&mac, 2, LDL_System_time());
break;
case DIO3_Pin:
LDL_MAC_interrupt(&mac, 3, LDL_System_time());;
break;
default:
break;
}
}
If you use an interrupt, make sure to define LORA_SYSTEM_ENTER_CRITICAL and _LEAVE_CRITICAL. This should work:
#define LORA_SYSTEM_ENTER_CRITICAL(APP) volatile uint32_t primask = __get_PRIMASK();__disable_irq();
#define LORA_SYSTEM_LEAVE_CRITICAL(APP) __set_PRIMASK(primask);
I recommend to put this in a header file, then define LORA_TARGET_INCLUDE to be the name of that file (e.g. -DLORA_TARGET_INCLUDE='"your_header.h"'). All the other LDL build options can go there.
My custom header file is named LDL_options.h and goes as follows:
#ifndef LDL_OPTIONS_H_
#define LDL_OPTIONS_H_
#include "cmsis_gcc.h"
// http://stm32f4-discovery.net/2015/06/how-to-properly-enabledisable-interrupts-in-arm-cortex-m/
static volatile uint32_t primask = 0;
#define LORA_SYSTEM_ENTER_CRITICAL(APP) do { primask = __get_PRIMASK(); __disable_irq(); } while (0);
#define LORA_SYSTEM_LEAVE_CRITICAL(APP) __set_PRIMASK(primask);
#endif /* LDL_OPTIONS_H_ */
You will need to also define:
LORA_ENABLE_SX1276
LORA_ENABLE_US_902_928
LORA_DISABLE_FULL_CODEC
I assume you've already done this somewhere I can't see. I mean, you can do it all from the makefile if you prefer.
Yep, I define those symbols on the IDE, but now i think it's better to have them on the LDL_options.h file so others doesn't need to load the project into the IDE just to see the configuration. So the LDL_options.h file ends up like this:
#ifndef LDL_OPTIONS_H_
#define LDL_OPTIONS_H_
#include "cmsis_gcc.h"
// http://stm32f4-discovery.net/2015/06/how-to-properly-enabledisable-interrupts-in-arm-cortex-m/
#define LORA_ENABLE_SX1276
#define LORA_ENABLE_US_902_928
#define LORA_DISABLE_FULL_CODEC
// #define LORA_TARGET_INCLUDE /* See lora_platform */
static volatile uint32_t primask = 0;
#define LORA_SYSTEM_ENTER_CRITICAL(APP) do { primask = __get_PRIMASK(); __disable_irq(); } while (0);
#define LORA_SYSTEM_LEAVE_CRITICAL(APP) __set_PRIMASK(primask);
#endif /* LDL_OPTIONS_H_ */
Once you get that sorted you will probably find that the example sends too frequently since the US doesn't have a duty cycle limit. To slow it down you can either add a delay in your app, or set LDL_MAC_setAggregatedDutyCycle() to impose a global duty cycle limit. A setting of 7 will give you a ~1% duty cycle limit.
edit: made mistake on global duty cycle
Will edit the comment once i get some results later today.
Thanks for the help and patience.
10km is probably too far for initial debug. Too close (i.e. sitting right next to the gateway) can also be an issue.
It's often useful to print time (i.e. LDL_System_time()) with each event for double checking timing.
Hi,
Thanks for the tips. I had to modify the logging macros (replacing PRIu32 for %d) because of my underlying functions. Everything else seems to be working as expected. Will report back when I get a gateway.
|
2025-04-01T06:38:12.183351
| 2016-06-24T14:24:09
|
162158572
|
{
"authors": [
"mafrosis",
"thicklord"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4700",
"repo": "cjlucas/rtorrent-python",
"url": "https://github.com/cjlucas/rtorrent-python/issues/29"
}
|
gharchive/issue
|
Docs for 1.0-alpha?
Are these available somewhere convenient?
Are these available anywhere at all ?
|
2025-04-01T06:38:12.278210
| 2021-07-30T16:00:23
|
956883400
|
{
"authors": [
"alex60217101990",
"claffin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4701",
"repo": "claffin/cloudproxy",
"url": "https://github.com/claffin/cloudproxy/issues/49"
}
|
gharchive/issue
|
always destroy instances
Expected Behavior
an GCP and DO cloud providers always destroy instances:
2021-07-30 15:59:33.646 | INFO | uvicorn.protocols.http.h11_impl:send:461 - <IP_ADDRESS>:60922 - "GET /destroy HTTP/1.1" 200
2021-07-30 15:59:36.640 | INFO | uvicorn.protocols.http.h11_impl:send:461 - <IP_ADDRESS>:60922 - "GET /destroy HTTP/1.1" 200
2021-07-30 15:59:30.937 | INFO | uvicorn.protocols.http.h11_impl:send:461 - <IP_ADDRESS>:60924 - "GET /destroy HTTP/1.1" 200
The UI regularly calls the /destroy endpoint to get a list of all the proxy instances pending deletion. This get request you are seeing is just that and it's not actually deleting the proxies.
If you start Cloudproxy and don't open the UI, then you'll notice those requests aren't there.
|
2025-04-01T06:38:12.280171
| 2023-09-21T13:01:08
|
1906923693
|
{
"authors": [
"keighrim"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4702",
"repo": "clamsproject/app-swt-detection",
"url": "https://github.com/clamsproject/app-swt-detection/issues/2"
}
|
gharchive/issue
|
image preprocessing for training
For the first batches of training, we plan to use video frames extracted from AAPB videos. However, there's concerns around how to "normalize" different image size and aspect ratios from videos from different decades. This thread is to discuss how we implement the normalizing strategies.
We decided to go only with 4:3 videos (before HD broadcasting era, circa early 2000s) in the early rounds of annotation.
Note that as long as we are using the pre-trained backbone weights, those weights in torch-vision package come with their own preprocessing code. That is, we can add some additional preprocessing based on some domain knowledge before the torch-shipped preprocessing. However, as of now we don't see a lot of needs for doing so.
|
2025-04-01T06:38:12.296843
| 2024-03-16T16:10:54
|
2190117065
|
{
"authors": [
"HighCommander4",
"jcarrete5"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4703",
"repo": "clangd/clangd",
"url": "https://github.com/clangd/clangd/issues/1975"
}
|
gharchive/issue
|
clangd "--query-driver" option doesn't follow symlinks
When I run clangd with the --query-driver option set to a path containing a symlink, the argument has no effect and I still have missing system headers. In the following case, <iostream> can't be found.
#include <iostream>
int main(int argc, char* argv[])
{
std::cout << "hi\n";
return 0;
}
I can work around this issue by using the complete path without any symlinks.
Within the logs, the oe-workdir path item is a symlink to another location.
Logs
Please attach the clangd stderr log if you can. (Usually available from the editor)
If possible, run with --log=verbose - note that the logs will include the contents of open files!
If this is a crash, try to put llvm-symbolizer on your PATH per the troubleshooting instructions.
[START][2024-03-16 12:09:47] LSP logging initiated
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'I[12:09:47.584] clangd version 17.0.6\nI[12:09:47.584] Features: linux\nI[12:09:47.584] PID: 102869\nI[12:09:47.584] Working directory: /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight\nI[12:09:47.584] argv[0]: /usr/bin/clangd\nI[12:09:47.584] argv[1]: --log=verbose\nI[12:09:47.584] argv[2]: --query-driver=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/oe-workdir/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++\nV[12:09:47.584] User config file is /home/jasonc/.config/clangd/config.yaml\nI[12:09:47.584] Starting LSP over stdin/stdout\nV[12:09:47.584] <<< {"id":1,"jsonrpc":"2.0","method":"initialize","params":{"capabilities":{"offsetEncoding":["utf-8","utf-16"],"textDocument":{"callHierarchy":{"dynamicRegistration":false},"codeAction":{"codeActionLiteralSupport":{"codeActionKind":{"valueSet":["","quickfix","refactor","refactor.extract","refactor.inline","refactor.rewrite","source","source.organizeImports"]}},"dataSupport":true,"dynamicRegistration":false,"isPreferredSupport":true,"resolveSupport":{"properties":["edit"]}},"completion":{"completionItem":{"commitCharactersSupport":true,"deprecatedSupport":true,"documentationFormat":["markdown","plaintext"],"insertReplaceSupport":true,"insertTextModeSupport":{"valueSet":[1,2]},"labelDetailsSupport":true,"preselectSupport":true,"resolveSupport":{"properties":["documentation","detail","additionalTextEdits","sortText","filterText","insertText","textEdit","insertTextFormat","insertTextMode"]},"snippetSupport":true,"tagSupport":{"valueSet":[1]}},"completionItemKind":{"valueSet":[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25]},"completionList":{"itemDefaults":["commitCharacters","editRange","insertTextFormat","insertTextMode","data"]},"contextSupport":true,"dynamicRegistration":false,"editsNearCursor":true,"insertTextMode":1},"declaration":{"linkSupport":true},"definition":{"linkSupport":true},"documentHighlight":{"dynamicRegistration":false},"documentSymbol":{"dynamicRegistration":false,"hierarchicalDocumentSymbolSupport":true,"symbolKind":{"valueSet":[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]}},"hover":{"contentFormat":["markdown","plaintext"],"dynamicRegistration":false},"implementation":{"linkSupport":true},"publishDiagnostics":{"relatedInformation":true,"tagSupport":{"valueSet":[1,2]}},"references":{"dynamicRegistration":false},"rename":{"dynamicRegistration":false,"prepareSupport":true},"semanticTokens":{"augmentsSyntaxTokens":true,"dynamicRegistration":false,"formats":["relative"],"multilineTokenSupport":false,"overlappingTokenSupport":true,"requests":{"full":{"delta":true},"range":false},"serverCancelSupport":false,"tokenModifiers":["declaration","definition","readonly","static","deprecated","abstract","async","modification","documentation","defaultLibrary"],"tokenTypes":["namespace","type","class","enum","interface","struct","typeParameter","parameter","variable","property","enumMember","event","function","method","macro","keyword","modifier","comment","string","number","regexp","operator","decorator"]},"signatureHelp":{"dynamicRegistration":false,"signatureInformation":{"activeParameterSupport":true,"documentationFormat":["markdown","plaintext"],"parameterInformation":{"labelOffsetSupport":true}}},"synchronization":{"didSave":true,"dynamicRegistration":false,"willSave":true,"willSaveWaitUntil":true},"typeDefinition":{"linkSupport":true}},"window":{"showDocument":{"support":true},"showMessage":{"messageActionItem":{"additionalPropertiesSupport":false}},"workDoneProgress":true},"workspace":{"applyEdit":true,"configuration":true,"didChangeWatchedFiles":{"dynamicRegistration":false,"relativePatternSupport":true},"semanticTokens":{"refreshSupport":true},"symbol":{"dynamicRegistration":false,"hierarchicalWorkspaceSymbolSupport":true,"symbolKind":{"valueSet":[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]}},"workspaceEdit":{"resourceOperations":["rename","create","delete"]},"workspaceFolders":true}},"clientInfo":{"name":"Neovim","version":"0.9.5"},"initializationOptions":{},"processId":102865,"rootPath":"/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight","rootUri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight","trace":"off","workspaceFolders":[{"name":"/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight","uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight"}]}}\n\nI[12:09:47.584] <-- initialize(1)\nI[12:09:47.585] --> reply:initialize(1) 0 ms\nV[12:09:47.585] >>> {"id":1,"jsonrpc":"2.0","result":{"capabilities":{"astProvider":true,"callHierarchyProvider":true,"clangdInlayHintsProvider":true,"codeActionProvider":{"codeActionKinds":["quickfix","refactor","info"]},"compilationDatabase":{"automaticReload":true},"completionProvider":{"resolveProvider":false,"triggerCharacters":[".","<",">",":","\\"","/","*"]},"declarationProvider":true,"definitionProvider":true,"documentFormattingProvider":true,"documentHighlightProvider":true,"documentLinkProvider":{"resolveProvider":false},"documentOnTypeFormattingProvider":{"firstTriggerCharacter":"\\n","moreTriggerCharacter":[]},"documentRangeFormattingProvider":true,"documentSymbolProvider":true,"executeCommandProvider":{"commands":["clangd.applyFix","clangd.applyTweak"]},"foldingRangeProvider":true,"hoverProvider":true,"implementationProvider":true,"inactiveRegionsProvider":true,"inlayHintProvider":true,"memoryUsageProvider":true,"referencesProvider":true,"renameProvider":{"prepareProvider":true},"selectionRangeProvider":true,"semanticTokensProvider":{"full":{"delta":true},"legend":{"tokenModifiers":["declaration","definition","deprecated","deduced","readonly","static","abstract","virtual","dependentName","defaultLibrary","usedAsMutableReference","usedAsMutablePointer","constructorOrDestructor","userDefined","functionScope","classScope","fileScope","globalScope"],"tokenTypes":["variable","variable","parameter","function","method","function","property","variable","class","interface","enum","enumMember","type","type","unknown","namespace","typeParameter","concept","type","macro","modifier","operator","bracket","label","comment"]},"range":false},"signatureHelpProvider":{"triggerCharacters":["(",")","{","}","<",">",","]},"standardTypeHierarchyProvider":true,"textDocumentSync":{"change":2,"openClose":true,"save":true},"typeDefinitionProvider":true,"typeHierarchyProvider":true,"workspaceSymbolProvider":true},"offsetEncoding":"utf-8","serverInfo":{"name":"clangd","version":"clangd version 17.0.6 linux x86_64-pc-linux-gnu"}}}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.699] <<< {\"jsonrpc\":\"2.0\",\"method\":\"initialized\",\"params\":{}}\n\nI[12:09:47.699] <-- initialized\nV[12:09:47.699] <<< {\"jsonrpc\":\"2.0\",\"method\":\"textDocument/didOpen\",\"params\":{\"textDocument\":{\"languageId\":\"cpp\",\"text\":\"/*\\n * SPDX-License-Identifier: GPL-3.0-or-later\\n *\\n * Copyright (C) 2023 Jason Carrete\\n *\\n * This file is part of Flight Controller.\\n *\\n * Flight Controller is free software: you can redistribute it and/or modify\\n * it under the terms of the GNU General Public License as published by\\n * the Free Software Foundation, either version 3 of the License, or\\n * (at your option) any later version.\\n *\\n * This program is distributed in the hope that it will be useful,\\n * but WITHOUT ANY WARRANTY; without even the implied warranty of\\n * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\\n * GNU General Public License for more details.\\n *\\n * You should have received a copy of the GNU General Public License\\n * along with this program. If not, see <https://www.gnu.org/licenses/>.\\n */\\n\\n#include \\\"appinfo.h\\\"\\n#include \\\"version.h\\\"\\n\\n#include <iostream>\\n\\nnamespace ffd = freeflight_daemon;\\n\\nint main(int argc, char* argv[])\\n{\\n std::cout << ffd::get_app_name() << ' ' << freeflight::get_version() << ' '\\n << freeflight::get_name() << '\\\\n';\\n return 0;\\n}\\n\",\"uri\":\"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\",\"version\":0}}}\n\nI[12:09:47.699] <-- textDocument/didOpen\nV[12:09:47.699] System include extraction: driver clang expanded to /usr/bin/clang\nV[12:09:47.699] System include extraction: not allowed driver /usr/bin/clang\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.700] <<< {"id":2,"jsonrpc":"2.0","method":"textDocument/semanticTokens/full","params":{"textDocument":{"uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp"}}}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.700] <-- textDocument/semanticTokens/full(2)\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.700] config note at /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd:1:0: Parsing config fragment\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.700] config note at /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd:1:0: Parsed 1 fragments from file\nV[12:09:47.700] Config fragment: compiling /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd:1 -> 0x0000724F50003730 (trusted=false)\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.701] --> textDocument/publishDiagnostics\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.701] >>> {"jsonrpc":"2.0","method":"textDocument/publishDiagnostics","params":{"diagnostics":[],"uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/.clangd"}}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.701] Loaded compilation database from /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/compile_commands.json\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.701] Broadcasting compilation database from /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.701] System include extraction: not allowed driver /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.701] ASTWorker building file /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0 with command \n[/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git]\n/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++ --target=aarch64-poky-linux --driver-mode=g++ --sysroot=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot -I/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd -I/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/freeflight/public -mcpu=cortex-a57 -march=armv8-a+crc -mbranch-protection=standard -fstack-protector-strong -O2 -D_FORTIFY_SOURCE=2 -Wformat -Wformat-security -Werror=format-security --sysroot=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot -O2 -pipe -g -feliminate-unused-debug-types -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native= -fvisibility-inlines-hidden -O2 -g -DNDEBUG -std=gnu++20 -o src/flightd/CMakeFiles/flightd.dir/main.cpp.o -c -resource-dir=/usr/lib/clang/17 -- /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.702] Loaded compilation database from /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git/compile_commands.json\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.702] --> window/workDoneProgress/create(0)\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.702] >>> {"id":0,"jsonrpc":"2.0","method":"window/workDoneProgress/create","params":{"token":"backgroundIndexProgress"}}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.702] Enqueueing 1 commands for indexing\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.702] Driver produced command: cc1 -cc1 -triple aarch64-poky-linux -fsyntax-only -disable-free -clear-ast-before-backend -disable-llvm-verifier -discard-value-names -main-file-name main.cpp -mrelocation-model pic -pic-level 2 -pic-is-pie -mframe-pointer=non-leaf -fmath-errno -ffp-contract=on -fno-rounding-math -mconstructor-aliases -funwind-tables=2 -target-cpu cortex-a57 -target-feature +neon -target-feature +v8a -target-feature +crc -target-abi aapcs -msign-return-address=non-leaf -msign-return-address-key=a_key -mbranch-target-enforce -debug-info-kind=constructor -dwarf-version=5 -debugger-tuning=gdb -fcoverage-compilation-dir=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git -resource-dir /usr/lib/clang/17 -I /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd -I /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/freeflight/public -D _FORTIFY_SOURCE=2 -D NDEBUG -isysroot /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot -internal-isystem /usr/lib/clang/17/include -internal-isystem /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot/usr/local/include -internal-externc-isystem /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot/include -internal-externc-isystem /home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot/usr/include -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fmacro-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -O2 -Wformat -Wformat-security -Werror=format-security -std=gnu++20 -fdeprecated-macro -fdebug-compilation-dir=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/freeflight-1.0+git=/usr/src/debug/freeflight/1.0+git-r0 -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot= -fdebug-prefix-map=/home/jasonc/projects/quadcopter/albatros/build/tmp/work/cortexa57-poky-linux/freeflight/1.0+git/recipe-sysroot-native= -ferror-limit 19 -fvisibility-inlines-hidden -stack-protector 2 -fno-signed-char -fgnuc-version=4.2.1 -fno-implicit-modules -fcxx-exceptions -fexceptions -vectorize-loops -vectorize-slp -no-round-trip-args -target-feature -fmv -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -x c++ /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.702] Building first preamble for /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0\nV[12:09:47.702] BackgroundIndex: building version 1 after loading index from disk\nV[12:09:47.702] <<< {"id":0,"jsonrpc":"2.0","result":null}\n\nI[12:09:47.702] <-- reply(0)\nI[12:09:47.702] --> $/progress\nV[12:09:47.702] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"begin","percentage":0,"title":"indexing"}}}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.702] BackgroundIndex: serving version 1 (50980 bytes)\nI[12:09:47.702] --> $/progress\nV[12:09:47.703] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"report","message":"0/1","percentage":0}}}\n\nI[12:09:47.703] --> $/progress\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.703] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"report","message":"0/1","percentage":0}}}\n\nI[12:09:47.703] --> $/progress\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.703] >>> {"jsonrpc":"2.0","method":"$/progress","params":{"token":"backgroundIndexProgress","value":{"kind":"end"}}}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.716] Built preamble of size 529320 for file /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0 in 0.01 seconds\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'I[12:09:47.717] --> workspace/semanticTokens/refresh(1)\nV[12:09:47.717] >>> {"id":1,"jsonrpc":"2.0","method":"workspace/semanticTokens/refresh","params":null}\n\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.717] <<< {"jsonrpc":"2.0","method":"$/cancelRequest","params":{"id":2}}\n\nI[12:09:47.717] <-- $/cancelRequest\nV[12:09:47.717] indexed preamble AST for /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0:\n symbol slab: 5 symbols, 5376 bytes\n ref slab: 0 symbols, 0 refs, 128 bytes\n relations slab: 0 relations, 24 bytes\nV[12:09:47.717] <<< {"id":3,"jsonrpc":"2.0","method":"textDocument/semanticTokens/full","params":{"textDocument":{"uri":"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp"}}}\n\nI[12:09:47.717] <-- textDocument/semanticTokens/full(3)\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.717] <<< {"id":1,"jsonrpc":"2.0","result":null}\n\nI[12:09:47.717] <-- reply(1)\nV[12:09:47.717] Build dynamic index for header symbols with estimated memory usage of 22004 bytes\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" 'V[12:09:47.724] Trying to fix unresolved name "cout" in scopes: [std::]\n'
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.724] Dex query tree: false\nV[12:09:47.724] Dex query tree: false\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "E[12:09:47.725] IncludeCleaner: Failed to get an entry for resolved path : No such file or directory\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.726] indexed file AST for /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp version 0:\n symbol slab: 2 symbols, 4680 bytes\n ref slab: 2 symbols, 2 refs, 4272 bytes\n relations slab: 0 relations, 24 bytes\nV[12:09:47.726] Build dynamic index for main-file symbols with estimated memory usage of 12040 bytes\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "I[12:09:47.726] --> textDocument/publishDiagnostics\n"
[ERROR][2024-03-16 12:09:47] .../vim/lsp/rpc.lua:734 "rpc" "/usr/bin/clangd" "stderr" "V[12:09:47.726] >>> {\"jsonrpc\":\"2.0\",\"method\":\"textDocument/publishDiagnostics\",\"params\":{\"diagnostics\":[{\"code\":\"pp_file_not_found\",\"message\":\"'iostream' file not found\",\"range\":{\"end\":{\"character\":19,\"line\":24},\"start\":{\"character\":9,\"line\":24}},\"relatedInformation\":[],\"severity\":1,\"source\":\"clang\"},{\"code\":\"undeclared_var_use\",\"message\":\"Use of undeclared identifier 'std'\",\"range\":{\"end\":{\"character\":7,\"line\":30},\"start\":{\"character\":4,\"line\":30}},\"relatedInformation\":[],\"severity\":1,\"source\":\"clang\"},{\"code\":\"misc-unused-alias-decls\",\"codeDescription\":{\"href\":\"https://clang.llvm.org/extra/clang-tidy/checks/misc/unused-alias-decls.html\"},\"message\":\"Namespace alias decl 'ffd' is unused (fix available)\",\"range\":{\"end\":{\"character\":0,\"line\":27},\"start\":{\"character\":0,\"line\":26}},\"relatedInformation\":[],\"severity\":2,\"source\":\"clang-tidy\",\"tags\":[1]}],\"uri\":\"file:///home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\",\"version\":0}}\n\nI[12:09:47.726] --> reply:textDocument/semanticTokens/full(2) 26 ms, error: Task was cancelled.\nV[12:09:47.726] >>> {\"error\":{\"code\":-32800,\"message\":\"Request cancelled\"},\"id\":2,\"jsonrpc\":\"2.0\"}\n\nV[12:09:47.726] ASTWorker running SemanticHighlights on version 0 of /home/jasonc/projects/quadcopter/albatros/build/workspace/sources/freeflight/src/flightd/main.cpp\nI[12:09:47.726] --> reply:textDocument/semanticTokens/full(3) 8 ms\nV[12:09:47.726] >>> {\"id\":3,\"jsonrpc\":\"2.0\",\"result\":{\"data\":[26,10,3,15,65537,0,6,17,15,131072,2,4,4,3,131075,0,9,4,2,16387,0,12,4,2,16387],\"resultId\":\"1\"}}\n\n"
System information
Output of clangd --version:
clangd version 17.0.6
Features: linux
Platform: x86_64-pc-linux-gnu
Editor/LSP plugin:
Neovim/nvim-lspconfig
Operating system:
Arch Linux
Kernel version: 6.7.9
Duplicate of #1605
|
2025-04-01T06:38:12.304715
| 2020-06-16T11:26:52
|
639588308
|
{
"authors": [
"HighCommander4",
"kadircet",
"lh123",
"njames93"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4704",
"repo": "clangd/clangd",
"url": "https://github.com/clangd/clangd/issues/433"
}
|
gharchive/issue
|
Include completion fails at the end of file
clangd version: clangd version 11.0.0 (https://github.com/llvm/llvm-project.git 3badd17b6989621b5aa2732800f697dabbda034d)
OS: Windows 10
I think this issue still exists. @HighCommander4
Originally posted by @lh123 in https://github.com/clangd/clangd/issues/38#issuecomment-643115928
You're right, I see the same issue (global completion instead of include completions).
I think what happened when I was testing #38, is that I typed in #include "llvm/Sup manually. When I type the " character, my editor auto-inserts a matching closing quote, so the actual test case I was testing was:
#include "llvm/Sup^"
(not the presence of the closing quote), which works fine.
I can't seem to reproduce this, neither on LLVM or a dummy project.
Could you please share clangd logs ?
It reproduces if the file has no code in and the include brackets are unbalanced and followed by eof
It reproduces if the file has no code in and the include brackets are unbalanced and followed by eof
Interesting I thought I had fixed this one... Well good thing is, this means at least I know the fix :D
Sent out https://reviews.llvm.org/D95419
Interesting I thought I had fixed this one... Well good thing is, this means at least I know the fix :D
Sent out https://reviews.llvm.org/D95419
Works like a charm now on my end, thanks.
Works like a charm now on my end, thanks.
|
2025-04-01T06:38:12.317299
| 2020-03-23T01:36:40
|
585869355
|
{
"authors": [
"chmanie",
"fannheyward",
"sam-mccall",
"stokhos"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4705",
"repo": "clangd/coc-clangd",
"url": "https://github.com/clangd/coc-clangd/issues/25"
}
|
gharchive/issue
|
Clangd formatting issue?
Hi:
I notice that clang has formatting feature, but I don't know how to set it up. Is there any docs for this?
best regards
Peiyun Jin
clangd is using clang-format to do formatting, https://clangd.llvm.org/features.html#formatting , you can set .clang-format for style options.
Hi I have installed coc-clangd, have .clang-format file in the project folder, and set
"coc.preferences.formatOnSaveFiletypes": ["cpp"],
"coc.preferences.formatOnSave": true,
But neither :Format nor format on save worked.
Could you tell me how can fix this?
Best
I think the problem is caused by this line "clangd.arguments": ["-Wall", "-Werror", "-std=c++17"] in above in my coc-settings.json. After adding this line into my coc-settings.json. The language server stops working for some reason.
{
"coc.preferences.formatOnSaveFiletypes": ["cpp"],
"coc.preferences.formatOnSave": true,
"clangd.arguments": ["-Wall", "-Werror", "-std=c++17"],
"languageserver": {
"python": {
"command": "python",
"args": [
"-mpyls",
"-vv",
"--log-file",
"/tmp/lsp_python.log"
],
"trace.server": "verbose",
"filetypes": [
"python"
],
"settings": {
"pyls": {
"enable": true,
"trace": {
"server": "verbose"
},
"commandPath": "",
"configurationSources": [
"pycodestyle"
],
"plugins": {
"jedi_completion": {
"enabled": true
},
"jedi_hover": {
"enabled": true
},
"jedi_references": {
"enabled": true
},
"jedi_signature_help": {
"enabled": true
},
"jedi_symbols": {
"enabled": true,
"all_scopes": true
},
"mccabe": {
"enabled": true,
"threshold": 15
},
"preload": {
"enabled": true
},
"pycodestyle": {
"enabled": true
},
"pydocstyle": {
"enabled": false,
"match": "(?!test_).*\\.py",
"matchDir": "[^\\.].*"
},
"pyflakes": {
"enabled": true
},
"rope_completion": {
"enabled": true
},
"yapf": {
"enabled": true
}
}
}
}
}
}
}```
Look like it's caused by server, @sam-mccall can you look into this?
I think the problem is caused by this line "clangd.arguments": ["-Wall", "-Werror", "-std=c++17"] in above in my coc-settings.json. After adding this line into my coc-settings.json. The language server stops working for some reason.
The reason is clangd.arguments is extra flags to pass to clangd, and those are not valid clangd arguments:
$ clangd -Wall -Werror -std=c++17
clangd: Unknown command line argument '-Wall'. Try: 'clangd --help'
clangd: Did you mean '--help'?
clangd: Unknown command line argument '-Werror'. Try: 'clangd --help'
clangd: Did you mean '--color'?
clangd: Unknown command line argument '-std=c++17'. Try: 'clangd --help'
clangd: Did you mean '--log=c++17'?
If you want to set the flags for parsing your code, this is configured using compile_commands.json or compile_flags.txt: https://clangd.llvm.org/installation.html#project-setup
@sam-mccall Hi Sam, Is there a way to enable clang-tidy and -j inside coc-settings.json? Or how can I enable them with vim?
best regards
Sure: "clangd.arguments": ["-j=3", "-clang-tidy-checks=bugprone-*"] - those really are clangd flags.
Similar to clang-format, for finer-grained clang-tidy config you should use the standard .clang-tidy config file in your source tree, clangd should respect it.
@sam-mccall Do I need to enable clang-tide first, "clangd.arguments": ["-j=3", "-clang-tide=true","-clang-tidy-checks=bugprone-*"]?
What's clangd --version?
It's on by default as of clangd 9.
It may barely work in clangd 8, it was experiencing and off by default.
I'm using fedora 32 beta. Clangd is version 10.
With clangd 10 you shouldn't have to explicitly enable clang tidy.
On Sat, Mar 28, 2020, 7:30 PM Stokhos<EMAIL_ADDRESS>wrote:
I'm using fedora 32 beta. Clangd is version 10.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/clangd/coc-clangd/issues/25#issuecomment-605500662,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AAEGBAMIBI3LZ3LECRB7SNTRJY65TANCNFSM4LRQVAMQ
.
Sorry to add to this but it seems related. I am very confused about what this plugin can do. Should I be able to do formatting? Issuing a coc :Format command doesn't do anything for me. Would then clang-format be used or clang-tidy with the --fix option (or both)?
|
2025-04-01T06:38:12.321682
| 2019-01-31T08:06:54
|
405125196
|
{
"authors": [
"clangen",
"ibrokemypie"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4706",
"repo": "clangen/musikcube",
"url": "https://github.com/clangen/musikcube/issues/240"
}
|
gharchive/issue
|
Ogg Opus Files Dont Show
On macOS Mojave, installed via homebrew. MP3 files show fine, but the majority of my library which is in ogg opus just doesnt show in the library viewer.
Oh, I see, tag parsing failed? That's possible, I don't have tons of Opus files. Is there any way you can share a file with tags that don't seem to be working?
sure, try this one
01 Tree Village.ogg.zip
Ah! So... taglib (our tag parsing library) sees the .ogg extension and assumes it's a vorbis file, then fails when it's not. When this happens, we should probably detect the error and then check to see if it's an opus file.
Are there any other file formats that commonly have an .ogg extension that may not be vorbis? :thinking:
Got it. I modified the taglib parser reader plugin so it can detect opus files in an ogg container -- your example file seems to parse fine now! I also made it such that as soon as I can get examples of other formats in ogg containers it should be trivial to add support to them as well.
seems to be fixed after re-adding the library on latest head, thanks!
Thanks for confirming the fix!
|
2025-04-01T06:38:12.346644
| 2021-05-21T09:52:44
|
897884262
|
{
"authors": [
"joaopaulovieira",
"leticiafernandes",
"pedrochamberlain",
"unique1o1"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4707",
"repo": "clappr/clappr",
"url": "https://github.com/clappr/clappr/issues/2013"
}
|
gharchive/issue
|
Browser starts to lag after playing 3-5 videos when using player.configure to update the source url
Browser:
Firefox 88.0.1
OS:
Pop OS 20.04 LTS
Clappr Version:
latest from http://cdn.clappr.io/latest/clappr.min.js
Steps to reproduce:
play 3-5 video one after another.
I was expecting to have a smooth playback but instead it started lagging(my browser)
Did you try to reproduce this issue at http://cdn.clappr.io/
No
I'm trying to update the video source using the following code:
// [...]
player.configure({
source: 'another-url',
// ...
});
which updates the video but after playing around 3-5 videos one after another the browser starts to lag.
I also tried to reproduce this issue on multiple devices, the same issue occurred.
NOTE : source URL is an HLS video encrypted with AES-128
Did you try to reproduce this issue at http://cdn.clappr.io/
No
Can't you reproduce the reported behavior or just haven't tried?
I also tried to reproduce this issue on multiple devices, the same issue occurred.
Yes I have but not on http://cdn.clappr.io/
Yes, I understand that you have tested it on other devices. The recommendation to test on http://cdn.clappr.io/ (specifically on http://cdn.clappr.io/demo) is that it becomes a common point that we can also validate on our side.
I was unable to reproduce this issue. Can you generate any evidence of this? (CPU / Memory consumption of the tab that is running Clappr, some video showing the problem visually)
Hi @unique1o1, is the problem still happening? Can you generate any evidence of this? (CPU / Memory consumption of the tab that is running Clappr, some video showing the problem visually)
I'm closing this issue due to inactivity. If needed, please feel free to reopen it.
|
2025-04-01T06:38:12.348139
| 2015-10-25T03:38:44
|
113204453
|
{
"authors": [
"leandromoreira",
"polaris-zx"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4708",
"repo": "clappr/clappr",
"url": "https://github.com/clappr/clappr/issues/609"
}
|
gharchive/issue
|
no video on galaxy s5
There is no video on Galaxy s5 when a live video or a replay is playing. But the video is diaplayed after you click the progress bar.
only the audio.
are you still facing this issue @polaris-zx ? also, did you test with different stream sources? is it only with hls ?
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.