added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:37:03.685506
| 2017-04-11T01:57:03
|
220817802
|
{
"authors": [
"DomT4",
"MikeMcQuaid"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1255",
"repo": "Homebrew/brew",
"url": "https://github.com/Homebrew/brew/pull/2474"
}
|
gharchive/pull-request
|
xcode: 8.3.1 is latest version.
Our CI images are already updated with it.
when "10.12" then "802.0.38"
Is out-of-date FWIW. Latest CLT is:
~> /Library/Developer/CommandLineTools/usr/bin/clang --version
Apple LLVM version 8.1.0 (clang-802.0.38)
Latest Xcode is:
~> clang --version
Apple LLVM version 8.1.0 (clang-802.0.41)
I can't read, ignore the comment I posted last night 😓.
|
2025-04-01T06:37:03.691045
| 2017-09-25T17:46:56
|
260360336
|
{
"authors": [
"maxim-belkin"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1256",
"repo": "Homebrew/brew",
"url": "https://github.com/Homebrew/brew/pull/3209"
}
|
gharchive/pull-request
|
install.rb: report caveats for installed formulae
Pull Request template:
[x] Have you followed the guidelines in our Contributing document?
[x] Have you checked to ensure there aren't other open Pull Requests for the same change?
[x] Have you added an explanation of what your changes do and why you'd like us to include them?
[ ] Have you written new tests for your changes? Here's an example.
[ ] Have you successfully run brew tests with your changes locally?
Detailed explanation:
When installing multiple formulae, caveats might get lost in the normal
output that accompanies normal installation. This PR reminds the user of
all the caveats at the end of the installation. Caveats are repeated only if
there are more than one formulae being installed. Following calls to
brew install <formula-with-caveats> does not result in caveat message
being printed again.
Example:
$ brew install go python
...
[snip]
...
==> Please take a note of the caveats for the following installed formulae:
==> Caveats for "go":
A valid GOPATH is required to use the `go get` command.
If $GOPATH is not specified, $HOME/go will be used by default:
https://golang.org/doc/code.html#GOPATH
You may wish to add the GOROOT-based install location to your PATH:
export PATH=$PATH:/usr/local/opt/go/libexec/bin
==> Caveats for "python":
This formula installs a python2 executable to /usr/local/bin.
If you wish to have this formula's python executable in your PATH then add
the following to ~/.bash_profile:
export PATH="/usr/local/opt/python/libexec/bin:$PATH"
Pip and setuptools have been installed. To update them
pip2 install --upgrade pip setuptools
You can install Python packages with
pip2 install <package>
They will install into the site-package directory
/usr/local/lib/python2.7/site-packages
See: https://docs.brew.sh/Homebrew-and-Python.html
now caveats for installed_on_request deps are also reported. What is not taken into account is whether these dependencies are beings upgraded in the current run (not sure if commented out line would do exactly that or not).
not ready yet, argh...
Closing for now... will reopen when I have more time to work on it.
|
2025-04-01T06:37:03.855616
| 2017-03-24T16:19:49
|
216835766
|
{
"authors": [
"raza15"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1257",
"repo": "Homebrew/homebrew-tex",
"url": "https://github.com/Homebrew/homebrew-tex/issues/42"
}
|
gharchive/issue
|
:tex is deprecated
Hi. I'm new to contributing to homebrew. As my first commit, I was looking for warnings in the homebrew formulas. One package showed the error * :tex is deprecated.. So to fix that error, can I just remove that :tex dependency?
Thanks.
@MikeMcQuaid can you please help me? Thanks,
|
2025-04-01T06:37:03.860871
| 2015-12-22T18:19:38
|
123531417
|
{
"authors": [
"Michael-Jing",
"MikeMcQuaid",
"afh",
"apjanke",
"bceverly",
"dunn"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1258",
"repo": "Homebrew/homebrew-x11",
"url": "https://github.com/Homebrew/homebrew-x11/issues/170"
}
|
gharchive/issue
|
i3 + XQuartz + (maybe) El Cap
Hi,
I'm trying to get i3 running under XQuartz on my 13" retina Macbook Pro running 10.11.2 and the only last remaining issue is that I can't for the life of me get my meta key working. I've tried every combination of the input preferences checkboxes on XQuartz, I've tried full screen and not full screen, I've tried using .Xmodmap to explicitly map Meta_R and Meta_L to mod1 and I'm stumped.
If anyone has this working and can share their .config/i3/config file, along with any other tweaks they had to do in .Xmodmap, .Xresources, or .xinitrc.d/* I would massively appreciate it.
Thanks.
If you could try installing i3 outside of Homebrew, that would help us confirm that this is a Homebrew-specific problem. Thanks!
Trying to build from source but bumping my head against the wall. I have created a CPATH that pulls in the include files I need; however, now I'm hitting an undefined type of CARDINAL. Any suggestions? I searched for documentation on how to build / install on the Mac and outside of building from source, I came up dry. Perhaps my Google-fu is weak...
What key are you using for Meta? Option (aka "alt")? One or both sides? And what isn't working about it? From what I understand about XQuartz and Mac keyboards, you don't really have a Meta key, and should maybe be mapping Alt instead.
@bceverly I have i3 (4.12 see #210) running on XQuartz (2.7.9) on Mac OS X El Capitan (10.11.4 (15E65)) and use both command keys (⌘) as a modifier.
Hope the following helps:
% grep 'set $mod' ~/.config/i3/config
set $mod Mod2
% xmodmap
xmodmap: up to 2 keys per modifier, (keycodes in parentheses):
shift Shift_L (0x40), Shift_R (0x44)
lock Caps_Lock (0x41)
control Control_L (0x43), Control_R (0x46)
mod1 Mode_switch (0x42), Mode_switch (0x45)
mod2 Meta_L (0x3f), Meta_R (0x47)
mod3
mod4
mod5
We'll accept PRs for this but we're not actively working on it at this time.
Could anyone share a step by step guide on how to set up i3 on mac osx? I like the i3 style very much, and I got a macbook pro recently, I have installed Ubuntu in order to use i3, but even though Ubuntu is good enough, the energy saving system is much better on osx.
|
2025-04-01T06:37:03.888734
| 2021-09-01T20:42:01
|
985654893
|
{
"authors": [
"ClashLuke",
"bionboy"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1259",
"repo": "HomebrewNLP/HomebrewNLP",
"url": "https://github.com/HomebrewNLP/HomebrewNLP/pull/4"
}
|
gharchive/pull-request
|
Basic CLI and Project Structuring
Changes
Basic CLI
Main functions: [preprocess, train, inference]
Created project structure
Reflecting CLI functions
Formatter Utility
Syntax Highlighted printing
Pretty printing/logging
Color Traceback
Demo
Usage/Help output
Syntax output
Training output
Traceback
I changed the style to keep it consistent with the rest of the repo, except in one newly-created file. Let's discuss code style and formatters in discord so that we won't accidentally botch each-others code.
I'd also ignore deepsource for now as it complains about all issues from all moved files. Let's fix the problems some other time.
|
2025-04-01T06:37:03.918622
| 2019-05-29T19:57:15
|
449993054
|
{
"authors": [
"Hoohm",
"hisplan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1260",
"repo": "Hoohm/CITE-seq-Count",
"url": "https://github.com/Hoohm/CITE-seq-Count/issues/57"
}
|
gharchive/issue
|
local variable 'bcs_corrected' referenced before assignment
Due to v1.4.2 being too slow, I tried the develop branch (542ec344fe26b1cbbd72194c6ed7b809a82de19b), as you suggested in the other thread at some point, but it failed with the following error:
Loading whitelist
Creating barcode tree for whitelist
Counting number of reads
Started mapping
Processing 72 reads
CITE-seq-Count is running with one core.
Mapping done for process 10349. Processed 8 reads
Mapping done
Counter({'TCCCAAGCATTAAGCT': 2, 'CGCGGAAGAGCACACG': 2, 'CCTTTCTGAAAGGAGA': 2, 'ACTTGCCACCAAGTCC': 2, 'ACAATTTAGACATGGC': 2, 'ACGGGGAAGGACGTCA': 2, 'CGAATATCCTTAAGAG': 1, 'GCATAAAGTGCACCGC': 1, 'AAGCACATCACCTTGA': 1, 'ATCGGAAGAGCACACG': 1, 'TTAGCATCAACAGGCC': 1, 'CAACCACTAATAGGTA': 1, 'CCTCTAATCGGTCGTC': 1, 'CCCGGCGTACGGGGAA': 1, 'CCATTAATAATGTTTT': 1, 'CGTGAATTCTGAGGCC': 1, 'CCTTTACCAGCTTTAG': 1, 'TTCCATTCTTTAGCTC': 1, 'ATAAATCACCTCACTT': 1, 'AGGCCGTCCGATCTAG': 1, 'AAGTTGCCATACAAAA': 1, 'ATTCAAACGGCCTGTC': 1, 'TAAACGCAAGCCTCAA': 1, 'GCAAAAAATTTAGGGT': 1, 'GGTGGTCTATAGTGTT': 1, 'GCGCGATTCGATCTGC': 1, 'TAGCAAGGCCACGACG': 1, 'TTCGGGAGGGTAGTCG': 1, 'CGTTTGGTCAGTTCCA': 1, 'TTTTCTTCTGCGTCAG': 1, 'CAGTAGACTCCTTCTG': 1, 'GATGTGGTAGAAGTCG': 1, 'GGCTGCGGACGACCAG': 1, 'ATAGCAAAGCCTCTAC': 1, 'GGCGCATAACGATACC': 1, 'GACCAATCTGACCAGC': 1, 'ACGTATTTAGCCACAT': 1, 'AAACGTCGGCTACAGT': 1, 'TGCCCTACTTGCCCTA': 1, 'CTTGCTGCTAAAGGTC': 1, 'CTCGGAGGAGCACACG': 1, 'GTGAGTTGTTCCATTC': 1, 'GAGTCTACACAGTGTT': 1, 'TACTGCTTGTTTACGA': 1, 'AATTCATCCATTAACT': 1, 'TAAGAGACCATCTTAA': 1, 'TCATAAGAGGTTTTAC': 1, 'GATCGAAGAGCACACG': 1, 'CGAGCAGTAGACTCCT': 1, 'CGCATTGCATTCATCA': 1, 'AGATTGAGGCTGGGAA': 1, 'AGAACGTGAAAAAGCG': 1, 'TCTGATTGTCCAGTTG': 1, 'AACGTACCTTCAAGAA': 1, 'AAGGTTCCCGATCTAA': 1, 'AATCCGACCAATCCCA': 1, 'GTACCTCGCAACGGCT': 1, 'GCCGATACTTGGAACA': 1})
Correcting umis
Traceback (most recent call last):
File "/home/ubuntu/miniconda/bin/CITE-seq-Count", line 11, in <module>
sys.exit(main())
File "/home/ubuntu/miniconda/lib/python3.6/site-packages/cite_seq_count/__main__.py", line 474, in main
bcs_corrected=bcs_corrected,
UnboundLocalError: local variable 'bcs_corrected' referenced before assignment
Any idea?
Hello @hisplan
I see that I made some changes that didn't pass running tests and still pushed them. I guess I needed to work on different systems and wanted the latest changes.
I'm gonna work on performance and fixes tomorrow. I'll push some changes by then. Let me know if you see improvements then.
You mentioned 1.4.2 being too slow, would you mind telling me which part was problematic for you?
I'm passing the 10x v2 whitelist and setting --expected_cells=0. I suspect that CB correction part?
Yes, that's the exact issue I want to fix. Trying to not go towards creating an index for it and this is not trivial.
Another quick fix for you would be to use the filtered list from 10x v2. It will go way faster.
Have you had a chance to work on this issue? I've tried 10x v3 whitelist (millions of barcodes;;) with --expected_cells=0, and it's been running for more than 3 days...
Hey @hisplan
One quick thing, if you're running a v3 chermistry, I would suggest a cell barcode correction of 0 --bc_collapsing_dist 0
Some barcodes have only on hamming distance between them, so collapsing would be a mistake.
I suspect they have some control code in there because they talk about 3M barcodes but the whitelist has 6M.
@hisplan On the performance part, I haven't been able to get a great increase in speed.
One more thing I have to test is to create an index and use this instead of a tree for barcode recovery.
This might take a lot of memory though. You can try this branch. I added some filtering of low content cells to speed up things.
Hi @Hoohm, thanks for your quick reply.
I also had a chance to talk to 10x a few months ago about the whitelist for 10x v3 chemistry. As far as I understand, the whitelist has 6.8M barcodes, but the half of it is for feature barcoding. And I was told that by design there is a mismatch of 2bp between GEX library and feature barcode. I asked if I could get separated lists for GEX and feature barcode libraries, but they said no;;;; I don't know why, but if you could get it, please do let me know :-)
Anyway, given this, would you still suggest --bc_collapsing_dist=0? I've been running with the value 1, but if I recall correctly, I did get a message something like this:
Testing cell barcode collapsing threshold of 1
Value is too high, reducing it by 1
Testing cell barcode collapsing threshold of 0
Using 0 for cell barcode collapsing threshold
Interesting.
I'm gonna try to get the info as well.
The response you got is strange because the mapping between cell barcodes for mRNA and ADTs is here and it has 6M lines... I'm very confused :D
Hmm, the branch I'm working on should skip the cell barcode correction completely if the number is at 0 whereas the older ones don't skip it properly.
Definitely try the brach with this modification. It also stops correcting unmapped UMIs and low UMI cells (1 or 0 UMI for a tag). Both of these filters reduce computation time. It's also reported in the report.
I didn't fully understand it either. But that was pretty much the response I got from 10x. And yeah, I'm using the exact same file. The filename suggests 3M, but it actually contains 6M (3M for GEX, 3M for feature barcoding).
Anyway, mainly I've been using the official 1.4.2 and the develop branch, but I will try feature/index_whitelist.
I'm having two problems with the feature/index_whitelist branch.
Problem 1
When I tried to load the matrix file using Seurat 3's Read10X function, it threw this exception:
Error in dimnamesGets(x, value) :
invalid dimnames given for “dgTMatrix” object
Problem 2
Using the --dense parameter threw this exception:
Writing dense format output
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 1653, in create_block_manager_from_blocks
mgr = BlockManager(blocks, axes)
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 114, in __init__
self._verify_integrity()
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 311, in _verify_integrity
construction_error(tot_items, block.shape[1:], self.axes)
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 1691, in construction_error
passed, implied))
ValueError: Shape of passed values is (5, 819821), indices imply (4, 819821)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/bin/CITE-seq-Count", line 10, in <module>
sys.exit(main())
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/cite_seq_count/__main__.py", line 483, in main
filename='dense_umis.tsv')
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/cite_seq_count/io.py", line 48, in write_dense
pandas_dense = pd.DataFrame(sparse_matrix.todense(), columns=columns, index=index)
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/frame.py", line 424, in __init__
copy=copy)
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/construction.py", line 167, in init_ndarray
return create_block_manager_from_blocks([values], [columns, index])
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 1660, in create_block_manager_from_blocks
construction_error(tot_items, blocks[0].shape[1:], axes, e)
File "/home/ubuntu/miniconda/envs/feature-index-whitelist/lib/python3.6/site-packages/pandas/core/internals/managers.py", line 1691, in construction_error
passed, implied))
ValueError: Shape of passed values is (5, 819821), indices imply (4, 819821)
Hello @hisplan
where are you on this? Have you been able to run it?
I'm wrapping things up for a new relase.
Great!
Don't use the V3 whitelist, it's more than double the number of barcodes. For V3 I would suggest not using the full one and just running without it.
Another important thing, the cells you will find in the Protein data with CSC is not going to match the RNA cells. You have to use the mapping they provide on their github page.
I feel bit uncomfortable running without the whitelist. The workaround we have now seems fine.
Anyway, which mapping file are you referring to?
This list
You have to map the whitelist you have from the RNA cell barcodes to the Protein cell barcodes. Yuo can use the file I linked here to do that.
If you use the whitelist from the RNA directly for the protein, you are actually getting RNA from a call and Protein data from another.
Oh, I’m doing this for hashtag. The scRNA-seq count matrix that I mentioned previously has the error-corrected barcodes which were generated using that list from 10x GitHub repo.
I'm not sure I understand. Which error-corrected barcodes are you talking about? The cell barcode of the TAGS barcode?
The cell barcodes in the scRNA-seq count matrix are already error-corrected. I'm feeding this to CITE-seq-Count via --whitelist (instead of using the whole 10x v3 whitelist). Since CITE-seq-count is working with a smaller set of whitelisted barcodes, the running time was reasonable for me.
By the way, you can close this ticket. I will try out the new release once it's out.
Yes, that is a sound strategy to use.
If you use only the RNA part of V3 and adding your own cell hashing protocol, all good.
If you use REAP seq, protein data from 10x V3, you need to use the mapping between RNA cell barcodes and the Protein cell barcodes.
I'll close the ticket then :)
|
2025-04-01T06:37:03.938306
| 2017-01-13T22:43:57
|
200750330
|
{
"authors": [
"Jaden-Giordano",
"jkleinsc",
"tleb"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1261",
"repo": "HospitalRun/hospitalrun-frontend",
"url": "https://github.com/HospitalRun/hospitalrun-frontend/issues/922"
}
|
gharchive/issue
|
Spaces in username, during logging
Expected behavior:
Space trimming
Actual behavior:
A Username or password is incorrect. error.
Steps to reproduce:
Go on demo, try an ID with a space in front or behind it ; it doesn't work.
Great projet BTW!
A simple fix would be to just trim the spaces, but couldn't it be a breach in security if you allowed the user access if they accidentally entered a space, or any invalid character, anywhere in the credentials?
I don't see why, you just delete the ones at the beginning and at the end and those in between are part of the username. But, should spaces be allowed at all?
Usernames shouldn't have spaces in them, and it makes sense to trim any spaces on login. It is a simple change that should produce a better user experience.
|
2025-04-01T06:37:03.943378
| 2022-09-01T02:36:46
|
1358163315
|
{
"authors": [
"CLAassistant",
"bennymelb"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1262",
"repo": "HospitalRun/hospitalrun-frontend",
"url": "https://github.com/HospitalRun/hospitalrun-frontend/pull/2981"
}
|
gharchive/pull-request
|
Bennylo/fix node
Fixes #[replace brackets with the issue number that your pull request addresses].
Changes proposed in this pull request:
[list out summary of changes here]
[list out summary of changes here]
[list out summary of changes here]
[etc]
Newly added dependencies with Bundlephobia links:
[Link of the new dependency]
[Link of the new dependency]
[etc]
Note: pull requests without proper descriptions may simply be closed without further discussion. We appreciate your contributions, but need to know what you are offering in clearly described format. Provide tests for all code that you add/modify. If you add/modify any components update the storybook. Thanks! (you can delete this text)
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Opened the wrong PR, closing this now
|
2025-04-01T06:37:03.987718
| 2018-12-12T01:09:37
|
390022198
|
{
"authors": [
"NoopDog",
"sampierson",
"theathorn"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1263",
"repo": "HumanCellAtlas/data-browser",
"url": "https://github.com/HumanCellAtlas/data-browser/issues/397"
}
|
gharchive/issue
|
Health check endpoints
Add route 53 health check as per https://github.com/HumanCellAtlas/dcp-monitoring/blob/master/README.md
┆Issue Number: HCAB-334
Does this issue even make sense? If the data browser is a JS app with no server side component, there is nowhere to put a health-check endpoint.
Exact thats why this is not implemented. It is not super clear what the goal is and how to meet the HCA alerting requirements with this. For the data portal we quickly did this:
https://dev.data.humancellatlas.org/health/ but the portal does have individual pages.
Not sure what we could test on the browser. That the index.html page loads? We could create a separate health.html page and load that just to check cloudfront and S3 but that wont tell if actual site is working.
We could have some testing endpoint run the JS and then look for something in the dom etc.
Not sure what the goal is here and what level of effort is justified.
Superseded by #398.
|
2025-04-01T06:37:04.000330
| 2023-09-20T11:44:21
|
1904810882
|
{
"authors": [
"Zahidpichen1",
"hogepodge"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1264",
"repo": "HumanSignal/label-studio",
"url": "https://github.com/HumanSignal/label-studio/issues/4804"
}
|
gharchive/issue
|
Enquiring something
Did you use image recognition tool in this program
Thanks for your question @Zahidpichen1. Please use the Label Studio community Slack forums for questions about label studio integrations.
|
2025-04-01T06:37:04.002863
| 2022-03-10T01:07:48
|
1164622648
|
{
"authors": [
"Humbedooh",
"sebbASF"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1265",
"repo": "Humbedooh/asfmm",
"url": "https://github.com/Humbedooh/asfmm/issues/11"
}
|
gharchive/issue
|
Meeting topic overly verbose
The default meeting topic does not need #asfmeeting prefix: indeed that is misleading as this is not IRC.
It could be something like:
Official ASF Members Annual Meeting.
Though it should probably start as:
Official ASF Members Annual Meeting. Starts at yyyy-mm-dd hh:mm UTC (in xx minutes)
Likewise #backchannel is not really appropriate for the other channel and could be dropped.
First bit addressed with b189657
|
2025-04-01T06:37:04.028620
| 2019-07-25T08:11:07
|
477161201
|
{
"authors": [
"Sbrenthughes",
"Thirumaran2011"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1266",
"repo": "Hygieia/hygieia-scm-bitbucket-collector",
"url": "https://github.com/Hygieia/hygieia-scm-bitbucket-collector/issues/5"
}
|
gharchive/issue
|
Unable to fetch data from bitbucket collector
Hi ,
I have tried to configure the bitbucket collector in hygieia .i am facing the following issues.
application.properties:
dbname=dashboarddb
# Database HostName - default is localhost
dbhost=localhost
# Database Port - default is 27017
dbport=27017
# MongoDB replicaset
dbreplicaset=false
dbhostport=localhost:27017
# Database Username - default is blank
dbusername=dashboarduser
# Database Password - default is blank
dbpassword=dbpassword
# Logging File location
logging.file=./logs/bitbucket.log
# Collector schedule (required)
git.cron=0 0/1 * * * *
# Mandatory parameters
git.host=mybitbucketrepo.com/
git.api=/rest/api/1.0/
# Maximum number of days to go back in time when fetching commits
git.commitThresholdDays=15
logging.level.com.capitalone.dashboard=DEBUG
logging.level.com.capitalone.dashboard.collector=DEBUG
# Page size for rest calls
# Only applicable to Bitbucket Server.
# Only applicable to Bitbucket Cloud.
git.pageSize=25
# Bitbucket product
# Set to 'cloud' to use Bitbucket Cloud (formerly known as Bitbucket)
# Set to 'server' to use Bitbucket Server (formerly known as Stash)
# More information can be found here:
https://github.com/capitalone/Hygieia/issues/609
git.product=server
# Bitbucket key for private repos
git.key=fM0sYnrWPYxozSNuMX9dcwhhJtkOKkNz
bitbucket.key=fM0sYnrWPYxozSNuMX9dcwhhJtkOKkNz
Issue resolved. Closing
|
2025-04-01T06:37:04.039947
| 2020-10-26T21:26:19
|
729919944
|
{
"authors": [
"evolverine",
"gcassel",
"thomasgwatson",
"tibetsprague"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1267",
"repo": "Hylozoic/hylo-evo",
"url": "https://github.com/Hylozoic/hylo-evo/issues/676"
}
|
gharchive/issue
|
Add emoji reactions to posts and comments
[x] For posts change upvoting to emoji reactions.
[x] For comments and posts.
[x] Mouse over the reactions to see how many of each one there are
[x] Show a list of who reacted (and in which way?)
[x] Do we want a limited set like FB or all of them like Slack? Let's go with all emojis, but have a default set like Slack
[x] What set of default emojis to use?- [ ] Should each post type have a different default set or the same one?
[x] How to deal with current "upvotes"? Make them all a certain emoji, make thumbs up?
[x] Emoji picker: Maybe use: https://github.com/missive/emoji-mart Or https://github.com/ealush/emoji-picker-react
[x] React button will use smiley face icon
[ ] How will list view show reactions?
[ ] Do we want to add to grid views?
evo-node integration checks
[ ] old vote graphQL allows for successful 'vote'
[ ] old vote graphQL allows for successful removal of 'vote'
[ ] post reaction creates reaction
[ ] post reaction updates post.reactions and post.num_people_reacts
[ ] comment reaction creates reaction
[ ] comment reaction updates post.reactions and post.num_people_reacts
[ ] post delete reaction creates reaction
[ ] post delete reaction updates post.reactions and post.num_people_reacts
[ ] comment delete reaction creates reaction
[ ] comment delete reaction updates post.reactions and post.num_people_reacts
@brodeur
Small grid view is the most obviously bad and there are several approaches we could take to shift that. Some of those approaches might be done by Aaron and others might need Tom (if we want to build another component for this use-case)
big grid view: At a minimum, scaling down the size of all the components in EmojiRow would go a long way here. We might also want to hide emojis when there are too many
List row view: Again scaling down EmojiRow stuff here would be nice.
Heya. How's this coming along? 🤞🏼 It's a much anticipated feature for us 😃
It’s so closseee, coming in the next couple weeks.
On Nov 14, 2022, at 1:09 PM, M @.***> wrote:
Heya. How's this coming along? 🤞🏼 It's a much anticipated feature for us 😃
—
Reply to this email directly, view it on GitHub https://github.com/Hylozoic/hylo-evo/issues/676#issuecomment-1314383271, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAA5HKC5P4MN362ADORUTDWIKTCPANCNFSM4S76BS4Q.
You are receiving this because you authored the thread.
Is this a good place to request a skin tone selector or slider for emoji? (Perhaps it should be added to the nearly-complete list, or else set aside for future development.) I suppose it's a good chunk of work, but it may be increasingly expected for matching the feature set of platforms such as Discord and Slack.
|
2025-04-01T06:37:04.052278
| 2022-12-15T08:55:15
|
1498039791
|
{
"authors": [
"Hzfinfdu",
"jzhang38",
"yujianll"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1268",
"repo": "Hzfinfdu/Diffusion-BERT",
"url": "https://github.com/Hzfinfdu/Diffusion-BERT/issues/6"
}
|
gharchive/issue
|
How to evaluate BLEU score on LM1B?
Dear authors,
I understand that you plan to release your code on January. But could you share more details regarding how you evaluate the BLEU score and PPL on the LM1B dataset? I am also working on Diffusion Model for text and may potentially cite your paper. Thanks!
Hi,
We computed the BLEU score with all test data as references and reported the average BLEU score of each generated sentence. We sampled 1K sentences respectively for evaluating BLEU and S-BLEU.
For PPL, the ELBO on the test set is an upper bound of token-wise NLL. And we first convert such bound to per-word NLL and use this to get the per-word PPL.
Hope this helps!
@yujianll Hi,
Yes, we sum up NLL for all tokens in the sequence as NLL for the sequence.
The validation ELBO is around 110. And the average number of words in each sequence in the test set is around 26. Thus per-word NLL is around 4.23. The test PPL is obtained by exp(4.23).
@Hzfinfdu Thanks for the reply!
I have another low-level question. When you calculate NLL on test set, do you sum for all T diffusion steps, or do you sample a few time steps for calculation? If you do sample, how many time steps do you use?
@yujianll Hi,
We trained DiffusionBERT with 512 steps and used DDIM sampling to uniformly sample 128 steps on test set, both for NLL calculation and generation.
Hope this helps!
Thanks, this helps a lot!
|
2025-04-01T06:37:04.067683
| 2024-02-13T17:41:53
|
2132845630
|
{
"authors": [
"aitronssesin",
"blaise-tk",
"cheesypy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1269",
"repo": "IAHispano/Applio",
"url": "https://github.com/IAHispano/Applio/issues/289"
}
|
gharchive/issue
|
[BUG] Wrong precompiled link on Release Page Applio 3.0.6
Describe the bug
A clear and concise description of what the bug is.
On release page for Applio 3.0.6, precompiled link points to:
https://huggingface.co/IAHispano/applio/resolve/main/Applio V3 Precompiled/ApplioV3.0.6.zip?download=true
Which gives me an error.
Maybe you meant this one ?
https://huggingface.co/IAHispano/Applio/resolve/main/Compiled/ApplioV3.0.6.zip?download=true
To Reproduce
Go to 'Applio 3.0.6 Release page'
Click on precompiled
See error
Expected behavior
I wanted to download the precompiled version of the project ApplioV3.0.6.zip from huggingface by clicking the link
yeah we are doing some changes, check the readme for the links
K, was just relating in case you weren't aware of it. The link for the precompiled directory in the readme page is also a 404.
Fixed
|
2025-04-01T06:37:04.106071
| 2023-04-27T11:41:02
|
1686673335
|
{
"authors": [
"irmsan",
"jacobemery"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1270",
"repo": "IBM/Ansible-OpenShift-Provisioning",
"url": "https://github.com/IBM/Ansible-OpenShift-Provisioning/issues/118"
}
|
gharchive/issue
|
OPEN VPN
Please implement a variable to decide whether OpenVPN setup is needed or not.
And use the OpenVPN deployment dependent of that variable (default should be FALSE i think)
Hi @irmsan, agreed! Could you please take a look at PR?
It adds a new setup_openvpn to the bastion's variables and is used only for openVPN tasks.
Will that resolve this issue? Let me know.
|
2025-04-01T06:37:04.109314
| 2021-02-11T20:59:40
|
806735908
|
{
"authors": [
"lmsurpre",
"prb112"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1271",
"repo": "IBM/FHIR",
"url": "https://github.com/IBM/FHIR/issues/1936"
}
|
gharchive/issue
|
Automated e2e integration tests for the fhir-notification module
While reviewing https://github.com/IBM/FHIR/pull/1932 I noted that we really don't have any coverage of our fhir-notification module in our CI pipeline.
Way back when we added NATS.io support for notifications, we did the work to spin up the NATS cluster as part of our docker-compose env (with db2), but we never did circle back around to adding automated tests for it:
From build/README-DB2.md:
Note: If you are testing NATS notifications, invoke the NATS subscriber via node fhir-server-test/src/test/nodejs/nats-subscriber. If this is your first time, install the dependencies first by installing Node.js (if not already installed) and running (cd fhir-server-test/src/test/nodejs && npm install).
I'm hoping we can expand on that and add automated tests that cover the create, update (and patch), and delete notification events within our existing CI.
We should probably upgrade the NATS dependency to 2.0 while we're at it...
I added Kafka (as it's the easiest to add, if we want to add nats, I can add that too)
|
2025-04-01T06:37:04.111463
| 2018-08-07T17:53:52
|
348424866
|
{
"authors": [
"cmgrote"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1272",
"repo": "IBM/ansible-role-infosvr-import-export",
"url": "https://github.com/IBM/ansible-role-infosvr-import-export/issues/12"
}
|
gharchive/issue
|
Simplify mappings
At the moment there are different mappings for different loading mechanisms:
IGC REST API mappings
istool mappings
IA REST API mappings
Would be better if there were just one consistent way of doing mappings, and we transparently translate that in the background to the form that we might require for a particular scenario (otherwise user needs to configure the same mapping multiple times across each area).
... or simply recommend an approach which means defining key attributes to be mapped in a separate vars file, and leave it to the playbook itself to determine how to apply those mappings? (That way users that define the target values for mapping don't need to be aware of / worry about the source values?)
|
2025-04-01T06:37:04.120398
| 2020-02-03T14:40:17
|
559126674
|
{
"authors": [
"enricoberti",
"stuckless",
"tay1orjones"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1273",
"repo": "IBM/carbon-addons-iot-react",
"url": "https://github.com/IBM/carbon-addons-iot-react/pull/870"
}
|
gharchive/pull-request
|
fix(floating-menu): add floating menu container to Storybook
Closes #576
Summary
To get the floating menu to scroll together with the page, it should be attached to its parent element instead of <body> itself. To do that, you just need to add data-floating-menu-container to the element you want the menu to attach to (in this case the storybook container)
Acceptance Test (how to verify the PR)
Verify it on ?path=/story/watson-iot-table--stateful-example-with-secondary-title
Please not that our storybook has a wrapping element with the styling position: fixed wrapping the stories, so the menu offset calculation fails sometimes. It's not a problem in normal apps.
If normal apps don't use position: fixed then maybe we shouldn't use it by default in storybook?
I'm seeing an issue when I scoll the page and then click on overflow menu that the overflow menu is not positioned correctly. Now this could be caused by what you described here... but, how can we be sure, and can we validate this? Should we try to turn off the position fixed in the story so that we can validate the correct behaviour will happen?
@stuckless thanks for going through this.
In our case the 'position:fixed' on storybook is from the 'centered' addon. My recommendation is to remove that, and with that at the same time we get the same experience of Carbon (they don't center the components in the stories) and also avoid positioning problems like this. (@davidicus ).
I updated this PR removing the centered addon for the specific story and it fixes the offset problems.
You can test it here: https://deploy-preview-870--carbon-addons-iot-react.netlify.com/?path=/story/watson-iot-table--stateful-example-with-secondary-title
If we implement it storybook-wise, this also eliminates the need of adding data-floating-menu-container to our storybook container.
Also, as you said, this doesn't fix the problem downstream but it helps with fixing it. I don't have access to the downstream implementation to check for it.
It is not really well documented on Carbon's component but it should be the choice of who's developing an app wether to attach an overflow menu to the body (should be 99.9% of the cases) or to another element so the positioning works correctly.
Also, given how the floating menu positioning is implemented, it cannot be arbitrarily added to the best available parent, like the table itself in our case.
@enricoberti Thanks for looking into this. So, I can confirm it appears to work ok in my browser, I guess the original bug was opened up against this only happening on on an ipad. I don't have an ipad (here) for testing... have you verified that even with your most recent change that it behaves the same on ipad vs desktop.
Nothing to do with your change :) but these kinds of behaviours scare me, a little, since it's like a mine field for the downstream consumer. ie, when do I set this, where, how, etc :(
@stuckless yes, verified on the ipad too :) i know how you feel with that - it should actually be documented better also on the Carbon side of things
:tada: This PR is included in version 2.39.3 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:37:04.123623
| 2018-11-13T19:44:46
|
380392525
|
{
"authors": [
"joshblack",
"shixiedesign"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1274",
"repo": "IBM/carbon-elements",
"url": "https://github.com/IBM/carbon-elements/issues/65"
}
|
gharchive/issue
|
Add @carbon/motion package with timing information
This issue is for creating the @carbon/motion package with SCSS and JS support for easing curves for components. Will need the values from @shixiedesign when she has the chance!
Easing
Productive
Expressive
Standard easing
cubic-bezier(0.2, 0, 0.38, 0.9)
cubic-bezier(0.4, 0.14, 0.3, 1)
Entrance easing
cubic-bezier(0, 0, 0.38, 0.9)
cubic-bezier(0, 0, 0.3, 1)
Exit easing
cubic-bezier(0.2, 0, 1, 0.9)
cubic-bezier(0.4, 0.14, 1, 1)
|
2025-04-01T06:37:04.146339
| 2023-02-01T13:17:05
|
1566100538
|
{
"authors": [
"LasseBlaauwbroek",
"fidel-schaposnik",
"jasonrute",
"mirefek"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1275",
"repo": "IBM/graph2tac",
"url": "https://github.com/IBM/graph2tac/issues/96"
}
|
gharchive/issue
|
Improve our use of tf.Dataset
In the TFGNN model we are using tf.Dataset to make a data processing pipeline. It has certain advantages and disadvantages that we are bumping up into. In this issue, I'm gathering details so that we can have a complete picture of the situation.
The good
Datasets are basically required (as I understand it) to split your data on multiple GPUs. So we would at least need a Dataset at some stage of the pipeline.
Datasets make it very easy to work with individual graphs and then batch them into batched graphs when the time comes for batched data. This greatly improves the maintainability of some parts of the code.
It is a standard tensorflow API for data so it is fully featured.
It makes it easy for us to mix the definitions dataset with the proof states dataset.
It gives us a convenient way to cache the data in memory (or in files).
To the extent that we can cache our processed dataset in memory, it makes it really fast to work with the dataset over hundreds of epochs.
The bad
Our current use of tf.Dataset doesn't scale well and is slow. Ideally we wouldn't need to worry about shuffling or caching the data in the dataset and instead let our loader handle this, but the problem is that processing the data is too slow. (See the observations section below for some thoughts on what is slow.)
The dataset needs to be compiled. This is fine if you only have one super dataset pipeline, but was bad when we tried in the predict server to make a new dataset for every new definition.
Observations
Ideally we would like this dataset processing to be fast enough that we don't need to cache or shuffle the data outside of the loader. But this may also not be practical and we need to look into better ways to cache and shuffle the preprocessed data.
I haven't done exact timings, but processing the 70,000 definitions seems to take a lot longer than the 500,000 proof states. This could be either because the definitions are much larger graphs or because there is some additional step to the definition processing which is making it slow. Figuring out what steps are slow could make a big difference in speeding up the code without "throwing the baby out with the bathwater".
We also have some abilities to process the graphs outside the dataset. This was required for the predict server. We should compare timing of the two approaches, and maybe considering doing some of the processing outside the dataset.
I'm experimenting right now with recomputing definitions between embeddings. I'm not using the dataset pipeline, but I wonder if I should. I guess it is a good way to compare the speeds of both approaches.
There are whole articles on optimizing the dataset pipeline that we could look into.
How would we like to in general want the caching to behave? Either, we can load all the data into the memory in the format ready for the network (but then, there can be the scaling issue if the data don't fit into memory), or we want to reload all the data each epoch (which can be slow). Or are there any other options? (like saving a cache to hard drive in TF format, I am not sure how feasible it would be).
I would like to first properly understand our aim, and then we can try to figure out the technical details of whether we want to use tf.Dataset, or not. Note that the loader basically offers random access to the datapoints (but it is computing them every time they are accessed).
To me, the basic requirements of training are like this:
During one training epoch, we train on randomly generated batches (subsets) of the dataset. A batch is basically the size of GPU memory.
At some point, we will no longer be able to hold the entire training data into RAM. Hence, we will have to load it in batches. These batches can either be the same size of the GPU-batch, or a more granular 'RAM-batch'.
The loader will have to be fast enough such that one batch can be fetched while the previous batch is being processed by the GPU. (It's hard to imagine this being a problem, the loader seems relatively fast now.)
To me, it seems that the proper way of dealing with this is as follows:
. When training starts, we load the dataset into mmap memory (note that this is an operation that takes minimal memory regardless of the size of the dataset, this is the beauty of Captn Proto). Then, we calculate an index that contains the root nodes of all proof states and definitions we want to train on. This index is kept in memory permanently, but should be fairly small. 2. While epoch nis being trained, the next epochn+1is being prepared: The index from (1) is randomly permuted and split into batches. This should be pretty quick, but if it is not, we have the entire epoch to calculate it. 3. While batchiof epochnis being trained, batchi+1` is fetched from the Captn Proto dataset. That is, we calculate the forward closures of all the root nodes in the batch and load them into a Numpy array or whatever other datastructure.
4. We pray that fetching a batch is fast enough to keep the GPU busy. But if it is not, I guess we can parallelize this.
I see, so we would like the Dataset to look ahead, and prepare a batch it was not asked about yet... I would have to look more into Dataset to see if it is happening by default, or how to do it.
Also, at some point, we were considering moving the graph loader into Cython in case that would be a bottleneck (but I think we concluded there were more serious speed issues).
Note that if it is indeed the case that we need tf.Dataset in order to parallelize over multiple GPU's, then I propose this scheme: A tf.Dataset basically corresponds to a RAM-batch. This is the largest size that we are willing to load into RAM. The tf.Dataset can the split this batch into smaller GPU-batches.
The only other alternative I see is to load the entire Capt'n Proto dataset into a tf.Dataset. But if we run out of memory, then the tensorflow code will have to be responsible for swapping part of the tf.Dataset to disk. Does this functionality exist? (I would think so, because surely we are not the only ones with datasets that exceed RAM?)
Looking though the tensorflow API, it looks to me like a lot of what I describe can be easily done using a combination of the prefetch functionality and the from_generator functionality.
I spend some time digging through the codebase and through the tensorflow documentation. My impression is that using tf.data.Dataset is a good idea in general, and there is no reason why we can't have our cake and eat it too. Here is what I would suggest as a 'plan of attack':
Get rid of the old C++ loader code (it clutters up the repo, and the next steps will break it)
Is the tfgnn model now superior in every way to the tf2 model? If so, let's get rid of the tf2 model (if we ever need it again, it is in git's history).
Let's merge tfgnn.dataset.Dataset and loader.py_data_server.DataServer into one class. This will save a lot of transformations in the pipeline and simplify the code. I don't really see any reason why we need two classes here.
Move the shuffling of the data as early into the pipeline as possible. Ideally at the first step, where a proof state and definition is still a single root node. The shuffle function requires a buffer, which scales linearly with the size of the data in the buffer. Hence, it is much cheaper to do this early in the pipeline. When the shuffling it the first step in the pipeline, it does not have to take any memory at all. Currently, the shuffling is happening way to late.
Remove any calls to cache. There are currently multiple in the pipeline. At most, there should be one call, but ideally none at all.
Experiment with different prefetch buffer sizes (instead of AUTOTUNE). Use the TF Profiler to see if we have any bottlenecks.
If there are still bottlenecks, experiment with adding num_parallel_calls to batch and any remaining map calls in the pipeline.
If there are still bottlenecks, implement the forward closure computation in Cython.
If there are still bottlenecks, use the TF interleave function to feed data into the pipeline at parallel.
If there are still bottlenecks, go sit in the corner and cry.
While I'm not very up-to-date on the code-level details of how the tf.data.Dataset is being used right now, let me just point out that tf.data.Dataset.cache supports on-disk caching too, and if I recall correctly this included "fancy" features such as sharding (for when e.g. you need to split the data into multiple files because otherwise it's too big or too slow to read a batch from a single disk). So in principle one should be able to mostly keep the same pipeline once the data becomes larger than the available RAM memory, except for some finetuning of the on-disk-caching. My impression is that trying to achieve this from scratch (handling serialization and deserialization, prefetching, sharding, etc) would be a lot of work and also mostly reinventing the wheel...
Attached is a diagram of my proposal. I suggest to entirely ditch the idea of a dataserver. This brings way to much overhead and complexity with it. Instead, I'd go for an entirely functional approach.
stream-proposal.pdf
Thanks for looking into this everyone, especially @LasseBlaauwbroek!
Is the tfgnn model now superior in every way to the tf2 model? If so, let's get rid of the tf2 model (if we ever need it again, it is in git's history).
Mostly. There were still a few more comparison experiments I wanted to run, but I could do those on v13 or v14 in old branches. I personally have no intention of going back to the TF2 model. All are new features are in in the TFGNN model.
Let's merge tfgnn.dataset.Dataset and loader.py_data_server.DataServer into one class. This will save a lot of transformations in the pipeline and simplify the code. I don't really see any reason why we need two classes here.
I'm not sure I understand the proposal here, but I'm open to it.
Remove any calls to cache. There are currently multiple in the pipeline. At most, there should be one call, but ideally none at all.
Yes we probably don't need all the calls to cache we have. One is because we split the data in the dataset and we need a cache before the split to prevent recomputing all the test data for the training data pipeline. If we handle training and validation data better, then this will go away.
I'm not sure I understand all your proposals in detail yet, Lasse, but here are some high level things I'd like from a pipeline in order of preference:
Robust to tricky-to-debug errors (for example indexing errors in graphs).
Easy to experiment with ideas. Both you and I have ideas for other things we would like to add to training. If it is easy to add different data (and avoid bugs when doing so) that would be great and improve our scientific productivity.
Scalable to larger data (and remain runnable on at least 2 GPUs).
Fast. It doesn't have to be lightning fast, but it would be nice for it to be fast enough to get results in a few days of training.
I think we can easily do all of those things.
|
2025-04-01T06:37:04.149080
| 2024-04-04T08:39:15
|
2224865294
|
{
"authors": [
"DaniyalMustafa"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1276",
"repo": "IBM/ibm-common-service-operator",
"url": "https://github.com/IBM/ibm-common-service-operator/pull/1908"
}
|
gharchive/pull-request
|
Removing ping
Issue: https://github.ibm.com/IBMPrivateCloud/roadmap/issues/62325.
We need not to check response of host to the ping
logs from running the job:
Changing the hostname from cp-test-ibm-common-services.apps.cat.cp.fyre.ibm.com to cp-test-ibm-common-services.apps.cat.cp.fyre.ibm.com
configmap/cs-onprem-tenant-config unchanged
Given Custom Hostname: cp-test-ibm-common-services.apps.cat.cp.fyre.ibm.com
Host is reachable. Proceeding further...
Custom secret not configured
Deleting old job of iam-custom-hostname if exists
job.batch "iam-custom-hostname" deleteda
Running custom hostname job
job.batch/iam-custom-hostname created
platform-auth-service is available.
job.batch/iam-custom-hostname condition met
successfully updated the custom hostname
token_output
Access token is present.
|
2025-04-01T06:37:04.157937
| 2024-11-04T13:22:14
|
2632807521
|
{
"authors": [
"baal-lgln"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1277",
"repo": "IBM/ibmcloud-object-storage-plugin",
"url": "https://github.com/IBM/ibmcloud-object-storage-plugin/issues/152"
}
|
gharchive/issue
|
Error while loading shared libraries: libfuse.so.2
IBM Kubernetes Cluster
Kubernetes Version: 1.29.9 (Node with Ubuntu 24)
ibm-object-storage-plugin: 2.2.32
Since updating our Kubernetes Nodes to Ubuntu 24 we get the following error in our container:
MountVolume.SetUp failed for volume "pvc-..." :
mount command failed, status: Failure, reason:
Error mounting volume: s3fs mount failed: s3fs: error while loading shared libraries: libfuse.so.2:
cannot open shared object file: No such file or directory
The container stays in pending mode. The PVC however is bound.
Example PVC:
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: test-pvc
labels:
billingType: hourly
region: eu-de
annotations:
ibm.io/add-mount-param: complement_stat,compat_dir
ibm.io/auto-create-bucket: 'false'
ibm.io/auto-delete-bucket: 'false'
ibm.io/bucket: BUCKET_NAME
ibm.io/endpoint: https://s3.direct.eu-de.cloud-object-storage.appdomain.cloud
ibm.io/secret-name: s3-access-secret
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 100Gi
storageClassName: ibmc-s3fs-smart-perf-regional
Problem still persists.
Also tested on a Cluster Node with Kubernetes version 1.31.1_1527 with Ubuntu 24.
PVC YAML:
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: test-pvc
annotations:
ibm.io/add-mount-param: complement_stat,compat_dir
ibm.io/auto-create-bucket: 'false'
ibm.io/auto-delete-bucket: 'false'
ibm.io/bucket: BUCKET_NAME
ibm.io/endpoint: https://s3.direct.eu-de.cloud-object-storage.appdomain.cloud
ibm.io/secret-name: SECRET_NAME
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 10Gi
storageClassName: ibmc-s3fs-smart-perf-regional
Status is bound and PersistentVolume is created, but the Pod will not start:
apiVersion: v1
kind: Pod
metadata:
name: pvc-test
spec:
volumes:
- name: data
persistentVolumeClaim:
claimName: test-pvc
containers:
- name: pvc-size-test
image: public.ecr.aws/docker/library/busybox:stable
command: ['/bin/sh', '-c']
args: ['while true; do du -sh /data; sleep 10;done']
volumeMounts:
- mountPath: '/data'
name: data
resources: {}
This still leads to the error FailedMount and a pod stuck in ContainerCreating:
MountVolume.SetUp failed for volume "pvc-3a8a9b35-75a5-4289-8aaa-fcd38143bcb3" : mount command failed, status:
Failure, reason: Error mounting volume: s3fs mount failed: s3fs: error while loading shared libraries:
libfuse.so.2: cannot open shared object file: No such file or directory
Just switching to Ubuntu 20 as the base OS for the Nodes resolves this error.
Ubuntu 20 is deprecated and will soon stop to be be supported by IBM Kubernetes Clusters, we need a solution for this!
|
2025-04-01T06:37:04.159684
| 2019-07-12T13:47:07
|
467423293
|
{
"authors": [
"germanattanasio",
"marloncarvalho"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1278",
"repo": "IBM/java-sdk-core",
"url": "https://github.com/IBM/java-sdk-core/pull/45"
}
|
gharchive/pull-request
|
Allow for authentication for configured proxy
This PR fixes issue #44. @lpatino10
@marloncarvalho you also need to configure git because your user doesn't match your commit
I see! It's using my enterprise username and e-mail. Sorry. I've changed it. What should I do now? Just commit something and push it? Fork again?
Nevermind! I didn't know I can change the Author with rebase and amend. Fixed. @lpatino10 @germanattanasio
|
2025-04-01T06:37:04.163027
| 2021-03-12T02:43:49
|
829703321
|
{
"authors": [
"ibm-devx-automation",
"mkistler"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1279",
"repo": "IBM/node-sdk-core",
"url": "https://github.com/IBM/node-sdk-core/pull/128"
}
|
gharchive/pull-request
|
Add getQueryParam utility method to support pagination
This PR adds a small utility method to extract the value of a query parameter from a URL string. This will be used in the pagination support for Node.
Checklist
[x] npm test passes (tip: npm run lint-fix can correct most style issues)
[x] tests are included
[ ] documentation is changed or added
:tada: This PR is included in version 2.10.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:37:04.163847
| 2020-05-06T16:23:49
|
613448346
|
{
"authors": [
"xjtustt"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1280",
"repo": "IBM/operand-deployment-lifecycle-manager",
"url": "https://github.com/IBM/operand-deployment-lifecycle-manager/pull/375"
}
|
gharchive/pull-request
|
fix the wrong path issue
Upgrade operator docs moved from /design to /user but docs have not been changed.
/assign @danielxlee
|
2025-04-01T06:37:04.168752
| 2018-05-21T19:22:00
|
325029802
|
{
"authors": [
"MHBauer",
"eggshell",
"mpetason"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1281",
"repo": "IBM/rotisserie",
"url": "https://github.com/IBM/rotisserie/issues/123"
}
|
gharchive/issue
|
http should redirect to https for dev.rotisserie.tv
http://dev.rotisserie.tv is the default nginx landing page.
the Thanks slide at https://blog.eggshell.me/rotisserie-talk/#39
links to dev.rotisserie.tv which by default is interpreted as http when you copy and paste it in.
http://dev.rotisserie.tv
It should be a direct clicky link to https://dev.rotisserie.tv
Additionally http://dev.rotisserie.tv should redirect to https://dev.rotisserie.tv.
[x] fix slides with direct link to https://dev.rotisserie.tv
[ ] make http://dev.rotisserie.tv redirect to https://dev.rotisserie.tv
Dev is for staging. Right now we are using Cert-Manager + Kube-Lego in the prod cluster, where dev also lives, which causes the issues with redirects. Cert-Manager + the IBM ingress controller should be controlling all traffic for the domain, however Kube-Lego also sees the annotation and is setting up ingress resources for the domain. There is currently a bug with the Ingress Controller which keeps it from pulling records for the same host from two different ingress resources.
Basically - As soon as PR #122 is merged and we re-deploy prod + dev + fortnite, we shouldn't run into this issue. The setup already assumes HTTP to HTTPS redirects.
Went ahead and updated the last slide. Thanks @MHBauer
confirm slide update.
http redirects to https after removing kube-lego and switching everything over to cert-manager.
confirm redirect.
thanks guys. good job!
|
2025-04-01T06:37:04.171929
| 2022-11-24T14:28:24
|
1463431362
|
{
"authors": [
"pgodowski",
"xcliu-ca"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1282",
"repo": "IBM/test-infra",
"url": "https://github.com/IBM/test-infra/pull/496"
}
|
gharchive/pull-request
|
do not error out docker login although multicloudlab credential unknown
Signed-off-by: Xinchun Liu<EMAIL_ADDRESS>What this PR does / why we need it:
Which issue(s) this PR fixes (optional, in fixes #<issue number>(, fixes #<issue_number>, ...) format, will close the issue(s) when PR gets merged):
Fixes #
Specify your PR reviewers:
/cc @ashank07 @pgodowski
with this, we can test build-tool / check-tool builds, will not be able to push to quay.io/multicloudlab though
/lgtm
|
2025-04-01T06:37:04.174189
| 2019-03-15T18:14:44
|
421646003
|
{
"authors": [
"scottdangelo",
"sidd-gupta"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1283",
"repo": "IBM/vr-speech-sandbox-cardboard",
"url": "https://github.com/IBM/vr-speech-sandbox-cardboard/issues/58"
}
|
gharchive/issue
|
new prefabs addition error
I am trying to add new prefabs (or objects). I added intent and entities in the IBM cloud and also in
the assets folder. But I am getting the error KeyNotFoundException.
Can you please tell me the changes in the code to add new prefabs.
@scottdangelo Can you please help me out with this?
Sorry, but that is beyond the scope of this code pattern. You will need to do some research on Unity Prefabs. You can look in the Unity editor at the keys/values in the scene where the existing prefabs are, and get some clues as to how to modify for your purposes.
|
2025-04-01T06:37:04.176469
| 2022-08-04T14:00:50
|
1328686163
|
{
"authors": [
"akirafujiu",
"arsuryan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1284",
"repo": "IBM/watson-openscale-samples",
"url": "https://github.com/IBM/watson-openscale-samples/pull/45"
}
|
gharchive/pull-request
|
add notebook for troubleshooting in 4.5.1 to migrate custom monitor i…
…nstances.
Target issue:
https://github.ibm.com/aiopenscale/tracker/issues/27047
Change sets:
A notebook to migrate monitor instances for custom monitor definition created in 4.0.X
@kishore-patel - Please review this notebook. Thanks.
@arsuryan @kishore-patel Could either of you please merge/rebase this PR?
|
2025-04-01T06:37:04.178057
| 2022-11-28T13:54:04
|
1466469102
|
{
"authors": [
"ogauneau",
"phaumer"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1285",
"repo": "IBM/zopeneditor-about",
"url": "https://github.com/IBM/zopeneditor-about/issues/292"
}
|
gharchive/issue
|
Need timestamps on user build output/log
To troubleshoot performance problems, it will be helpfull to have timestamps on the log output of the user build:
To better see if the time is spent in Z Open Editor or Zowe Explorer or RSE API calls,...
@ogauneau we now have timestamps in the user build log when the Z Open Editor v3.1.0 log level is set to DEBUG. Same setting for both logs.
|
2025-04-01T06:37:04.191123
| 2023-06-04T14:19:22
|
1740400524
|
{
"authors": [
"XinyueZ",
"rentainhe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1286",
"repo": "IDEA-Research/GroundingDINO",
"url": "https://github.com/IDEA-Research/GroundingDINO/issues/131"
}
|
gharchive/issue
|
How to optimise via IoU if too many similar objects are detected?
Hey, I always have a very large bbox when I query text-prompt "a house" , meanwhile single ones are detected correctly.
How to avoid the large bbox, at least I want to get an approach generally to filter the large ones.
thanks
You can add NMS after the output box
@rentainhe Does this repo provide an implementation of it?
@rentainhe Does this repo provide an implementation of it?
Please refer to here: https://github.com/IDEA-Research/Grounded-Segment-Anything/blob/8124fe737dc877ec49a0881785119fe222a4c868/automatic_label_simple_demo.py#L104
yeah, thank you I have seen it
|
2025-04-01T06:37:04.194411
| 2023-10-26T00:35:43
|
1962507719
|
{
"authors": [
"DoubleDeez"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1287",
"repo": "IDI-Systems/UnrealImGui",
"url": "https://github.com/IDI-Systems/UnrealImGui/pull/4"
}
|
gharchive/pull-request
|
Fix MetaClass warning for UImGuiSettings::ImGuiInputHandlerClass
Wed Oct 25 17:24:39 PDT 2023 Warning LogClass Property StructProperty UImGuiSettings::ImGuiInputHandlerClass defines MetaData key "MetaClass" which contains short type name "ImGuiInputHandler". Suggested pathname: "/Script/ImGui.ImGuiInputHandler". Module:ImGui File:Private/ImGuiModuleSettings.h
Looks like the relevant code was adding in 5.1
https://github.com/EpicGames/UnrealEngine/blob/5de4acb1f05e289620e0a66308ebe959a4d63468/Engine/Source/Runtime/CoreUObject/Private/UObject/Class.cpp#L3733C8-L3733C8
https://github.com/EpicGames/UnrealEngine/commit/43d504502bf8c100aa52b799e9dfb721c296d5ed
|
2025-04-01T06:37:04.297427
| 2019-02-25T22:39:58
|
414341947
|
{
"authors": [
"FelixGail",
"IIIuminator"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1290",
"repo": "IIIuminator/EnQ",
"url": "https://github.com/IIIuminator/EnQ/issues/41"
}
|
gharchive/issue
|
App crash on unknown suggester
Songs starred from a suggester that is no longer available crashes the app.
As the goal of starred songs is to remember interesting songs and the app can potentially be used on different bot instances with incompatible providers, the best solution would probably be trying to use the started song, and if no suitable provider is found, the song is simply searched on available providers.
will add a redirect to a search on all providers once #42 is implemented
|
2025-04-01T06:37:04.305028
| 2017-09-27T08:04:59
|
260882483
|
{
"authors": [
"mgylling",
"ottonomy"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1291",
"repo": "IMSGlobal/openbadges-validator-core",
"url": "https://github.com/IMSGlobal/openbadges-validator-core/pull/172"
}
|
gharchive/pull-request
|
Add support for validating "version", "@language", and "related" props
Add support for validating "version", "@language", and "related" properties. For "related", a partial validation is done, only on embedded properties that are actually included, so that the related object may be minimal and only declare properties that distinguish the related version from the entity core to the test. Related objects are not fetched. Resolves #70.
LGTM
|
2025-04-01T06:37:04.319387
| 2022-03-08T11:44:18
|
1162556823
|
{
"authors": [
"odscjames"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1292",
"repo": "INDIGO-Initiative/database-app",
"url": "https://github.com/INDIGO-Initiative/database-app/pull/113"
}
|
gharchive/pull-request
|
data-quality-report: Improvements
Now we only do stats on public data anyway, that makes no sense.
Build in background and store in cache (instead of build on demand). Building once per day is fine. We are going to add more stats so build times will only get longer - currently 10s on my laptop.
Also noting there is an inefficiency here (we look up how many public projects there are repeatedly when building) but as it builds in the background I'm not bothered at this stage.
|
2025-04-01T06:37:04.349471
| 2021-07-23T13:11:35
|
951567501
|
{
"authors": [
"lesteve",
"ogrisel"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1293",
"repo": "INRIA/scikit-learn-mooc",
"url": "https://github.com/INRIA/scikit-learn-mooc/issues/414"
}
|
gharchive/issue
|
Add "advanced pipeline", missing value, imputing module, maybe more
So this is an ambitious goal, with some moving parts. We need to reevaluate depending on how fast we progress with other things.
This does not really fit in any existing module and this seems to long to put it in module 1.
Here are a few things that we could show:
advanced pipeline: as currently in the wrap-up quiz 1. Need more details.
missue value imputing
Maybe we could present splines and dealing with time features in this module:
https://scikit-learn.org/dev/auto_examples/applications/plot_cyclical_feature_engineering.html
This will require scikit-learn 1.0 which is not released yet though.
I created a MOOC 3.0 milestone, to make it less of a priority that MOOC 2.0 which should primarily be focused on a improving the existing material. We can always re-milestone issues based on progress.
|
2025-04-01T06:37:04.388570
| 2023-12-03T23:37:13
|
2022769643
|
{
"authors": [
"TheFaik",
"konsheng",
"lipind"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1295",
"repo": "IPECTER/LighterAPI",
"url": "https://github.com/IPECTER/LighterAPI/issues/6"
}
|
gharchive/issue
|
1.20.2 Support?
When will version 1.20.2 be supported?
pls add 1.20.4 supported
pls add 1.20.6 supported
pls supported
|
2025-04-01T06:37:04.504068
| 2020-07-27T16:07:41
|
666410768
|
{
"authors": [
"Uyadav207",
"pranavi79",
"raghavg27"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1297",
"repo": "ISTESRMNCR/CODE-CAMP-2020",
"url": "https://github.com/ISTESRMNCR/CODE-CAMP-2020/pull/101"
}
|
gharchive/pull-request
|
GrowTogether
Application for Registering Team to CODE CAMP 1.0
This pull request template helps you complete an application to the CODE CAMP 1.0 HACKATHON. Use the checklist below to verify that you have followed the instructions correctly.
(Put x inside [] like [x])
Select the type of Pull Request
[] Registration Pull request.
[x] Submission Pull request.
Checklist ✅
[] I have read the instructions on the README file before submitting my application and gone through CODE_OF_CONDUCT.
[] I have Submitted my team's details by making a folder of my team as instructed in How to send Pull Request page.
[] I have used the Markdown file template to add my information for the Hackathon.
[] I have given all the necessary details as mentioned in template index.md file and the details are correct and best to my knowledge.
[] I understand that a reviewer will merge my pull request after examining it or ask for changes in case needed.
[] I understand I should not tag or add a reviewer to this Pull Request.
[] I understand the Details added to the template will be used as a means of communication at the time of result declaration.
[] I have added the event to my calendar.
@pranavi79 , instructions not followed.
The file made are in Teams folder, Needs to be below Team folder
@pranavi79 please follow instructions again. Delete this repo, fork it again and NOW - drop your folder inside FINALIST folder and these is already a file - INDEX.MD, update it
|
2025-04-01T06:37:04.545037
| 2024-11-22T14:43:47
|
2683442289
|
{
"authors": [
"mtfishman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1299",
"repo": "ITensor/ITensorPkgSkeleton.jl",
"url": "https://github.com/ITensor/ITensorPkgSkeleton.jl/issues/9"
}
|
gharchive/issue
|
Generated code blocks in README don't have syntax highlighting
Literate is generating code blocks that are marked with @example rather than julia in the docs/README, which means they don't have syntax highlighting. Investigate how to change that (see https://fredrikekre.github.io/Literate.jl/v2/outputformats/#Configuration, https://documenter.juliadocs.org/stable/man/syntax/#reference-at-example).
Probably this will require generating the README.md and docs/src/index.md from the examples/README.jl separately with different Literate.markdown commands, and also making the Literate check workflow check both of those.
See also #14.
|
2025-04-01T06:37:04.603094
| 2022-10-06T16:54:56
|
1400017695
|
{
"authors": [
"MrRobz",
"andrewbrey"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1304",
"repo": "Icon-Shelf/icon-shelf",
"url": "https://github.com/Icon-Shelf/icon-shelf/pull/135"
}
|
gharchive/pull-request
|
feat(devcontainer): enable dev environments in GH Codespaces and VS Code remote containers
This project is marked as open to contribution for Hacktoberfest, and as such I figured it would be desirable for contributors to have a simplified means of getting up and running with the developer setup. With the addition of the devcontainer configuration, this project is now compatible with GitHub Codespaces as well as VS Code Remote Containers. Contributors who use one of these options now get all of the dev setup steps completed for them, and you get peace of mind that contributions will come from setups with the correct usage of eslint, prettier, etc.
EDIT:
Demo
https://vimeo.com/757666222
@MrRobz yes, it's incredibly cool - I have already recorded a video, but it's 25Mb and apparently GH will only let you attach videos of up to 10Mb - I'm in the process of re-encoding it so it's only 10Mb :sweat: (i'll attach once that's done).
@MrRobz I've attached a link to vimeo which is a long demonstration of how this works - the video is being transcoded still, but should be watchable in a few minutes :+1:
@MrRobz I've attached a link to vimeo (https://vimeo.com/757666222) which is a long demonstration of how this works - the video is being transcoded still, but should be watchable in a few minutes 👍
That was pretty amazing. One small suggestion. If there were some sample svg icons is a folder already present in the code-space it would make developing and testing the app lot more easier.
I'm thinking if we could have these icons in these links downloaded and present in a folder when opening the code space.
https://github.com/tailwindlabs/heroicons/tree/master/src/20/solid
https://github.com/Remix-Design/RemixIcon/tree/master/icons
@MrRobz I just looked into this, and I think it's probably a little out of scope for the purpose of this PR due to the fact that collection names are stored in IndexDB so there's no way for me to, during the container build, establish a new icon collection and add files to it.
I think you could modify the application code to do an initial scan of the icon-library directory and automatically create some collections based on any directories present there, and that would be the way you support pre-seeding the application state with some libraries.
Basically, I think it would be super helpful if the application could be defaulted to have some collections, but I think that's probably a bit further than this PR is meant to go. If you were able to merge this PR in, perhaps it could be a good first issue item for someone else in Hacktoberfest to make the required application code changes to support seeding the collection database and then dropping some svgs into the file system at container launch :+1:
@MrRobz I just looked into this, and I think it's probably a little out of scope for the purpose of this PR due to the fact that collection names are stored in IndexDB so there's no way for me to, during the container build, establish a new icon collection and add files to it.
I think you could modify the application code to do an initial scan of the icon-library directory and automatically create some collections based on any directories present there, and that would be the way you support pre-seeding the application state with some libraries.
Basically, I think it would be super helpful if the application could be defaulted to have some collections, but I think that's probably a bit further than this PR is meant to go. If you were able to merge this PR in, perhaps it could be a good first issue item for someone else in Hacktoberfest to make the required application code changes to support seeding the collection database and then dropping some svgs into the file system at container launch 👍
I understand. Lets do something simple. Clone 2 other repos with icons in it to a folder in say home or desktop in the code-space machine.
I tried adding these steps to the dockerfile but these folders aren't showing.
Could you help with this pls.
RUN cd /home && git clone https://github.com/Remix-Design/RemixIcon.git
RUN cd /home && git clone https://github.com/tailwindlabs/heroicons.git
@MrRobz Sounds like a great way to get most of the way there! I believe your changes should actually be fine, but you've cloned to the /home directory, and not /home/node, which is where your file explorer is open to - I think you should probably be cloning to /home/node.
Also, I believe the -C flag to git clone will let you skip having to cd into the clone directory 👍
@MrRobz Sounds like a great way to get most of the way there! I believe your changes should actually be fine, but you've cloned to the /home directory, and not /home/node, which is where your file explorer is open to - I think you should probably be cloning to /home/node.
Also, I believe the -C flag to git clone will let you skip having to cd into the clone directory 👍
That made it work. Thank you
|
2025-04-01T06:37:04.626456
| 2017-01-17T07:15:22
|
201193252
|
{
"authors": [
"leastprivilege",
"sriram5052"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1305",
"repo": "IdentityServer/IdentityServer4",
"url": "https://github.com/IdentityServer/IdentityServer4/issues/705"
}
|
gharchive/issue
|
How to update claims on refresh tokens refresh?
it's not an issue, its a question.
Setup
Identity server- HybridAndClientCredentials
scope - IdentityServerConstants.StandardScopes.OpenId,
IdentityServerConstants.StandardScopes.Email,
IdentityServerConstants.StandardScopes.Profile,
IdentityServerConstants.StandardScopes.OfflineAccess,
"api","offline_access"
UpdateAccessTokenClaimsOnRefresh=true,
AllowOfflineAccess = true,
I was trying to update the custom claims on the client, but couldn't update the claim to the identity server.
It would be great if you could point me at right direction for updating the claim.
whether the way i am trying to update the claim by refreshing token is right or not?
i have claim named facility_id, which has to be changed when the user change the default facility.
i have updated the facility_id in the client.but how to update it to the database through identity server.
Data management is not our problem - go to your database and update it yourself ;)
Also - if the data frequently changes - it is probably not a good candidate for a claim in a token.
Thanks for the answer, i thought identity will take care of updating claim to database. Now it's clear.
|
2025-04-01T06:37:04.640098
| 2021-03-06T17:27:11
|
823701085
|
{
"authors": [
"aperezdc",
"psaavedra",
"zdobersek"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1306",
"repo": "Igalia/WPEBackend-fdo",
"url": "https://github.com/Igalia/WPEBackend-fdo/issues/141"
}
|
gharchive/issue
|
WebKit WPE main/HEAD doesn't build using the master/HEAD of WPEBackend-FDO
... the same doesn't happen with WPEBackend-FDO 1.8.0
--- /home/bot/toolchain/sysroots/armv7at2hf-neon-poky-linux-gnueabi/usr/lib/pkgconfig/wpe-1.0.pc 2021-03-05 05:36:02.854345912 +0000
+++ /home/bot/toolchain_1.9.1/sysroots/armv7at2hf-neon-poky-linux-gnueabi/usr/lib/pkgconfig/wpe-1.0.pc 2021-03-05 00:56:37.717953845 +0000
@@ -5,7 +5,7 @@
Name: wpe-1.0
Description: The wpe library
-Version: 1.9.1
+Version: 1.8.0
Requires: xkbcommon
Cflags: -I${includedir}/wpe-1.0
Libs: -L${libdir} -lwpe-1.0
There are changes in the includes that could be the source of this errors. Here the differences between 1.8.1 and 1.9: https://paste.debian.net/1188147/. Probably related with f461fd4d306436bcefa0bdbf1821a191d7462c38.
commit f461fd4d306436bcefa0bdbf1821a191d7462c38
Author: Adrian Perez de Castro<EMAIL_ADDRESS>Date: Thu Nov 26 15:47:43 2020 +0200
Simplify public headers
Remove unneeded inclusions of libwpe headers, preferring forward
declarations of the needed types and make all the inclusions use double
quoted paths (to prefer local header versions, as opposed to installed
ones).
The issue only affects the Tools/wpe/backends in WebKit so far. For example, you still can build WPE by disabling the tools (-DENABLE_TOOLS=OFF)
../../Tools/wpe/backends/HeadlessViewBackend.cpp:85:22: error: ‘wpe_view_activity_state_visible’ was not declared in this scope
Full list of errors: https://paste.debian.net/1188149/
Can confirm this. For backends code in WebKit, a simple explicit include of <wpe/wpe.h> in ViewBackend.h works.
This was because of the recent cleanups in the WPEBackend-fdo headers, which no longer end up including <wpe/wpe.h>; I think we should make <wpe/fdo.h> include it again so the expectations of existing applications stay the same. I'll make a PR later today.
|
2025-04-01T06:37:04.645588
| 2015-04-25T15:52:35
|
70930135
|
{
"authors": [
"IgnaceMaes",
"NoobTW"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1307",
"repo": "IgnaceMaes/MaterialSkin",
"url": "https://github.com/IgnaceMaes/MaterialSkin/issues/45"
}
|
gharchive/issue
|
WndProc &HF030&
I tried to catch the maximize event but failed.
Also I can't disable control box and buttons.
Are you using the NuGet package? This should be fixed in the project, but I simply haven't pushed the update to NuGet yet.
Yes, I am using NuGet package. Maybe I should wait for your updates or compile your source on my own, right?
Thanks in advanced.
Sorry for the delay, but MaterialSkin 0.2.1 has been pushed to NuGet. (Including the ability to disable the control box buttons)
|
2025-04-01T06:37:04.688034
| 2023-11-04T08:48:55
|
1977240298
|
{
"authors": [
"IgorKowalczyk",
"kordnddn"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1308",
"repo": "IgorKowalczyk/majo.exe",
"url": "https://github.com/IgorKowalczyk/majo.exe/issues/659"
}
|
gharchive/issue
|
is this hostable in replit?
Environment
asd
Steps To Reproduce
asdw
Software Version
asd
Expected Behavior
asd
Actual Behavior
asd
Screenshots
asd
Severity
Trivial
Priority
Low
Type
Other
Reproducible
[X] Yes
[X] No
Additional Information
asd
From what I know, the bot should work well on the Replit, it just requires extra configuration. Only the dashboard does not work on the Replit, It returns errors during deployment.
However, I would like to point out that replit is not a supported hosting and errors that occur there will not be fixed.
|
2025-04-01T06:37:04.700369
| 2020-11-26T07:27:23
|
751355255
|
{
"authors": [
"michael-lachmann",
"radlinsky",
"renkun-ken"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1309",
"repo": "Ikuyadeu/vscode-R",
"url": "https://github.com/Ikuyadeu/vscode-R/issues/469"
}
|
gharchive/issue
|
How to work remotely with a node behind a login server
Is your feature request related to a problem? Please describe.
I am working remotely with a cluster. I log into the login server, and then request a node, to which I then log in. I manage to run remotely on the server, and see plots and help files. When I connect to the node, I do manage to start radian and attach. However, plots and help are not displayed.
I set the environment variable TMPDIR so that tempdir() points to a directory that is mounted on the login server. And that directory does work when used from the server. I can see that when I'm on a node, plots are generated inside that directory (and I can manually see them). But I can't have them automatically appear.
I did try to do ssh port forwarding with -R when loging into the node from the server. That didn't help. I'm not I did it correctly.
Describe the solution you'd like
I'd like to be able to work remotely in that setup
Describe alternatives you've considered
See above
Additional context
You could manually start radian and attach, and the vscode status bar could show the pid of the attached session?
Would you like to share ls.str(.vsc) in your attached R session?
Below is the output. $HOME is my home directory, and $WORk is a work directory (I did search/replace to the output)
attach : function ()
capture_str : function (object)
check_null_dev : function ()
dataview_data_type : function (x)
dataview_table : function (data)
dir_extension : chr "$HOME/.vscode-R"
dir_plot_history : chr "$WORK/seir_regression/TMP/RtmpvxGtqk/vscode-R/images"
dir_session : chr "$WORK/seir_regression/TMP/RtmpvxGtqk/vscode-R"
get_timestamp : function ()
globalenv_file : chr "$WORK/seir_regression/TMP/RtmpvxGtqk/vscode-R/globalenv.json"
globalenv_lock_file : chr "$WORK/seir_regression/TMP/RtmpvxGtqk/vscode-R/globalenv.lock"
homedir : chr "$HOME"
new_plot : function ()
null_dev_id : Named int 2
null_dev_size : num [1:2] 10.1 10.1
path_to_uri : function (path)
pid : int 190091
plot_file : chr "$WORK/seir_regression/TMP/RtmpvxGtqk/vscode-R/plot.png"
plot_history_file : NULL
plot_lock_file : chr "$WORK/seir_regression/TMP/RtmpvxGtqk/vscode-R/plot.lock"
plot_updated : logi FALSE
print.help_files_with_topic : function (h, ...)
rebind : function (sym, value, ns)
request : function (command, ...)
request_browser : function (url, title, ..., viewer)
request_file : chr "$HOME/.vscode-R/request.log"
request_lock_file : chr "$HOME/.vscode-R/request.lock"
rstudioapi_enabled : function ()
show_browser : function (url, title = url, ..., viewer = getOption("vsc.browser", "Active"))
show_dataview : function (x, title, viewer = getOption("vsc.view", "Two"))
show_globalenv : logi TRUE
show_page_viewer : function (url, title = NULL, ..., viewer = getOption("vsc.page_viewer", "Active"))
show_plot : logi TRUE
show_view : logi TRUE
show_viewer : function (url, title = NULL, ..., viewer = getOption("vsc.viewer", "Two"))
show_webview : function (url, title, ..., viewer)
tempdir : chr "$WORK/seir_regression/TMP/RtmpvxGtqk"
unbox : function (x)
update_globalenv : function (...)
update_plot : function (...)
wd : chr "$WORK/seir_regression"
vscode status bar could show the pid of the attached session when the session is attached?
When I'm on the main server, the pid is updated. When I'm on the node it isn't.
Note that the remote ssh is to the server, not to the node. The files are served from the server. Since the directory structure on both is identical (other than /tmp), I can edit files and run things. But remote is connected to the initial server.
believe this is the same issue I'm having in #552 .. like @michael-lachmann I too want to be able to use r session watcher on a compute node, not the login node that I initially Remote-SSH into.
I added this line to my .Renviron to point the tmp dir to my $HOME directory:
TMPDIR="/home/cmr46993/tmp"
I think we're almost to a fix because if I use R: Create R Terminal now, the plots pngs go here:
/home/cmr46993/tmp/RtmpNZ3ffC/
BUT if I launch a new Terminal manually, connect to a compute node, and launch radian, the plots pngs go here:
/home/cmr46993/tmp/RtmpMmBQiz/
And then if I launch yet another Terminal manually, I see a third tmp directory 😆
I hope this is a simple fix @renkun-ken because this will be a game changer for developing code for me and apparently others too :)
Actually, perhaps this issue is related to:
https://github.com/microsoft/vscode-remote-release/issues/1722
There's currently no easy way to Remote SSH into an interactive session on a compute node.
|
2025-04-01T06:37:04.711020
| 2023-08-06T18:22:08
|
1838298629
|
{
"authors": [
"Im-Beast"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1310",
"repo": "Im-Beast/http_benchmarks",
"url": "https://github.com/Im-Beast/http_benchmarks/issues/5"
}
|
gharchive/issue
|
bug: updated Fastro weirdly hangs tests
Newer Fastro versions weirdly hang tests when I run them.
I don't know what exactly has changed, but that's no longer the case.
Most probably it was something on Fastro's end, however it seems to be resolved now
|
2025-04-01T06:37:04.715425
| 2024-08-06T21:37:05
|
2451732721
|
{
"authors": [
"colaub",
"sonoro1234",
"zaafar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1311",
"repo": "ImGuiNET/ImGui.NET",
"url": "https://github.com/ImGuiNET/ImGui.NET/issues/491"
}
|
gharchive/issue
|
Support linux older than glibc2.29
On linux lincimgui.so is built against glibc2.29.
Log from a RHEL8 / Rocky8 which based on glibc2.28 :
/lib64/libm.so.6: version `GLIBC_2.29' not found (required by [...]/bin/Debug/net8.0/runtimes/linux-x64/native/libcimgui.so)
rhel8/rocky8 are long-term distributions, that's why it could be interesting to have another linux version in the nuget package.
Is it an option?
Thank you for considering the question.
Colin.
that's a https://github.com/cimgui/cimgui question. Please post it there otherwise correct me if I am wrong.
cimgui is not built at all. You choose compiler.
oh i see, @colaub can you help me figure out which linux version i should use such that I get glibc2.28
ref: https://github.com/ImGuiNET/ImGui.NET-nativebuild/blob/master/.github/workflows/build.yml#L29
|
2025-04-01T06:37:04.781684
| 2023-09-27T16:38:59
|
1915956206
|
{
"authors": [
"fedorov"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1312",
"repo": "ImagingDataCommons/Cloud-Resources-Workflows",
"url": "https://github.com/ImagingDataCommons/Cloud-Resources-Workflows/issues/29"
}
|
gharchive/issue
|
Reorganize repository structure
We can probably have a more intuitive and clean look if we reorganize as follows:
Dockerfiles -> docker
sqlQueries (I would also rename this to "queries"), Notebooks, SevenBridges, Terra -> workflows
architectureDiagrams, pricingOptimization, sampleManifests -> docs
Thinking more about this, maybe "resources" is a better name than "docs" for the content mentioned above.
|
2025-04-01T06:37:04.791373
| 2018-08-21T13:13:15
|
352532352
|
{
"authors": [
"nemobis",
"richard-orr"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1313",
"repo": "Impactstory/oadoi",
"url": "https://github.com/Impactstory/oadoi/issues/95"
}
|
gharchive/issue
|
False negative for University of Illinois Libraries
https://doi.org/10.5210/fm.v18i6.4340 is available from the publisher at http://journals.uic.edu/ojs/index.php/fm/article/view/4340/3687 , although not in PDF.
API:
{
"best_oa_location": null,
"data_standard": 2,
"doi": "10.5210/fm.v18i6.4340",
"doi_url": "https://doi.org/10.5210/fm.v18i6.4340",
"genre": "journal-article",
"is_oa": false,
"journal_is_in_doaj": false,
"journal_is_oa": false,
"journal_issns": "1396-0466",
"journal_name": "First Monday",
"oa_locations": [],
"published_date": "2013-06-03",
"publisher": "University of Illinois Libraries",
"title": "Assigning Wikipedia editing: Triangulation toward understanding university student engagement",
"updated": "2018-06-17T07:18:14.930443",
"year": 2013,
"z_authors": [
{
"family": "Roth",
"given": "Amy"
},
{
"family": "Davis",
"given": "Rochelle"
},
{
"family": "Carver",
"given": "Brian"
}
]
}
https://support.unpaywall.org/public/tickets/d3736be7c80f599113502fe94c78c682436e7630ab317fdc4911de3aa11d9097
|
2025-04-01T06:37:04.796423
| 2018-07-11T08:56:15
|
340150602
|
{
"authors": [
"Kinle",
"ccarpenter04",
"devender-yadav"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1314",
"repo": "Impetus/Kundera",
"url": "https://github.com/Impetus/Kundera/issues/1021"
}
|
gharchive/issue
|
Bug in com.impetus.kundera.classreading.Reader - getResourceIterator
A Spring project was created with name that have spaces.
When persistence.xml doesn't have any configured, it scans for entity classes in classpath.
Since the path is URL encoded, spaces converts to %20 and the URL String doesn't seem to read the file.
Suggestion: String urlString = url.toString(); can be changed to url.getFile() in Line 137
@Kinle url.getFile() will also return string with spaces converts to %20
I've also encountered this issue, it'd be nice if someone could look into resolving it.
|
2025-04-01T06:37:04.876391
| 2022-02-18T20:31:24
|
1143723378
|
{
"authors": [
"sysarcher",
"wagmarcel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1315",
"repo": "IndustryFusion/DigitalTwin",
"url": "https://github.com/IndustryFusion/DigitalTwin/pull/119"
}
|
gharchive/pull-request
|
Add E2E test to Digital Twin Repo
This commit containts:
Change of registry variables in Scorpio deployments and Digital Twin created deployment to
allow easier usage of local registry (to avoid pulling from private docker repo)
Bats/Detik based Kubernetes tests to check k8s resource status
Tests check operators and horizontal platform status
Automated the ingress setup for keycloak.local, alerta.local, ngsild.local urls
and respective updates in README.md
Automated K3s setup with two nodes and local registry
Linting of bats test files
Take mangement of postgres operator out of olm (since version on operatorhub is too old)
Upgrade of Postgres-operator to 1.7.1
Kubernetes version whcih is tested is 1.22
Closes #118
@wagmarcel is this the latest run for the CI? Seems to be failing
https://github.com/wagmarcel/DigitalTwin/actions/workflows/k8s-tests.yaml
@wagmarcel is this the latest run for the CI? Seems to be failing
https://github.com/wagmarcel/DigitalTwin/actions/workflows/k8s-tests.yaml
No. That was my tests runs. The test of this PR is already listed in this PR - see the 2nd check:.
https://github.com/IndustryFusion/DigitalTwin/runs/5253256772?check_suite_focus=true
My bad... I clicked on the branch at the top and ended up on the fork.
|
2025-04-01T06:37:04.897511
| 2024-06-11T20:12:01
|
2347217375
|
{
"authors": [
"Infinidoge",
"Misterio77",
"YaroKasear",
"frantisekhanzlikbl"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1316",
"repo": "Infinidoge/nix-minecraft",
"url": "https://github.com/Infinidoge/nix-minecraft/issues/70"
}
|
gharchive/issue
|
Document that the module's files option does env substitution
Currently, neither the readme nor the option docs mention that services.minecraft-servers.servers.<name>.files performs environment substitution.
I just got bitten by this when working with binaries that were getting mysteriously corrupted, and the module mutating them was not something that I expected.
The ideal case scenario is IMHO that this is disabled by default and opt-in on each file, or at least opt-able out of, but it should really be documented. I can make a PR for either if it helps. :slightly_smiling_face:
It's technically documented under the environmentFile option, however I do agree that it should be documented under files as well. Making it toggleable should be pretty straight forward too, but would likely need some changes to the mkFiles script. Feel free to open PRs if you have the time :)
hi, sorry for not replying sooner. I kept thinking I'll find time for this some when, but I'm starting to see I won't be able to anytime soon. on top of that, I'm somewhat considering moving my infra over from systemd to nomad. while NixOS modules are superior in many ways, I'm starting to feel the need for a more flexible orchestrator, and would also prefer to spend less time worrying about every other NixOS service being completely unsandboxed. :)
I just found this without seeing this bug report.
This issue of doing env substitution blindly makes the files option unusable for setting up mods. And since symlinks puts a symlink directly to the Nix Store, it doesn't look like there's a proper way to declaratively install mods, at least not the way the README suggests.
The symlinks option is the proper way to declaratively install mods. Mods are binaries, so symlinking them to the nix store is perfectly fine and generally preferable, as it reduces on the space cost of copying files.
The symlinks option is the proper way to declaratively install mods. Mods are binaries, so symlinking them to the nix store is perfectly fine and generally preferable, as it reduces on the space cost of copying files.
The issue with this is that it's a symlink to the Nix Store and Paper (And probably other server codebases.) likes to create the plugin configs in the same directory as the jars... which it can't do since the Nix Store is read-only. This causes Paper to bail with a "read-only file system" because it can't write to the plugins directory symlink makes.
Unless there's some other way to get this to work? Is there a setting I overlooked where Paper can create plugin configs outside the plugins directory which is a symlink to the Nix Store?
<Tired opinion (not because of you but because I've dealt with this before)>
As far as I am concerned, this is because Paper is bad. Paper, and other plugin-centered projects, are the only ecosystem I've seen that feels like it has any dominion over the folder dedicated to putting binaries from the user. No other modloader does this. The fact that a mod port of a plugin did this (LuckPerms) frustrated me to the point where I just found an alternative instead.
</Tired opinion>
The proper way to work around this is to make symlinks individually in the manner you did with the files option, as opposed to symlinking directly to the mods folder. (I.e., "mods/mod.jar" = ...;) There isn't a better option because Paper tries to write to a directory it has no business writing to. You can, at minimum, do this a bit easier with mapAttrs.
Hmm, that's not a perspective I thought of. I honestly don't think it's a problem, but that's why it's opinion, I suppose.
What server do you recommend?
I exclusively use mod loaders and not plugins, so the use case is a bit different. For my part, I use Fabric and Quilt on all of my servers, including vanilla-compatible ones. Forge is also a possibility for highly modded ones, but since that isn't currently packaged (see #15), I just stick with the textiles.
Will be fixed by #116
|
2025-04-01T06:37:04.984890
| 2024-02-13T09:31:01
|
2131871547
|
{
"authors": [
"bclenet",
"cmaumet"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1317",
"repo": "Inria-Empenn/narps_open_pipelines",
"url": "https://github.com/Inria-Empenn/narps_open_pipelines/issues/166"
}
|
gharchive/issue
|
[L7J7] Pipeline reproduction (SPM - deriv)
Softwares
SPM12
Input data
derivatives (fMRIprep)
Additional context
List of tasks
Please tick the boxes below once the corresponding task is finished. :+1:
[ ] :ok_hand: A maintainer of the project approved the issue, by assigning a :checkered_flag:status: ready for dev label to it.
[ ] :deciduous_tree: Create a branch on your fork to start the reproduction.
[ ] :sunrise: Create a file team_{team_id}.py inside the narps_open/pipelines/ directory. You can use a file inside narps_open/pipelines/templates as a template if needed.
[ ] :inbox_tray: Create a pull request as soon as you completed the previous task.
[ ] :brain: Write the code for the pipeline, using Nipype and the file architecture described in docs/pipelines.md.
[ ] :blue_book: Make sure your code is documented enough.
[ ] :snake: Make sure your code is explicit and conforms with PEP8.
[ ] :microscope: Create tests for your pipeline. You can use files in tests/pipelines/test_team_* as examples.
[ ] :microscope: Make sure your code passes all the tests you created (see docs/testing.md).
@icorouge: I am adding you as assignee on this issue (just for the time of the hackathon) so that we can more easily see which pipelines are open for new contributions in https://github.com/orgs/Inria-Empenn/projects/1/views/1
Correlation results with 108 subjects : [0.93, 0.94, 0.93, 0.94, 0.94, 0.93, 0.94, 0.93, 0.95] with commit ba2b3dd
|
2025-04-01T06:37:04.988506
| 2022-11-22T13:33:29
|
1459883001
|
{
"authors": [
"Remi-Gau",
"cmaumet"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1318",
"repo": "Inria-Empenn/narps_open_pipelines",
"url": "https://github.com/Inria-Empenn/narps_open_pipelines/pull/21"
}
|
gharchive/pull-request
|
[ENH] add model U26C
[x] add bids stats model for U26C
Had to make some choices on how to define parametric regressors for SPM.
This seems to be handled differently for SPM in nipype: https://nipype.readthedocs.io/en/latest/api/generated/nipype.algorithms.modelgen.html#module-nipype.algorithms.modelgen
So I did put them in the software options.
I will try to do the same for the model from our teams to see if I can make this work without turning it into a headache.
Not sure if this is helpful here (and Remi you must know about this) but keeping a note so that I remember too...
Just heard about NARPS BIDS-stats-models example: https://bids-standard.github.io/model-zoo/exhibits/narps/model-narps_smdl.html from Alejandro.
Yup Rotem and I created this one and I think I took it as a first step to start building those ones: https://github.com/Inria-Empenn/narps_open_pipelines/tree/main/narps_open/models
|
2025-04-01T06:37:05.016178
| 2022-12-22T14:44:36
|
1508013959
|
{
"authors": [
"tbirdso"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1319",
"repo": "InsightSoftwareConsortium/ITKRemoteModuleBuildTestPackageAction",
"url": "https://github.com/InsightSoftwareConsortium/ITKRemoteModuleBuildTestPackageAction/issues/38"
}
|
gharchive/issue
|
BUG: Linux internal build failure is silent
An issue was observed in the HASI repository where Linux wheel builds failed but jobs were marked as having completed successfully. MacOS and Windows wheels failed and were correctly marked as failed.
https://github.com/KitwareMedical/HASI/actions/runs/3750380351/jobs/6370024283
Steps to Reproduce
The HASI module depends on the ITKBoneEnhancement project. If ITKBoneEnhancement build outputs are not provided then HASI wheels are expected to fail.
Fetch ITK build artifacts (without ITKBoneEnhancement)
Try building HASI with ITKPythonPackage build scripts
Observed behavior
The HASI build begins inside the appropriate docker container and fails at the config stage. However, the failure does not seem to propagate from the container back to the host.
A similar issue was encountered in https://github.com/InsightSoftwareConsortium/ITKRemoteModuleBuildTestPackageAction/issues/52 where a package failed to install inside the docker image, resulting in a build failure that is not caught by the GitHub Actions runner and subsequently reported as a success.
A minimum path to address this issue is to validate that the expected wheel / number of wheels is present in dist/ after the build process completes and before artifact upload is attempted.
Workaround introduced in https://github.com/InsightSoftwareConsortium/ITKRemoteModuleBuildTestPackageAction/pull/53 so that a job fails if expected wheel output is not produced. Closing for now with the potential to reopen if this fix proves insufficient for tracing errors.
|
2025-04-01T06:37:05.023122
| 2017-03-10T11:25:26
|
213316165
|
{
"authors": [
"JPawsey45",
"fujiy"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1320",
"repo": "InstaSharp/InstaSharp",
"url": "https://github.com/InstaSharp/InstaSharp/issues/132"
}
|
gharchive/issue
|
Using Instasharp without a session
I am creating an Umbraco package, to do this I cannot use any ActionResult or a Session could you help advise an alternative so that I can have a 'Stateless' controller.
Thanks
The library is not tied to Sessions. You can store the data in any place.
|
2025-04-01T06:37:05.032116
| 2017-01-06T06:49:11
|
199135754
|
{
"authors": [
"PhilCai1993",
"rnystrom"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1321",
"repo": "Instagram/IGListKit",
"url": "https://github.com/Instagram/IGListKit/issues/387"
}
|
gharchive/issue
|
Question about an assert in IGListAdapter.m
This is a little strange.
#if DEBUG
for (id object in objects) {
IGAssert([object isEqual:object], @"Object instance %@ not equal to itself. This will break infra map tables.", object);
}
#endif
When will [object isEqual:object] return NO? object won't be nil in the for-loop.
Could you please attach an example for this?
@PhilCai1993 before ab890fc6070f170a2db5a383a6296e62dcf75678 when we used -isEqual: to test diffable objects, we had a user bug where someone's equality implementation accidentally returned NO even when the object was the same pointer. The error was really difficult to track down, so we added this assert so it wont even happen again.
However, since we changed the equality method to -isEqualToDiffableObject:, this assert should be updated.
But I'm still confused...
How could [obj isEqualToDiffableObject:obj ] return NO?
-(BOOL)isEqualToDiffableObject:(id)object {
// it returns NO when object === self, how could it happen?
}
@PhilCai1993 Purely by developer error. Non-obvious example:
@interface MyClass: NSObject
@property NSString *text;
@end
@implementation MyClass
- (BOOL)isEqualToDiffableObject:(id)object {
if (![object isKindOfClass:[MyClass class]]) { return NO; }
return [self.text isEqualToString:[object text]];
}
@end
Then you create and compare:
MyClass *left = [MyClass new];
NSLog(@"%zi", [left isEqualToDiffableObject:left]);
// prints "0" (aka NO)
That's b/c passing a message to nil returns a 0 value (NO in this case). Obviously this is fixed with a self == object check, but people make mistakes. The assert just makes a tricky-to-catch mistake impossible while debugging.
Thanks a lot!
|
2025-04-01T06:37:05.041533
| 2024-01-18T05:58:27
|
2087565795
|
{
"authors": [
"haofanwang",
"renderless",
"wangqixun"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1322",
"repo": "InstantID/InstantID",
"url": "https://github.com/InstantID/InstantID/issues/9"
}
|
gharchive/issue
|
About Identity Similarity...
In technical report, Fig. 6, "Jackie Chan" does not looks like reference image especially his nose. I supposed antelopev2 model should able to extract Jackie Chan very well as his images should be in Glint360k training data. Is this the limitation of face id encoder?
Also, the picture quality seems over saturated. Is it because of SDXL base model or your prompt?
It should be related to the weight scale. We don't tune the parameters carefully. Anyway, talk is easy, we will show you the code.
In technical report, Fig. 6, "Jackie Chan" does not looks like reference image especially his nose. I supposed antelopev2 model should able to extract Jackie Chan very well as his images should be in Glint360k training data. Is this the limitation of face id encoder?
Also, the picture quality seems over saturated. Is it because of SDXL base model or your prompt?
It may be a problem with prompt. I changed the prompt and base model to generate a new image. At the same time, our model is constantly being optimized.
@wangqixun which model were using here?
base model = https://civitai.com/models/43977?modelVersionId=227916
prompt = "cinema 4d render, high contrast, vibrant and saturated, sico style, dark and moody close-up shot of a handsome Saint-Pierrais man with a tired expression, (renaissance theme:1.1), colorful northern warrior, (glowing eyes:1.05), dynamic pose, hooded robe, surrounded by magical glow, floating ice shards, snow crystals, cold, windy background, frozen natural landscape in background cinematic atmosphere, highly detailed, sharp focus, intricate design, 3d, unreal engine, octane render, CG best quality, highres, photorealistic, dramatic lighting, artstation, concept art, cinematic, epic Steven Spielberg movie still, sharp focus, smoke, sparks, art by pascal blanche and greg rutkowski and repin, trending on artstation, hyperrealism painting, detailed character design, matte painting, 4k resolution"
neg prompt = "asian, (worst quality, low quality, thumbnail:1.4), signature, artist name, web address, cropped, jpeg artifacts, watermark, username, collage, grid, nude, topless, nsfw, naked, nipples"
The style is not very stable. Generate the image again.
@wangqixun which model were using here?
base model = https://civitai.com/models/43977?modelVersionId=227916
prompt = "cinema 4d render, high contrast, vibrant and saturated, sico style, dark and moody close-up shot of a handsome Saint-Pierrais man with a tired expression, (renaissance theme:1.1), colorful northern warrior, (glowing eyes:1.05), dynamic pose, hooded robe, surrounded by magical glow, floating ice shards, snow crystals, cold, windy background, frozen natural landscape in background cinematic atmosphere, highly detailed, sharp focus, intricate design, 3d, unreal engine, octane render, CG best quality, highres, photorealistic, dramatic lighting, artstation, concept art, cinematic, epic Steven Spielberg movie still, sharp focus, smoke, sparks, art by pascal blanche and greg rutkowski and repin, trending on artstation, hyperrealism painting, detailed character design, matte painting, 4k resolution"
neg prompt = "asian, (worst quality, low quality, thumbnail:1.4), signature, artist name, web address, cropped, jpeg artifacts, watermark, username, collage, grid, nude, topless, nsfw, naked, nipples"
The style is not very stable. Generate the image again.
Correct it
base model = https://civitai.com/models/84040?modelVersionId=196039
|
2025-04-01T06:37:05.112583
| 2020-05-20T10:13:47
|
621639706
|
{
"authors": [
"Taggotty",
"maxammann"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1323",
"repo": "Integreat/integreat-webapp",
"url": "https://github.com/Integreat/integreat-webapp/pull/336"
}
|
gharchive/pull-request
|
WEBAPP-562: Test content not found error poi
This pull request belongs to an issue on our bugtracker.
You can find it there by looking for an issue with the key which is mentioned in the title of this pull request.
It starts with the keyword WEBAPP.
The changes in FailureSwitcher are just what I think fits better to our mindset to prefer testing functunality. If you don't like it feel free to object :)
Examples for non-snapshot tests:
it('should show correct text', () => {
const wrapper = shallow(<MyComponent />);
expect(wrapper.text().includes('my text')).toBe(true);
});
const wrapper = shallow(<div><button className='btn btn-primary'>OK</button></div>);
const button = wrapper.find('.btn');
expect(button.text()).to.be.eql('OK');
|
2025-04-01T06:37:05.114385
| 2018-04-20T03:38:25
|
316116477
|
{
"authors": [
"rkamudhan"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1324",
"repo": "Intel-Corp/multus-cni",
"url": "https://github.com/Intel-Corp/multus-cni/issues/56"
}
|
gharchive/issue
|
Mismatch between the Self Link in apiserver - K8s plumbing working group documentation and the implementation
apiVersion: apiextensions.k8s.io/v1beta1
kind: CustomResourceDefinition
metadata:
name: networks.kubernetes.cni.cncf.io
spec:
group: kubernetes.cni.cncf.io
version: v1
scope: Namespaced
names:
plural: networks
singular: network
kind: Network
shortNames:
- net
Implementation: https://github.com/Intel-Corp/multus-cni/blob/122dbfb345ae4a13fd1d592723a0ba1603278dd9/multus/multus.go#L386
@s1061123 @dougbtv Creating a patch for it.
#58 Fixed in the PR
|
2025-04-01T06:37:05.116225
| 2020-06-19T13:10:36
|
641957789
|
{
"authors": [
"alexelizarov",
"dmitryermilov"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1325",
"repo": "Intel-Media-SDK/MediaSDK",
"url": "https://github.com/Intel-Media-SDK/MediaSDK/pull/2178"
}
|
gharchive/pull-request
|
[hevce] Enable QP modulation
Pass QP mod parameters to driver
@saosipov , @lakulako , @vilichev , please review
Need to wait while typo would be fixed in libva interface.
Now it contain filed hierachical_flag but it should be hieraRchical_flag.
Relates to this https://github.com/intel/libva/issues/429
|
2025-04-01T06:37:05.118933
| 2023-03-16T22:52:17
|
1628414799
|
{
"authors": [
"IntelOrca",
"Shfan3"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1326",
"repo": "IntelOrca/biorand",
"url": "https://github.com/IntelOrca/biorand/issues/301"
}
|
gharchive/issue
|
RE1 closes windowed or full screen at beginning
It closes when you first enter the mansion or when you go into the dining room. Never had RE1 issues before this update the seed is
R710-XVX3-XNE7603LCF020149K49K0WZZZZZZZZZZ00Z
I was playing as Jill.
Does it happen on vanilla as well or just when you load the biorand mod?
Does it happen if you select some NPCs and BGMs, or disable those two things?
it doesn't happen on vanilla. It happens with and without the NPCs and BGMs disabled
Does it happen if use an older version of biorand?
No everything is fine with an older version of BioRand.
Unfortunately I can't reproduce it, so I am not sure why it is crashing for you.
ok the 3.01 update fixed it brother so I'm gonna close this
|
2025-04-01T06:37:05.156913
| 2023-01-05T05:16:47
|
1520090171
|
{
"authors": [
"MartyG-RealSense",
"weicuiting"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1327",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/11293"
}
|
gharchive/issue
|
The accuracy of the distance between 2 points is high in the center but is low on the edge, how can I improve it?
Before opening a new issue, we wanted to provide you with some useful suggestions (Click "Preview" above for a better view):
Consider checking out SDK examples.
Have you looked in our documentations?
Is you question a frequently asked one?
Try searching our GitHub Issues (open and closed) for a similar issue.
All users are welcomed to report bugs, ask questions, suggest or request enhancements and generally feel free to open new issue, even if they haven't followed any of the suggestions above :)
Required Info
Camera Model
D435i
Operating System & Version
Win10
Language
python
Issue Description
Hi, sorry to bother you. I have read many issues but don't find the proper solution. I'm trying to measure the diameters of rebars in the RGB photos taken by realsense D435i. There is the problem that the accuracy of diameters is high for the rebars in the center but is low for the rebars on the dege As shown in the following figure(the reality is on the up and left, the calculation is on the down and right using the Euclidean distance).
I'm thinking is there any key point ignored. I don't know whether is the problem that the pointcloud gained by python is not aligned correctly, because the rebars seems not to lie on the highest plcace and have a bias.
This is the code I transport the points from pixels to the camera coordinates(using the method of point cloud):
import pyrealsense2 as rs
import numpy as np
import cv2
''' camera setting '''
pipeline = rs.pipeline()
config = rs.config()
config.enable_stream(rs.stream.depth, 1280,720, rs.format.z16, 30)
config.enable_stream(rs.stream.color, 1920,1080, rs.format.bgr8, 30)
profile = pipeline.start(config)
pc = rs.pointcloud()
points = rs.points()
#Define filters
#Decimation:
decimation = rs.decimation_filter()
#Depth to disparity
depth_to_disparity = rs.disparity_transform(True)
disparity_to_depth = rs.disparity_transform(False)
#Spatial:
spatial = rs.spatial_filter()
spatial.set_option(rs.option.holes_fill, 0) # between 0 and 5 def = 0
spatial.set_option(rs.option.filter_magnitude, 2) # between 1 and 5 def=2
spatial.set_option(rs.option.filter_smooth_alpha, 0.5) # between 0.25 and 1 def=0.5
spatial.set_option(rs.option.filter_smooth_delta, 20) # between 1 and 50 def=20
#Temporal:
temporal = rs.temporal_filter()
temporal.set_option(rs.option.filter_smooth_alpha, 0.4)
temporal.set_option(rs.option.filter_smooth_delta, 20)
colorizer = rs.colorizer()
#Get info about depth scaling of the device
depth_sensor = profile.get_device().first_depth_sensor()
depth_scale = depth_sensor.get_depth_scale()
print("Depth Scale is: " , depth_scale)
#align to color
align_to = rs.stream.color
align = rs.align(align_to)
def get_aligned_images():
frames = pipeline.wait_for_frames()
#Apply filters
pc_filtered = decimation.process(frames)
pc_filtered = depth_to_disparity.process(pc_filtered)
pc_filtered = spatial.process(pc_filtered)
pc_filtered = temporal.process(pc_filtered)
pc_filtered = disparity_to_depth.process(pc_filtered).as_frameset()
#Align the depth frame to color frame
aligned_frames = align.process(pc_filtered)
aligned_depth_frame = aligned_frames.get_depth_frame()
aligned_color_frame = aligned_frames.get_color_frame()
img_color = np.asanyarray(aligned_color_frame.get_data())
img_depth = np.asanyarray(aligned_depth_frame.get_data())
aligned_depth_color_frame = colorizer.colorize(aligned_depth_frame)
img_depth_mapped = np.asanyarray(aligned_depth_color_frame.get_data())
return img_color, img_depth, img_depth_mapped, aligned_color_frame, aligned_depth_frame, aligned_frames
def get_3d_camera_coordinate(depth_pixel, aligned_color_frame, aligned_depth_frame, aligned_frames):
x = np.round(depth_pixel[1]).astype(np.int64)
y = np.round(depth_pixel[0]).astype(np.int64)
#pointcloud
pc.map_to(aligned_color_frame)
points = pc.calculate(aligned_depth_frame)
points.export_to_ply("../frame_test.ply", aligned_color_frame)
vtx = np.asanyarray(points.get_vertices())
#print('vtx_before_reshape: ', vtx.shape)
vtx = np.reshape(vtx, (1080, 1920, -1))
#print('vtx_after_reshape: ', vtx.shape)
camera_coordinate = vtx[y][x][0]
#print ('camera_coordinate: ',camera_coordinate)
dis = camera_coordinate[2]
return dis, camera_coordinate
Hi @weicuiting As the Decimation filter will be reducing the resolution of the depth image, are measurements more accurate if you comment out the Decimation filter, please?
It's really appreciate for the quick apply. I just tried to comment out the Decimation filter, and the trend did not change(low accuracy onthe edge) even worse.
If the measurements are accurate at the center but become increasingly inaccurate when moving towards the edge of the image, this is usually because of an inaccuracy in the depth-color alignment. You seem to have applied alignment correctly in your script though and placed the align process after the post-processing filter list as Intel recommend.
Does accuracy improve if you align color to depth instead of depth to color by changing the align_to instruction from 'color' to 'depth'.
align_to = rs.stream.depth
As you are using pc.calculate to generate the point cloud and map_to to map color onto the depth points, it may not actually be necessary to use align_to to align depth to color.
Thank you for the responce! I tried 'align_to = rs.stream.depth', the trend still didn't changed. More important, it is not suitbale for my program to use 'align_to = rs.stream.depth' . This would result in the decreasing of RGB resolution and the black edges which affects the segmentation accuracy of rebars in the RGB photos(as the green masks shown in the following pictures).
Are you able to comment out the align instructions and let map_to perform the pointcloud alignment as suggested above at https://github.com/IntelRealSense/librealsense/issues/11293#issuecomment-1371935666
I comment out the align part, icluding 3 lines:
align to color
#align_to = rs.stream.color ## commen out(1)
#align = rs.align(align_to) ## commen out(2)
def get_aligned_images():
frames = pipeline.wait_for_frames()
#Apply filters
pc_filtered = decimation.process()
pc_filtered = depth_to_disparity.process(pc_filtered)
pc_filtered = spatial.process(pc_filtered)
pc_filtered = temporal.process(pc_filtered)
pc_filtered = disparity_to_depth.process(pc_filtered).as_frameset()
#Align the depth frame to color frame
#aligned_frames = align.process(pc_filtered) ## commen out(3)
aligned_depth_frame = pc_filtered.get_depth_frame()
aligned_color_frame = pc_filtered.get_color_frame()
img_color = np.asanyarray(aligned_color_frame.get_data())
img_depth = np.asanyarray(aligned_depth_frame.get_data())
aligned_depth_color_frame = colorizer.colorize(aligned_depth_frame)
img_depth_mapped = np.asanyarray(aligned_depth_color_frame.get_data())
return img_color, img_depth, img_depth_mapped, aligned_color_frame, aligned_depth_frame, aligned_frames
But in the process of calculate the 'vtx', an IndexError occurred: index 1578 is out of bounds for axis 0 with size 640, where the 1578 is from RGB while the 640 is from depth. Should I change the x and y pixel values according to the resolution rate(rate x = 640/1920, rate y = 360/1080)?
def get_3d_camera_coordinate(depth_pixel, aligned_color_frame, aligned_depth_frame, aligned_frames):
x = np.round(depth_pixel[1]).astype(np.int64)
y = np.round(depth_pixel[0]).astype(np.int64)
# 计算点云
pc.map_to(aligned_color_frame)
points = pc.calculate(aligned_depth_frame)
#points.export_to_ply("../frame_test.ply", aligned_color_frame)
vtx = np.asanyarray(points.get_vertices())
# print('vtx_before_reshape: ', vtx.shape) # 921600
vtx = np.reshape(vtx, (360,640, -1))
# print('vtx_after_reshape: ', vtx.shape) # (720, 1280, 1)
camera_coordinate = vtx[y][x][0]
# print ('camera_coordinate: ',camera_coordinate)
dis = camera_coordinate[2]
"""dis = aligned_depth_frame.get_distance( np.round(x).astype(np.int64), np.round(y).astype(np.int64)) # 获取该像素点对应的深度
# print ('depth: ',dis) # 深度单位是m
camera_coordinate = rs.rs2_deproject_pixel_to_point(depth_intrin, depth_pixel, dis)
# print ('camera_coordinate: ',camera_coordinate)"""
return dis, camera_coordinate
Yes, I would recommend the x and y pixel values according to the resolution rate. align_to uses the RealSense SDK's 'align processing block' to automatically adjust for differences between the depth and color streams such as different resolutions. As far as I am aware, these automatic adjustments for differences do not take place with map_to.
Can I understand by this way that it's better to keep the same resolution between RGB and depth(such as 1280*720) if I use 'map_to'. Because the resolution rate maybe cause the bias too?
Yes, if using map_to I recommend using the same resolution for both.
It's really sorry to let you waiting. I was just calculating the results. I used the same solution(1280*720) for the RGB and depth. And comment out the 'Decimation filter' and 'align_to' to keep the depth resolution. But the result is the worst until now:
Would the big bias cause by the different HOV between RGB and depth?When using the 'map to' alone didn't align the RGB and depth even they have the same resolution?
Currently, the original method(using 'align_to' and 'map to') is better. Is there any other method to improve the edge inaccuracy using depth-color alignment? Can I get the pointcloud using python as accurately as using realsense viewer?
Comparing your code to the align_depth2color.py example that the script seems to be based on, I note that you use this line:
aligned_frames = align.process(pc_filtered)
Whilst in align_depth2color_py it uses frames in the brackets instead of pc_filtered.
aligned_frames = align.process(frames)
The frames information comes from the frames = pipeline.wait_for_frames() line a little earlier in the script.
Does this mean that I can't use the filters or use the filters after align?
frames = pipeline.wait_for_frames()
aligned_frames = align.process(frames)
or
frames = pipeline.wait_for_frames()
aligned_frames = align.process(frames)
pc_filtered = decimation.process(aligned_frames)
pc_filtered = depth_to_disparity.process(pc_filtered)
pc_filtered = spatial.process(pc_filtered)
pc_filtered = temporal.process(pc_filtered)
pc_filtered = disparity_to_depth.process(pc_filtered).as_frameset()
You can still use the filters, yes.
Intel's recommendation is to place align.process after the post-processing filter list. This is a recommendation rather than a requirement though, and there are rare cases where an application has performed much better when placing align.process before the post-processing filters.
I have tried using align before filters, but it's a pity that the method didn't work.Is there other methods to improve the align accuracy or depth quality? (I'm think is there any problem with the depth quality)
How does the realsense viewer get the pointcloud? How can I get the pointcloud using python as accurately as using realsense viewer ? (can any setting files export from viewer and then import to python)
My understanding is that the RealSense Viewer pointcloud in its 3D mode is based on pc.calculate and map_to, and does not make use of align_to. RealSense Viewer is also a C++ application rather than a Python one.
You could check whether there is a mis-calibration of your camera's depth sensing by resetting it to its factory-new default calibration in the RealSense Viewer using instructions at https://github.com/IntelRealSense/librealsense/issues/10182#issuecomment-1019854487
OK, I'll calibrate the camera again. Does I just need to do the on_chip calibration, tare calibration and dynamic calibration?
Except the problem of align, how can I improve the depth quality, which kind of setting is useful to reduce the volatility of rebar lines that is stright actually?
Whilst on-chip calibration can be used to calibrate the camera, simply using the Viewer's factory-default calibration reset can work just as well.
On-chip calibration improves depth image quality, whilst tare calibration improves depth measurement accuracy.
Dynamic Calibration is a different method of calbration to on-chip that has the benefit of being able to calibrate the RGB sensor too.
The grid of rebar objects has the potential to confuse the depth sensing algorithm of the camera by forming a repetitive pattern (a series of similar looking objects in horizontal and vertical arrangements, like ceiling / floor tiles). Intel have a guide at the link below to reducing the neative impact of repetitive patterns.
https://dev.intelrealsense.com/docs/mitigate-repetitive-pattern-effect-stereo-depth-cameras
Thank you for the links, I'll have a try!
Sorry to bother again, is there any method to plot the points on the pointcloud.ply imported from map_to?(I want to check the point location on the pointcloud)
Once a ply is exported then you can import it into other tools and also pointcloud processing libraries such as PCL and Open3D but not import a .ply directly back into the RealSense SDK and access its depth information. A .bag file is the best format for reading recorded depth data back into an SDK script.
Is there any recommended samples for Open3D?
There are some Open3D examples for RealSense at the link below.
http://www.open3d.org/docs/0.12.0/tutorial/sensor/realsense.html
The official RealSense documentation for the Open3D wrapper also has some Python example code.
https://github.com/IntelRealSense/librealsense/tree/master/wrappers/open3d
Thank you very much! I'll have a try!
Hi @weicuiting Do you require further assistance with this case, please? Thanks!
Case closed due to no further comments received.
|
2025-04-01T06:37:05.166477
| 2023-06-30T14:10:24
|
1782597612
|
{
"authors": [
"MartyG-RealSense",
"jpfsaunders"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1328",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/11957"
}
|
gharchive/issue
|
hardware_reset generates "failed to set power state" in C++ application
|---------------------------------|------------------------------------------- |
| Camera Model = D455
| Firmware Version = <IP_ADDRESS>
| Operating System & Version Linux = Ubuntu 18.04
| Kernel Version (Linux Only) = ?
| Platform = nVidia Jetson Nano
| SDK Version = ?
| Language = C++ |
|
Issue Description
My C++ app operates three D455 cameras. Normally it starts and runs correctly. I would like to add a feature that monitors some info and then may decide to perform a hardware reset on a single camera (not all three), and then re-enable the steams and re-start the pipeline for that one camera (leaving the other two running as they originally were).
My attempt to do this is not working. All I did was run my code to determine if a reset is warranted (this part does work), and if it is, perform a hardware reset and then repeat the same code I used to originally (successfully) start up the camera.
When I try the hardware_reset(), the OS throws an exception "failed to set power state" and the app stops. The reset/restart code I tried using appears below:
// Below, variables g_list_of_cameras, serials, and ctx have been previously defined before this code block executes
// This next block assumes that the devices of g_list_of_cameras are in the order Camera 1, 2, 3. If not, it won't work.
// Note that each "for" loop should only do anything for 1 device and skip all others:
int cn = 0;
for (auto&& dev : g_list_of_cameras)
{
cn += 1;
if (cn != camera) continue;
cout << "resetting camera #" << camera << ", please wait ..." << endl;
dev.hardware_reset();
rs2::device_hub hub(ctx);
dev = hub.wait_for_device();
sleep(3);
}
// Re-Start streaming pipe for the reset device
cn = 0;
for (auto&& serial : serials)
{
cn += 1;
if (cn != camera) continue;
cout << serial << "\n\r";
rs2::pipeline pipe(ctx);
rs2::config cfg;
cfg.enable_device(serial);
cfg.disable_all_streams();
waitKey(10);
cfg.enable_stream(RS2_STREAM_COLOR, _FWIDTH, _FHEIGHT, RS2_FORMAT_BGR8, _CAM_FRAME_RATE);
cfg.enable_stream(RS2_STREAM_DEPTH, _FWIDTH, _FHEIGHT, RS2_FORMAT_Z16, _CAM_FRAME_RATE);
cfg.enable_stream(RS2_STREAM_INFRARED, _FWIDTH, _FHEIGHT, RS2_FORMAT_Y8, _CAM_FRAME_RATE);
pipeline_profile selection = pipe.start(cfg);
}
Should it be possible to use hardware_reset() on a camera more than once after initial powerup? What might be wrong with what I am doing?
Hi @jpfsaunders There is not a C++ example of resetting a single specific camera by its serial number, though a RealSense team member provided one for Python at https://github.com/IntelRealSense/librealsense/issues/5428#issuecomment-564482167 that takes the approach of generating a list of all attached cameras with the ctx.query_devices() instruction and then querying a serial number in that list.
There is also a C++ reset script at https://github.com/IntelRealSense/librealsense/issues/9287#issuecomment-867826974 that cycles through all attached cameras. This script also uses ctx.query_devices()
Hi @MartyG-RealSense ,
Thank you for the quick response. The Python example is exactly what I am doing in C++; for me it only works once on powerup and then if I try it a second time after that I get an error. I will look through the C++ reset script to see what they are doing but it looks more involved so it may take me a bit to figure out what is happening there.
Thanks very much, @jpfsaunders - I look forward to your next report. Good luck!
Hi @jpfsaunders Do you have an update about this case that you can provide, please? Thanks!
Hi @MartyG-RealSense ,
I have not made any progress, and this particular issue has been bumper down on my priority list. The C++ script does not appear to be doing anything different than what I am already doing. About the only thing I can think of is to try to reset the USB controller on my hardware (Jetson Nano) separately, either before or at the same time as performing the hardware_reset(). I am not sure how to do that, but I will experiment once I can get back to looking at this.
I would say we can close this issue. I will open a new one later if I can't figure out a workaround.
Okay, thanks very much @jpfsaunders for the update.
|
2025-04-01T06:37:05.171494
| 2020-06-17T12:56:26
|
640420719
|
{
"authors": [
"RealSenseCustomerSupport",
"RealSenseSupport",
"liukelinlin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1329",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/6624"
}
|
gharchive/issue
|
python application crashed when import pyrealsense2 on Win7
The device is a D415(Driver verision: <IP_ADDRESS>), and installed SDK2.0 (Intel RealSense SDK - win7 - <IP_ADDRESS>7). The viewer works as expected.
And installed python3.6.10 by miniconda, and "pip install pyrealsense2" successfully.
when i tried to "import pyrealsense2", it crashed. The error is attached.
https://support.intelrealsense.com/hc/user_images/JEkOxjDK-7M-hQ0fmg6Zng.jpeg
thanks,
@liukelinlin, could you please try to build the pyrealsense2 from source for Python3 for Win7?
https://github.com/IntelRealSense/librealsense/tree/master/wrappers/python#windows
Building the SDK2 for Win7 on either WIndows 7 or Windows 10 environment:
https://github.com/IntelRealSense/librealsense/blob/master/doc/installation_windows.md
https://github.com/IntelRealSense/librealsense/blob/master/doc/installation_win7.md#building-from-source
Thank you
Please let us know if further assistance is needed. Thank you.
The ticket will be closed in 7 days if there is no other question on the same topic.
|
2025-04-01T06:37:05.243130
| 2024-01-18T10:05:30
|
2087939904
|
{
"authors": [
"gordonhu608",
"hank-nguyen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1330",
"repo": "InternLM/InternLM-XComposer",
"url": "https://github.com/InternLM/InternLM-XComposer/issues/124"
}
|
gharchive/issue
|
Could not find a model in internlm/internlm-xcomposer-7b-4bit with a name in gptq_model-4bit-128g.safetensors, model.safetensors
I encountered this error trying to run python examples/example_chat_4bit_en.py. Thanks for any help.
FileNotFoundError: Could not find a model in internlm/internlm-xcomposer-7b-4bit with a name in gptq_model-4bit-128g.safetensors, model.safetensors. Please specify the argument model_basename to use a custom file name.
Same issue here.
check #51 works for me.
|
2025-04-01T06:37:05.245661
| 2023-09-04T03:03:51
|
1879424000
|
{
"authors": [
"QichangZheng",
"fairyshine"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1331",
"repo": "InternLM/InternLM",
"url": "https://github.com/InternLM/InternLM/issues/270"
}
|
gharchive/issue
|
[Bug] ImportError: This modeling file requires the following packages that were not found in your environment: configuration_internlm. Run pip install configuration_internlm
Describe the bug
This bug rises recently when running the code provided on huggingface, which was not in previous version.
ImportError: This modeling file requires the following packages that were not found in your environment: configuration_internlm. Run pip install configuration_internlm
Environment
python=3.8
transformers=4.31.0
Other information
No response
Unless I pass revision='' to fix a previous version. So it seems that there is something wrong with the latest version.
同样的问题
|
2025-04-01T06:37:05.374463
| 2020-08-28T14:11:41
|
688104156
|
{
"authors": [
"dannylamb"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1333",
"repo": "Islandora-Devops/ansible-role-matomo",
"url": "https://github.com/Islandora-Devops/ansible-role-matomo/pull/11"
}
|
gharchive/pull-request
|
Master -> Main Rename
Part of https://github.com/Islandora/documentation/issues/1595
@Islandora-Devops/committers bump
@rosiel :bowing_man:
|
2025-04-01T06:37:05.379240
| 2020-01-09T19:55:07
|
547687673
|
{
"authors": [
"seth-shaw-unlv",
"whikloj"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1334",
"repo": "Islandora/documentation",
"url": "https://github.com/Islandora/documentation/issues/1396"
}
|
gharchive/issue
|
Configuration page WSOD when Tomcat is down
If Tomcat is down when visiting the general configuration page '/admin/config' will WSOD and report to the log GuzzleHttp\Exception\ConnectException: cURL error 7: Failed connect to <IP_ADDRESS>:8080; Connection refused (see http://curl.haxx.se/libcurl/c/libcurl-errors.html) in GuzzleHttp\Handler\CurlFactory::createRejection() (line 185 of /var/www/html/drupal/vendor/guzzlehttp/guzzle/src/Handler/CurlFactory.php)..
To be honest, I don't know why the general configuration page crashes when Tomcat is down, but it does.
To reproduce:
Spin up a box
Visit `http://localhost:8000/admin/config' which should load fine.
vagrant ssh
sudo systemctl stop tomcat8
Visit `http://localhost:8000/admin/config' which will WSOD.
Visit the recent log messages page (http://localhost:8000/admin/reports/dblog) and see the error listed.
Resolved with https://github.com/Islandora/islandora/commit/2a1024c19a4b495c64fb88d4260e8d7ebea1f9b5
|
2025-04-01T06:37:05.382672
| 2022-02-07T17:40:33
|
1126300068
|
{
"authors": [
"DonRichards",
"rosiel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1335",
"repo": "Islandora/islandora_ci",
"url": "https://github.com/Islandora/islandora_ci/pull/9"
}
|
gharchive/pull-request
|
Update travis_setup_drupal.sh
update scripts because of bad update to phpcs.
Require 8.3.13 since 8.3.14 breaks, see https://www.drupal.org/project/coder/issues/3262291
Currently testing via this PR: https://github.com/Islandora/islandora_defaults/pull/64
@Islandora/8-x-committers
This is holding up at least 3 PR's right now:
https://github.com/Islandora/controlled_access_terms/pull/78
https://github.com/Islandora/islandora_defaults/pull/64
https://github.com/Islandora/islandora/pull/862
Using the examples I was able to determine this should work.
|
2025-04-01T06:37:05.398373
| 2024-01-24T14:03:25
|
2098338302
|
{
"authors": [
"LcsGrz",
"jena-chakour"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1336",
"repo": "Iterable/react-native-sdk",
"url": "https://github.com/Iterable/react-native-sdk/issues/530"
}
|
gharchive/issue
|
[ANDROID] - Notifications do not trigger onForegroundEvent
Notifications do not trigger onForegroundEvent (Android)
I did several tests, I do not understand how to solve it.
In my project I am using iterable, firebase messaging and notifee.
the question is that when a new notification arrives from iterable I receive it from firebase and it is shown automatically.
In android when I press that notification with the app open, nothing happens, no notifee event is executed where I can give an action in the code, the opposite happens in IOs, if the onForegroundEvent function of notifee is executed.
How can I implement an onPress in android, or how can I solve the problem?
I don't know if this issue should go here or in another of the packages I mentioned.
I am going to show the versions of the libraries and how I implemented it.
"@iterable/react-native-sdk": "1.3.17",
"@notifee/react-native": "7.8.2",
"@react-native-firebase/analytics": "18.7.3",
"@react-native-firebase/app": "18.7.3",
"@react-native-firebase/crashlytics": "18.7.3",
"@react-native-firebase/dynamic-links": "18.7.3",
"@react-native-firebase/messaging": "18.7.3",
"@react-native-firebase/perf": "18.7.3",
"@react-native-firebase/remote-config": "18.7.3"
// Firebase notification events setup
FirebaseMessaging().setBackgroundMessageHandler(handleRemoteNotification);
FirebaseMessaging().onMessage(handleRemoteNotification);
FirebaseMessaging().onTokenRefresh(updatePushNotificationToken);
FirebaseMessaging().setAutoInitEnabled(true);
// Push android notifications setup
await notifee.createChannel({
...channel,
vibration: true,
importance: AndroidImportance.HIGH,
sound: 'default',
});
// Handle onNotificationPress while the app is running
notifee.onForegroundEvent(({ type, detail }) => {
console.log('\n\n', '# foreground', type, '\n\n');
if (!detail.notification) return;
switch (type) {
case EventType.PRESS:
case EventType.ACTION_PRESS:
handlePushNotificationPressed(detail.notification);
}
});
// Handle onNotificationPress while the app is closed
notifee.onBackgroundEvent(async ({ type, detail }) => {
console.log('\n\n', '# background', type, '\n\n');
if (!detail.notification) return;
switch (type) {
case EventType.PRESS:
case EventType.ACTION_PRESS: {
handlePushNotificationPressed(detail.notification);
// Remove the notification
if (detail.notification?.id) {
await notifee.cancelNotification(detail.notification.id);
}
}
}
});
// Iterable setup
const config = new IterableConfig();
config.autoPushRegistration = true;
const initialized = await Iterable?.initialize(
env.iterableApiKey,
config,
);
Iterable?.setEmail(user);
the code looks too simple, it is worth noting that when I receive an event from firebase and the notification is iterable I do not show it a second time with notifee
EDIT 1:
to explain a little bit about the execution flow
I send a notification from iterable with the app open but in the background.
Firebase catches it and executes onMessage or the callback of setBackgroundMessageHandler
As it is a notification from iterable, I don't execute any action with Notifee, otherwise it would be duplicated.
Then if I press that notification, it opens the application but nothing happens, no Notifee event is executed.
I was testing and the difference I can find between this notification and the one I launch from notifee is the channelID, which is the firebase default, although I don't think it has any relevance fcm_fallback_notification_channel
Hi @LcsGrz, thanks for reaching out. Can you test extending your Firebase service to the Iterable service so it can forward onMessageReceived and onNewToken calls to IterableFirebaseMessagingService.handleMessageReceived and IterableFirebaseMessagingService.handleTokenRefresh, respectively?
Reference: https://support.iterable.com/hc/en-us/articles/360035019712-Iterable-s-Android-SDK#handling-firebase-push-messages-and-tokens
@jena-chakour
Hi :D , I hope you are doing well, I was testing but I am not so sure how I can debug this in android, what I did was to create the java file in the android folder and in the manifest I added the following line
<service android:name=".MyFirebaseMessagingService" />
package com.p;
import android.util.Log;
import com.google.firebase.messaging.FirebaseMessagingService;
import com.google.firebase.messaging.RemoteMessage;
import com.iterable.iterableapi.IterableFirebaseMessagingService;
import java.lang.reflect.Field;
import java.lang.reflect.Modifier;
public class MyFirebaseMessagingService extends FirebaseMessagingService {
@Override
public void onMessageReceived(RemoteMessage remoteMessage) {
Log.d( "onMessageReceived");
IterableFirebaseMessagingService.handleMessageReceived(this, remoteMessage);
}
@Override
public void onNewToken(String s) {
Log.d( "onNewToken");
IterableFirebaseMessagingService.handleTokenRefresh();
}
}
and then run npx react-native log-android
but nothing happens :(
|
2025-04-01T06:37:05.400816
| 2022-02-22T23:28:40
|
1147470594
|
{
"authors": [
"ItsDrike"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1337",
"repo": "ItsDrike/mcstatus",
"url": "https://github.com/ItsDrike/mcstatus/pull/2"
}
|
gharchive/pull-request
|
Run command for generating paylond with bash
This is just for testing to see if the workflows succeed here. It needs to be in a PR because otherwise the workflow isn't trigerred.
works, yey
|
2025-04-01T06:37:05.413897
| 2016-07-10T22:04:46
|
164736615
|
{
"authors": [
"Ivshti",
"milu2003"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1338",
"repo": "Ivshti/linvodb3",
"url": "https://github.com/Ivshti/linvodb3/issues/50"
}
|
gharchive/issue
|
Why should I use it very slow?
My data is just a few thousand。
not helpful in any way, no details
plus, performance depends largely on the back-end store you're using
@Ivshti i use it in electron.
no back-end store。
you can't possibly use it without a back-end store, as linvodb3 cannot be used without a back-end store
|
2025-04-01T06:37:05.484788
| 2020-08-18T08:02:27
|
680785601
|
{
"authors": [
"AIprogrammer",
"FabioTarocco"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1339",
"repo": "JDAI-CV/Down-to-the-Last-Detail-Virtual-Try-on-with-Detail-Carving",
"url": "https://github.com/JDAI-CV/Down-to-the-Last-Detail-Virtual-Try-on-with-Detail-Carving/issues/26"
}
|
gharchive/issue
|
Can't extract pretrained models
Hi,
We tried to extract the pretrained models from googledrive but when we are going to open them in both linux and windows, they appear to be corrupted.
Hi, you don't need to extract the checkpoint which can directly loaded by pytorch.
|
2025-04-01T06:37:05.489529
| 2024-09-22T20:42:18
|
2541249730
|
{
"authors": [
"DJOA-UP",
"Tonix22"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1340",
"repo": "JERL88/FMDAF",
"url": "https://github.com/JERL88/FMDAF/pull/6"
}
|
gharchive/pull-request
|
Task_excercise code review request
Modified code to match google standard, added more comments, modified classes and changed main.
Remove files
Then add a git ignore files for the build directory. Exclude something that is not [ cmake, cpp , h or .hpp. ]
|
2025-04-01T06:37:05.498422
| 2023-10-11T20:47:47
|
1938702412
|
{
"authors": [
"JGeek00",
"zekabra"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1341",
"repo": "JGeek00/adguard-home-manager",
"url": "https://github.com/JGeek00/adguard-home-manager/issues/62"
}
|
gharchive/issue
|
TROJ_GEN.R002V01IT23
This is probably a false positive, but since the penultimate version this has been detected in Virustotal
Obviously it's a false positive. Don't know why now it's reporting the app as a malware but it's completely false. The project is open source and everyone can read the code and search malicious behaviors.
|
2025-04-01T06:37:05.568826
| 2024-05-23T19:57:05
|
2313720122
|
{
"authors": [
"restelli"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1342",
"repo": "JQIamo/jane",
"url": "https://github.com/JQIamo/jane/issues/6"
}
|
gharchive/issue
|
Microzed freezes after some hours of operation
This is an unpredictable bug: the server on Microzed looses connection with BLACS worker after some operation time.
I suggest using sipyco, zprocess or Cap&Proto
Also protobuf could work since it is well maintained.
|
2025-04-01T06:37:05.577437
| 2020-03-06T18:22:46
|
577099016
|
{
"authors": [
"JWO719",
"tanepiper"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1343",
"repo": "JWO719/rxjs-web",
"url": "https://github.com/JWO719/rxjs-web/pull/1"
}
|
gharchive/pull-request
|
Possible improvements for library
This PR adds:
Add types for Observable Notification and Network
Add documentation to operators
Change network.connection API to use fromEventPattern
Add errors where approriate
These probably need some test coverage running in a real browser such as Karma tests.
This is amazing! thanks a lot for all the work you put into that! Are you still working on something or can I merge this?
@JWO719 if the tests pass green it's ready to merge :)
At the moment they are highly mocked. I tried investigating Jest in the browser, I can get puppeteer to work but not webdriver. Also JSDom causes similar issues.
Without figuring out a way to test them natively (and not sure geolocation is possible) this was the best solution I could come up with.
@JWO719 See #2 for more improvements on this branch. That change has a bigger change in the repo itself (using @nrwl/nx) but now also has the outline of an app that can be used to run in browser, also run E2E tests (as Jest will only ever be mockable)
I'll close this in favor of #2
|
2025-04-01T06:37:05.609074
| 2016-11-09T12:25:52
|
188235538
|
{
"authors": [
"JWally",
"Toffi-123",
"bchevalier"
],
"license": "unlicense",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1344",
"repo": "JWally/jsLPSolver",
"url": "https://github.com/JWally/jsLPSolver/issues/47"
}
|
gharchive/issue
|
Operators in Constraints
I would like to know if it was possible to include the option to add up properties of variables in the constraints section. Let's see an example: I have a scenario where a farmer wants to optimise his profits by evaluating the most profitable growing option:
{ "name": "Crop Rotation Problem", "optimize": "Gross Margin", "opType": "max", "constraints": { "field0": { "max": 1 }, "field1": { "max": 1 }, "Wheat": { "max": 10 } }, "variables": { "field00": { "field0": 1, "Wheat": 3.6, "Gross Margin": 3052.44 }, "field01": { "field0": 1, "Summer Wheat": 3.6, "Gross Margin": 1787.58 }, "field02": { "field0": 1, "Summer Barley": 3.6, "Gross Margin": 1789.65 }, "field10": { "field1": 1, "Corn": 11.9, "Gross Margin": 7243.830000000001 }, "field11": { "field1": 1, "Soybeans": 11.9, "Gross Margin": 4710.235 }, "field12": { "field1": 1, "Onions": 11.9, "Gross Margin": 46602.325000000004 }, }, "ints": { "field00": 1, "field01": 1, "field02": 1, "field10": 1, "field11": 1, "field12": 1, } }
Now for example imagine there was a policy restriction only allowing the farmer to grow a maximum of 75% of his entire acerage with the sum of Wheat and Summer Barley. To solve this, it would be nice if the Solver would be able to handle the following:
{ "name": "Crop Rotation Problem", "optimize": "Gross Margin", "opType": "max", "constraints": { "field0": { "max": 1 }, "field1": { "max": 1 }, "Wheat" + "Summer Barley": { "max": 10 } },...
Maybe there is a nicer and mathematically more appealing way to achieve this, but for now that's the only way I could think of.
Thanks already for the awesome work!
Hi Toffi,
I think that the simplest way would be to add the contribution to a constraint that corresponds to the maximum total area (maxFields here) for all the variables that have an impact on the total area (all the variables in this case):
"constraints": {
"maxFields": { "max": 10 },
"field0": { "max": 1 },
"field1": { "max": 1 },
"Wheat": { "max": 10 } }
},
"variables": {
"field00": { "maxFields": 1, "field0": 1, "Wheat": 3.6, "Gross Margin": 3052.44 }, }
"field01": { "maxFields": 1, "field0": 1, "Summer Wheat": 3.6, "Gross Margin": 1787.58 },
"field02": { "maxFields": 1, "field0": 1, "Summer Barley": 3.6, "Gross Margin": 1789.65 },
"field10": { "maxFields": 1, "field1": 1, "Corn": 11.9, "Gross Margin": 7243.83 },
"field11": { "maxFields": 1, "field1": 1, "Soybeans": 11.9, "Gross Margin": 4710.235 },
"field12": { "maxFields": 1, "field1": 1, "Onions": 11.9, "Gross Margin": 46602.325 },
}
What do you think?
Nice, that makes a lot of sense! Though I'm still struggling with another concept related to this issue: the policy may force our example farmer to grow a minimum of 2 crops on his farm. In GAMS one could create a variable CropsUsed and check whether the sum of CropsUsed was greater than 2. Now I don't see a way to get this to match the setup above, I was thinking of assigning different values for each crop, but realised that it would not get me any further. Do you probably have an idea? Thanks a lot already!
I am not sure I understand. If you want to force a minimum number of fields, you can achieve it the following way:
"constraints": {
"cropsUsed": { "min": 2, "max": 10 },
"field0": { "max": 1 },
"field1": { "max": 1 },
"Wheat": { "max": 10 } }
},
"variables": {
"field00": { "cropsUsed": 1, "field0": 1, "Wheat": 3.6, "Gross Margin": 3052.44 }, }
"field01": { "cropsUsed": 1, "field0": 1, "Summer Wheat": 3.6, "Gross Margin": 1787.58 },
"field02": { "cropsUsed": 1, "field0": 1, "Summer Barley": 3.6, "Gross Margin": 1789.65 },
"field10": { "cropsUsed": 1, "field1": 1, "Corn": 11.9, "Gross Margin": 7243.83 },
"field11": { "cropsUsed": 1, "field1": 1, "Soybeans": 11.9, "Gross Margin": 4710.235 },
"field12": { "cropsUsed": 1, "field1": 1, "Onions": 11.9, "Gross Margin": 46602.325 },
}
Notice that I just renamed maxFields into cropsUsed and added a lower bound to the constraint.
Sorry for being unclear, I think the first example is too minified. The farmer can only grow one kind of plant on each of his fields. He has multiple options though, for field0 example he has 9 options (field00 - field 08), for field1 he has only 5 options. He would like to chose the combination of options leading to the highest gross margin, which in this case would be to grow Onions on both field0 and field1. But the policy requires him to grow at least 2 crops on his entire farm (or even 3 if he had more than 2 fields), so any other combination would be acceptable, like field0 Onions and field1 Winter Wheat.
{
"name": "Crop Rotation",
"optimize": "Gross Margin",
"opType": "max",
"constraints": {
"field0": {
"max": 1
},
"feld1": {
"max": 1
},
"Potato": {
"max": 10
}
},
"variables": {
"field00": {
"field": 1,
"Sugar Beets": 11.9,
"Gross Margin": 9915.792
},
"field01": {
"field": 1,
"Greening": 11.9,
"Gross Margin": 11.9
},
"field02": {
"field": 1,
"Winter Wheat": 11.9,
"Gross Margin": 10121.07
},
"field03": {
"field": 1,
"Winter Barley": 11.9,
"Gross Margin": 8123.695000000001
},
"field04": {
"field": 1,
"Winter Rye": 11.9,
"Gross Margin": 6995.218
},
"field05": {
"field": 1,
"Summer Barley": 11.9,
"Gross Margin": 5935.2
},
"field06": {
"field": 1,
"Oats": 11.9,
"Gross Margin": 7517.308
},
"field07": {
"field": 1,
"Triticale": 11.9,
"Gross Margin": 6527.24
},
"field08": {
"field": 1,
"Onions": 11.9,
"Gross Margin": 46494.196
},
"feld10": {
"feld1": 1,
"Summer Wheat": 2.8,
"Gross Margin": 1389.98
},
"feld11": {
"feld1": 1,
"Summer Barley": 2.8,
"Gross Margin": 1391.65
},
"feld12": {
"feld1": 1,
"Winter Rye": 2.8,
"Gross Margin": 1639.4109999999998
},
"feld13": {
"feld1": 1,
"Triticale": 2.8,
"Gross Margin": 1529.98
},
"feld14": {
"feld1": 1,
"Onions": 11.9,
"Gross Margin": 46494.196
}
},
"ints": {
"field0": 1,
"field1": 1,
"field2": 1,
"field3": 1,
"field4": 1,
"field5": 1,
"field6": 1,
"field7": 1,
"field8": 1,
"feld10": 1,
"feld11": 1,
"feld12": 1,
"feld13": 1,
"feld14": 1
}
}
I hope that helps clarifying the issue, I've been thinking about it all day and it's driving me crazy. Thought about assigning prime numbers to each crop and then dividing the sum of them through each individual prime number (if the results was even, then just one crop would be grown), but a) this only works for 2 crops (not 3 or even more) and b) I still can't use operators in the constraint section.
Thanks for helping me out!
Basically you want to get at least n different types of crops. The solution would be to use the "big M" method. In your example, the idea is to force diversity of crops to 2. Let's say there are only 3 types of crops b, r and o (for barley, rye and onion), we can create 3 binary variables y_b, y_r and y_o corresponding to whether each crop is grown or not (binary variable => value either 0 or 1).
Then we can create a constraint y_b + y_r + y_o >= 2 (diversity constraint), that will force our diversity to be at least 2.
Now, we need to force those binary variables to 0 or 1 when the correspond crops are grown. 2 constraints per type of crops are needed: one to restrain the binary variable to 0 and one to force it to 1:
The restrain constraint is quite straightforward: y_o <= nb_onion_fields (can be rewritten 0 <= nb_onion_fields - y_o). This constraint ensures that if no onion field is grown, the associated binary variable will be 0.
The force constraint is a tiny bit more tricky: nb_onion_fields <= y_o * M (can be rewritten nb_onion_fields - y_o * M <= 0). This constraint ensures that if onion is grown, the associated binary variable will be 1. Notice that the value for M should be big enough so that it does not constraint the number of onion fields you can grow.
Here is the json formulation for what I stated above (by the way, the simpler the example the better).
{
"name": "Crop Rotation",
"optimize": "Gross Margin",
"opType": "max",
"constraints": {
"field0": {
"max": 6
},
"feld1": {
"max": 6
},
"cropsUsed": {
"max": 10
},
"diversity": {
"min": 2
},
"restrainOnion": {
"min": 0
},
"restrainSummerBarley": {
"min": 0
},
"restrainWinterRye": {
"min": 0
},
"forceOnion": {
"max": 0
},
"forceSummerBarley": {
"max": 0
},
"forceWinterRye": {
"max": 0
}
},
"variables": {
"growOnion": {
"diversity": 1,
"restrainOnion": -1,
"forceOnion": -999
},
"growSummerBarley": {
"diversity": 1,
"restrainSummerBarley": -1,
"forceSummerBarley": -999
},
"growWinterRye": {
"diversity": 1,
"restrainWinterRye": -1,
"forceWinterRye": -999
},
"field00": {
"field0": 1,
"cropsUsed": 1,
"Winter Rye": 11.9,
"Gross Margin": 6995.218,
"restrainWinterRye": 1,
"forceWinterRye": 1
},
"field01": {
"field0": 1,
"cropsUsed": 1,
"Summer Barley": 11.9,
"Gross Margin": 5935.2,
"restrainSummerBarley": 1,
"forceSummerBarley": 1
},
"field02": {
"field0": 1,
"cropsUsed": 1,
"Onions": 11.9,
"Gross Margin": 46494.196,
"restrainOnion": 1,
"forceOnion": 1
},
"feld10": {
"feld1": 1,
"cropsUsed": 1,
"Summer Barley": 2.8,
"Gross Margin": 1391.65,
"restrainSummerBarley": 1,
"forceSummerBarley": 1
},
"feld11": {
"feld1": 1,
"cropsUsed": 1,
"Winter Rye": 2.8,
"Gross Margin": 1639.411,
"restrainWinterRye": 1,
"forceWinterRye": 1
},
"feld12": {
"feld1": 1,
"cropsUsed": 1,
"Onions": 11.9,
"Gross Margin": 46494.196,
"restrainOnion": 1,
"forceOnion": 1
}
},
"ints": {
"field00": 1,
"field01": 1,
"field02": 1,
"feld10": 1,
"feld11": 1,
"feld12": 1,
},
"binaries": {
"growOnion": 1,
"growSummerBarley": 1,
"growWinterRye": 1
}
}
You can see that I decided to set M = 999 but M=10 would have been sufficient since no more than 10 crops can be grown.
Wow that perfectly solves my issue! Thanks so much for your advice, hope farmers will benefit from it someday 👍
no problem, let us know if you need further help
Sorry for coming back at you, but as I was implementing and checking the code I realised it wasn't solving as expected. I attached a MWE for recreating the issue, sorry for that it's so long.
When solving the attached problem, the diversity constraint (diversity > 3) is fulfilled by growPotatoes having a value of 2, and growOnions having a value of 1, despite them being declared as binary variables. It looks like my declaration of these variables as binaries is being ignored somehow. Another constraint that restricts the amount of each pair of crops grown to 95% of the entire acerage also seems to be ignored (as this constraint should also enforce the diversity of crops to a minimum of 3).
Thanks for your help once again in advance!
crop_rotation.txt
Sorry for the delay,
One possible reason why it won't solve for particular problem configurations is because the solver can be numerically unstable (and loop forever) but it's rare. It is a work in progress to make it stable but it takes a consequent amount of time. If that is the reason, you might just have to wait for the update (a reasonable ETA may be not before 3 months).
Is it important for that configuration to work? Does your work relying on the solver need to be ready soon?
Well thanks for having a look once again!
The solver is used (as you already figured I guess) to optimize crop rotations for farmers considering different aspects. It's a university project, so it will be free of charge when available. Right now, I'm hoping to get it online some time next year, also releasing it on github then. If you're interested in the project or the way the tableau is created just drop me an email, as I don't feel the code is ready to be published yet.
Without digging too deep into it, what about something like this:
You have a farm that has 4 fields on it. You can only grow 1 crop per field. Your crop choices are:
Barley (not available in field 4 because ...)
Wheat
Summer Wheat
Soy (not available for field 3 because of ...)
Onions
You can only have 1 crop / field.
{
name: "problem_2",
opType: "max",
optimize: "profit",
constraints: {
acres: {max: 100},
a_limiter: {max: 75},
f1: {max: 1},
f2: {max: 1},
f3: {max: 1},
f4: {max: 1}
},
variables: {
// Field 1 can grow these crops
f1_barley: {acres: 1, a_limiter: 1, f1: 1, profit: ?},
f1_wheat: {acres: 1, a_limiter: 1, f1: 1, profit: ?},
f1_summer_wheat: {acres: 1, f1: 1, profit: ?},
f1_soy: {acres: 1, profit: f1: 1, profit: ?},
f1_onions: {acres: 1, f1: 1, profit: ?},
// Field 2 can grow these crops
f2_barley: {acres: 1, a_limiter: 1, f2: 1, profit: ?},
f2_wheat: {acres: 1, a_limiter: 1, f2: 1, profit: ?},
f2_summer_wheat: {acres: 1, f2: 1, profit: ?},
f2_soy: {acres: 1, f1: 1, f2: 1, profit: ?},
f2_onions: {acres: 1, f2: 1, profit: ?},
// Field 3 can grow these crops...Notice Soy is Gone
f3_barley: {acres: 1, a_limiter: 1, f3: 1,profit: ?},
f3_wheat: {acres: 1, a_limiter: 1, f3: 1,profit: ?},
f3_summer_wheat: {acres: 1, f3: 1, profit: ?},
f3_onions: {acres: 1, f3: 1, profit: ?},
// Field 4 can grow these crops...Notice Barley is Gone
f4_wheat: {acres: 1, a_limiter: 1, f4: 1,profit: ?},
f4_summer_wheat: {acres: 1, f4: 1, profit: ?},
f4_soy: {acres: 1, f4: 1, profit: ?},
f4_onions: {acres: 1, f4: 1, profit: ?},
},
ints: {
f1_barley: 1,
f2_barley: 1,
...
..
.
f4_onions: 1
}
}
Thanks for the contribution! Being new to GitHub I just realized that I should close this issue, as the original question was resolved by bchevalier's answer (using the big M-Method).
The setup currently used is working similar to your approach, with some differences in the constraints to fulfill the requirements of the Common Agricultural Policy of the EU. The first issue is that each field is different in size, so restricting each crop production area to 75% of the total has to be done through the actual crop area grown. Assuming the total area would be 100, then
Wheat: 75 Soy: 75 ...
and in the variable
field1_option1: {field1: 1, Wheat: 5.6, Profit: ?} field1_option2: {field1: 1, Soy: 5.6, Profit: ?}
Then, each combination of crops cannot exceed 95% of the cropping area, therefore the following is added to the constraints
WheatBarley: 95 WheatSoy: 95 ...
and in the variables
field1_option1: {field1: 1, Wheat: 5.6, WheatBarley: 5.6, WheatSoy: 5.6, ..., Profit: ?} field1_option2: {field1: 1, Soy: 5.6, WheatSoy: 5.6, SoyOnions: 5.6, ..., Profit: ?}
As you can already see, the more growing options there are, the longer the constraints are going to be for each field.
Right now, as described above, this seems to be the source of the issues whith really long solve times for problems where this 95% constraints become binding.
To get a little closer to reality, even more constraints will be added, like available work hours during different times of the year, available machine hours and others. It would be perfect to achieve all this within JavaScript, as the available options for including GAMS are not as promising.
|
2025-04-01T06:37:05.656279
| 2024-10-27T15:15:36
|
2616666463
|
{
"authors": [
"JackPlowman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1346",
"repo": "JackPlowman/github-stats",
"url": "https://github.com/JackPlowman/github-stats/pull/211"
}
|
gharchive/pull-request
|
ci: Adjust size label thresholds for pull requests
Pull Request
Description
Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.
#211 👈
main
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @JackPlowman and the rest of your teammates on Graphite
Merge activity
Oct 27, 11:19 AM EDT: A user merged this pull request with Graphite.
|
2025-04-01T06:37:05.706181
| 2017-07-08T17:33:38
|
241468114
|
{
"authors": [
"Jackzmc"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1347",
"repo": "Jackzmc/JackzCo-Bot",
"url": "https://github.com/Jackzmc/JackzCo-Bot/issues/8"
}
|
gharchive/issue
|
custom_commands.json, command groups, and aliases
still need to work on:
custom_commands system,
command groups (secondary/main),
command aliases,
parameters
Command Group: Ready
Custom Commands: Ready
Command aliases/parameters: Not Ready
Still need to move some of my custom commnds, doing it later
|
2025-04-01T06:37:05.707694
| 2022-01-14T07:01:25
|
1103133309
|
{
"authors": [
"Jacob953",
"LegendZi"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1348",
"repo": "Jacob953/evalcsu",
"url": "https://github.com/Jacob953/evalcsu/issues/10"
}
|
gharchive/issue
|
计算机学院:请求完成通信工程2019级部分
[ ] Update more experimental reports
[ ] Update more notes
[ ] Much more professional evaluations
Add "TODO-List heading" and thanks for your future contributions!
|
2025-04-01T06:37:05.713661
| 2016-05-25T14:33:14
|
156766326
|
{
"authors": [
"Sodaaaa",
"szymczdm"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1349",
"repo": "JacquesCarette/literate-scientific-software",
"url": "https://github.com/JacquesCarette/literate-scientific-software/issues/7"
}
|
gharchive/issue
|
Bullets within a bullet
Currently unable to list bullets within a bullet.
Jingwei
This is solved in my newest version. I'll close the issue once I merge my branch.
I've merged branches, this is now implemented.
|
2025-04-01T06:37:05.750362
| 2015-03-05T09:59:27
|
59929313
|
{
"authors": [
"Jahaja",
"techtonik"
],
"license": "cc0-1.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1350",
"repo": "Jahaja/psdash",
"url": "https://github.com/Jahaja/psdash/pull/46"
}
|
gharchive/pull-request
|
Add main.py to make package executable
This is convenient for running repository version with python -m psdash
Thanks!
|
2025-04-01T06:37:05.762751
| 2016-10-17T07:15:00
|
183342082
|
{
"authors": [
"JakeWharton",
"zouzhenglu"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1351",
"repo": "JakeWharton/RxBinding",
"url": "https://github.com/JakeWharton/RxBinding/issues/296"
}
|
gharchive/issue
|
RxBinding need a new action for every click
I found I cannot use it to manager all my click code together. I alway setOnClickListenr for the same one ,and than I can create less OnClickListener object.
This is not a problem since these objects are so small. You would need to be allocating thousands of click listeners per second before it started to matter. No plans to change this.
thx. #close
thx.
|
2025-04-01T06:37:05.785284
| 2023-07-29T09:37:30
|
1827421069
|
{
"authors": [
"JaleelB",
"rodrigoricky"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1352",
"repo": "JaleelB/callsquare",
"url": "https://github.com/JaleelB/callsquare/issues/4"
}
|
gharchive/issue
|
Type '{ participants: true; }' is not assignable to type 'never'.ts(2322) (property) include: never
Type '{ participants: true; }' is not assignable to type 'never'.ts(2322)
(property) include: never
Thank you for reporting this issue, and for providing the screenshot. To help me understand and resolve the problem, could you please provide more details?
What were you doing when this issue occurred?
Can you provide any error messages or logs?
What browser and operating system are you using?
Are there specific steps I can follow to reproduce the issue?
Any additional information you can provide will be very helpful.
This happened from a fresh installation, just after I setup the database then I tried to run it, it happened.
Commands I used on setting up prisma;
yarn
yarn prisma init
yarn prisma migrate
yarn prisma migrate dev
yarn prisma db pull
npm run build && npm start
The error is likely happening because the Prisma Client that's being used to run the code doesn't match the Prisma schema. This can happen if the Prisma Client hasn't been regenerated after making changes to the schema.
Since you did a fresh clone and ran the Prisma setup commands, it's possible that the Prisma Client was not regenerated after running the migrations.
Here's what you can do to try and fix the issue:
You need to regenerate the Prisma Client to ensure that it matches the schema. You can do this by running the following command:
npx prisma generate
After regenerating the Prisma Client, you should rebuild the project to make sure that the newly generated client is being used. You can do this by running:
yarn build
Restart your development server to make sure that the changes are picked up.:
yarn dev
Also, based on the commands you listed, it seems you are using both npm and yarn at the same time, which I would recommend you not doing.
|
2025-04-01T06:37:05.794294
| 2016-12-19T22:41:40
|
196539445
|
{
"authors": [
"miguelmoraleda",
"njam3"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1353",
"repo": "Jam3/generator-jam3",
"url": "https://github.com/Jam3/generator-jam3/issues/255"
}
|
gharchive/issue
|
Add support for Google Secure Scaffold
Its being every time more common the need of work on top of that scaffold.
https://github.com/google/gae-secure-scaffold-python
It would be great to have it as an option in the generator.
Is it doable?
Definitely is doable, but I think we should wait until we have a viable angular project for the generator as there isn't really a point to setting up the google scaffold for a react project.
|
2025-04-01T06:37:05.802909
| 2019-04-29T13:59:13
|
438330424
|
{
"authors": [
"JamesMatchett"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1354",
"repo": "JamesMatchett/AWS-Discord-Bot",
"url": "https://github.com/JamesMatchett/AWS-Discord-Bot/pull/3"
}
|
gharchive/pull-request
|
Add status command to bot
Might require some further formatting and code quality fixes but making the initial PR for now if anyone wants to have a look at it.
Fixes #2
Happy enough that the feature works and no compromising information is revealed from the status command, only things revealed are:
Instance name:
Instance State:
Public IP Address: (Should not be compromising assuming server network traffic rules are correctly configured, plus bot hoster can remove this line if they wish)
Last startup time:
May be worth moving the status lines messaged out to the config.json file but this will be under another issue if I decide it's worthwhile
|
2025-04-01T06:37:05.824185
| 2023-09-28T15:53:19
|
1917829529
|
{
"authors": [
"JanEricNitschke"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1355",
"repo": "JanEricNitschke/pymend",
"url": "https://github.com/JanEricNitschke/pymend/issues/8"
}
|
gharchive/issue
|
Add more tests
For example to make sure that positional and keyword only identifiers are parsed correctly. *args and **kwargs.
Plucking some docstring_parser test holes here: https://github.com/JanEricNitschke/pymend/pull/69
|
2025-04-01T06:37:05.859447
| 2021-02-08T14:19:15
|
803607847
|
{
"authors": [
"JarlPatrick",
"kaarelkivisalu",
"marko213"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1356",
"repo": "JarlPatrick/eoa",
"url": "https://github.com/JarlPatrick/eoa/issues/23"
}
|
gharchive/issue
|
Statistics pages sorting
https://eoa.ee/?kool=true and https://eoa.ee/?hof=true are not sorted properly.
https://eoa.ee/?kool=true is not sorted at all.
https://eoa.ee/?hof=true seems to be sorted by "1. KOHTI" but secont sorting parameter is random.
What would be good criteria for sorting these tables? First places, then second places, then third places? Total of the count of the first three places?
I think overall participation might be the most relevant. For example, a student that has participated in 7 competitions but does not have any high places should be higher on the list than a person who has participated in one olympiad and achieved the third place there.
In that case, should all people who have participated in at least one contest be shown in the hall of fame? Currently, only students with at least one placement 1st-3rd are shown (source).
I prefer sorting based on participation. But I think that we have to decide how many participations are required to get there. Some statistics (participation-how many students):
25-1 24-1 23-2 22-1 21-4
20-3 19-1 18-3 17-6 16-3
15-5 14-11 13-10 12-13 11-23
10-16 9-24 8-26 7-55 6-95
5-110 4-238 3-362 2-599 1-1729
So I think we should include everyone who has got TOP3 or at least 8 participations (This is open for discussion).
Note that there are currently 450 people with at least one top 3 placement in the database (502 in total when also counting others with >= 8 participations). Should the hall of fame display be based on how many people would be shown?
|
2025-04-01T06:37:05.885992
| 2022-12-07T20:30:09
|
1482773665
|
{
"authors": [
"JasonBock"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1357",
"repo": "JasonBock/Transpire",
"url": "https://github.com/JasonBock/Transpire/issues/22"
}
|
gharchive/issue
|
Refactoring for Binary Number Literals, Constants, etc.
If this exists in code:
private const int MaximumSize = 80000000;
var maximumSize = 80000000;
Offer a refactoring to reformat it like this:
private const int MaximumSize = 80_000_000;
var maximumSize = 80_000_000;
I personally find that much easier to read.
Note that VS has a "Separate thousands" refactoring:
However, I'd like this to be solution-wide, so I'll still do this as an interested reader looking for exercise :)
|
2025-04-01T06:37:05.897334
| 2019-09-17T09:54:14
|
494521403
|
{
"authors": [
"A-Soltani",
"JasonGT"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1358",
"repo": "JasonGT/NorthwindTraders",
"url": "https://github.com/JasonGT/NorthwindTraders/issues/166"
}
|
gharchive/issue
|
Throwing Exception in validation flow
Hello, I inspect the repository and I got an issue about the validation way in the application layer.
According to following link, there is a RequestValidationBehavior to validate input data by using FluentValidation that is similar to ValidatorBehavior in eShopOnContainers. One of the issue is throwing exception in the code below.
https://github.com/JasonGT/NorthwindTraders/blob/c942193b02762561968c3981d9172e6e6fc6d274/Northwind.Application/Infrastructure/RequestValidationBehavior.cs#L33
I like ValidatorBehavior that MediatR have provided for us, but if throwing exception is used in handling validation there could be a big problem by using it. The problem is controlling the flow of business by exceptions that is wrong. I have thought about modifying this line of code to avoid throwing exception, but I couldn't come up with an idea.
Hi @A-Soltani - thanks for your feedback. I think about this approach sometimes too. Perhaps a better approach would be to use a RequestResult. This could include properties such as Succeeded (was the request successful) and Errors[] (if not successful, what were the errors).
In this way, we have a consistent approach to return results, errors or otherwise.
What do you think - do you like this approach better?
Feedback from everyone else welcome.
As an example:
public class RequestResult<T>
{
public bool Succeeded { get; set; }
public IEnumerable<RequestError> Errors { get; set; }
public T Result { get; set; }
}
public class RequestError
{
public string Code { get; set; }
public string Description { get; set; }
}
public class GetCustomerDetailQuery : IRequest<RequestResult<CustomerDetailVm>>
{
public string Id { get; set; }
}
The problem with this approach, is that the RequestError is quite generic in nature. An error type might be required to differentiate between authorisation, validation, not found and so on. Different types will be handled differently by an API client, 401, 400, 404, etc.
Thoughts?
Just reviewing eShopOnContainers, and they take the same ValidationException approach:
if (failures.Any())
{
_logger.LogWarning("Validation errors - {CommandType} - Command: {@Command} - Errors: {@ValidationErrors}", typeName, request, failures);
throw new OrderingDomainException($"Command Validation Errors for type {typeof(TRequest).Name}",
new ValidationException("Validation exception", failures));
}
The goal should be to find the simplest approach. In this case, we could state that the initial responsibility is with the client to validate the request. If the client fails to do so, then this is an exceptional circumstance - hence the ValidationException is thrown. This is therefore the simplest approach.
Thoughts?
No feedback so I am going to close this issue.
If you're interested in an alternative approach take a look at the new implementation for the UserManagerService which returns a Result type; https://github.com/JasonGT/NorthwindTraders/blob/master/Src/Infrastructure/Identity/UserManagerService.cs
|
2025-04-01T06:37:06.092887
| 2020-10-23T14:40:55
|
728274906
|
{
"authors": [
"Jean-Tinland",
"Obl1que",
"WarpWing"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1359",
"repo": "Jean-Tinland/simple-bar",
"url": "https://github.com/Jean-Tinland/simple-bar/issues/45"
}
|
gharchive/issue
|
A lot of "Something went very wrong errors"
Hello Jean :D. Good to see this repository is doing well. I'm here because I decided to mess around with my MacOS Rice a little. I decided to reinstall your version of simple-bar rather than a friend's take on it. I see that on the usage of the command
git clone https://github.com/Jean-Tinland/simple-bar $HOME/Library/Application\ Support/Übersicht/widgets/simple-bar
This freaks simple-bar out and I don't know how to solve it
Thanks in advance for any help you can provide!
Also, I can't interact with the bar like I did the first time. It should be also be worth noting that yabai isn't installed. Is it a requirement?
Actually, installing Yabai fixed the issue.
Hi, I was indeed thinking about that since the 2 widgets which weren't working were depending on yabai to get their data.
No problem, I think I need to cleanup the readme because the "Compatibility & requirements" is a bit lost in the middle of everything.
Glad it is working now!
This happened to me, but I had yabai installed. I managed to fix it by entering the right yabai path in the cmd + , settings, and upgrading/reinstalling yabai. Works beautifully now. Just wanted to say it for anyone that is having the same problem
This happened to me, but I had yabai installed. I managed to fix it by entering the right yabai path in the cmd + , settings, and upgrading/reinstalling yabai. Works beautifully now. Just wanted to say it for anyone that is having the same problem
|
2025-04-01T06:37:06.095171
| 2024-05-30T22:49:12
|
2326648784
|
{
"authors": [
"JeanExtreme002",
"yunathan51"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1360",
"repo": "JeanExtreme002/FlightRadarAPI",
"url": "https://github.com/JeanExtreme002/FlightRadarAPI/issues/75"
}
|
gharchive/issue
|
Is it possible to only retrieve information about a specific flight by searching for the flight number?
Por exemplo, apenas retornar as informações do voo LA3467 da LATAM
Unfortunately, no. You can only retrive by its airline or its registration.
|
2025-04-01T06:37:06.111814
| 2020-11-10T11:57:31
|
739847571
|
{
"authors": [
"Jeffail",
"nicktelford"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1361",
"repo": "Jeffail/benthos",
"url": "https://github.com/Jeffail/benthos/issues/547"
}
|
gharchive/issue
|
Syntax errors in bloblang expressions not checked by linter
The linter assumes that every bloblang expression is valid, this can cause invalid bloblang expressions anywhere in your configuration to prevent benthos from starting, despite the lint passing.
Expected behaviour: Linter flags all syntax errors in all bloblang expressions (including interpolated expressions).
This is done and will be in the next release: https://github.com/Jeffail/benthos/commit/cd97a28ab8d87c95fd07029309d92136d28bf5a0
|
2025-04-01T06:37:06.115179
| 2018-06-26T07:49:08
|
335698428
|
{
"authors": [
"bellwolf",
"rahulha",
"rax87"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1362",
"repo": "Jefferson-Henrique/GetOldTweets-python",
"url": "https://github.com/Jefferson-Henrique/GetOldTweets-python/issues/197"
}
|
gharchive/issue
|
Want to differentiate between retweets and "normal" tweets
Is there a way to differentiate between retweets, and normal tweets? I am scraping tweets from a set of usernames, eg. tweetCriteria = got.manager.TweetCriteria().setUsername('barackobama').
Using this criteria I get both normal tweets and retweets, however they are not differentiated.
Any advice on making adjustments to the script is welcome.
This would be very useful, and I too would like this feature.
There are several ways, try following in Python if you want to modify code and do it
Search for div with class "QuoteTweet u-block js-tweet-details-fixer". If this div exist that means it's a retweet.
Check if div with class "js-tweet-text-container" has or plain text. If it has Plain text, its not Retweet, if it has then it is
You can get the details of Retweeted data by looking into <'div class="QuoteTweet u-block js-tweet-details-fixer"><'div class="QuoteTweet-container">...
you can also search for text " Retweeted " in entire json response. If it exists, it is a Retweet.
|
2025-04-01T06:37:06.129814
| 2017-01-17T02:11:57
|
201158599
|
{
"authors": [
"JeffreyWay",
"OwenMelbz",
"jkudish",
"strebl"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1363",
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/109"
}
|
gharchive/issue
|
Super Large Sourcemap
Related to #108, but thought this would warrant it's own possible discussion thread.
As you can see in the output below the app.js.map file is 12.4mb. What would cause a sourcemap file to grow so large? Anything I'm possibly doing wrong?
> cross-env NODE_ENV=production webpack --progress --hide-modules
DONE Compiled successfully in 34886ms
Asset Size Chunks Chunk Names
/app.js 2.53 MB 1 [emitted] [big] app
fonts/icons.eot?25a32416abee198dd821b0b17a198a8f 76.5 kB [emitted]
fonts/icons.ttf?1dc35d25e61d819a9c357074014867ab 153 kB [emitted]
fonts/icons.woff?c8ddf1e5e5bf3682bc7bebf30f394148 90.4 kB [emitted]
fonts/icons.woff2?e6cf7c6ec7c2d6f670ae9d762604cb0b 71.9 kB [emitted]
/0.js 65.3 kB 0 [emitted]
fonts/icons.svg?d7c639084f684d66a1bc66855d193ed8 392 kB [emitted] [big]
/styles.css 496 kB 1 [emitted] [big] app
/0.js.map 470 kB 0 [emitted]
/app.js.map 12.4 MB 1 [emitted] app
/styles.css.map 88 bytes 1 [emitted] app
manifest.json 539 bytes [emitted]
Here are my config files
Any ideas?
Thanks again!
Your app.js is very large too. As far as I know, sourcemaps are often 5 to 10 times larger that the source file itself. But the browser doesn't download the sourcemaps, if they are not needed.
But I think 2.53 MB after minification is your bigger problem.
I've got it down to 1.1mb now after a few other optimizations, but still getting 10mb+ sourcemaps. You're saying that's to be expected?
I think you're talking about these?
new webpack.optimize.UglifyJsPlugin({
sourceMap: true,
compress: {
warnings: false,
screw_ie8: true,
conditionals: true,
unused: true,
comparisons: true,
sequences: true,
dead_code: true,
evaluate: true,
if_return: true,
join_vars: true,
},
output: {
comments: false
},
})
I was curious and tried that too. My output was exactly the same size. Only the the output.comments = false saved 1 KB on one file (from 280KB to 279KB).
I have no idea why that saved so much in your situation.
But I also don't see the mix-manifest.json in your output. Do you use the current laravel-mix version?
Here my output to compare:
> cross-env NODE_ENV=production webpack --progress --hide-modules
DONE Compiled successfully in 40789ms
Asset Size Chunks Chunk Names
/js/0.e97fe5c0195b81c24655.js.map 5.48 MB 0, 6 [emitted]
/js/0.e97fe5c0195b81c24655.js 600 kB 0, 6 [emitted] [big]
/js/2.e1d5ef9a8751c57c2226.js 5.17 kB 2, 6 [emitted]
/js/vendor.75c7220f46ad969b6a63.js 280 kB 3, 6 [emitted] [big] vendor
/js/frontend.c5d9b369644b972943d0.js 50.7 kB 4, 6 [emitted] frontend
/js/backend.2aef4bf88b748fd757aa.js 6.05 kB 5, 6 [emitted] backend
/js/manifest.d41d8cd98f00b204e980.js 1.59 kB 6 [emitted] manifest
/css/app.1842a9963041feec44c9.css 159 kB 4, 6 [emitted] frontend
/js/1.f1f62619ad496dab1015.js 13.9 kB 1, 6 [emitted]
/js/1.f1f62619ad496dab1015.js.map 73.8 kB 1, 6 [emitted]
/js/2.e1d5ef9a8751c57c2226.js.map 37.5 kB 2, 6 [emitted]
/js/vendor.75c7220f46ad969b6a63.js.map 2.05 MB 3, 6 [emitted] vendor
/js/frontend.c5d9b369644b972943d0.js.map 337 kB 4, 6 [emitted] frontend
/css/app.1842a9963041feec44c9.css.map 110 bytes 4, 6 [emitted] frontend
/js/backend.2aef4bf88b748fd757aa.js.map 41.7 kB 5, 6 [emitted] backend
/js/manifest.d41d8cd98f00b204e980.js.map 14.4 kB 6 [emitted] manifest
mix-manifest.json 628 bytes [emitted]
As you can see my souremaps are also ~10 times as large as my source files are. I think thats normal. The same is true if I look out the output from vue.js project which was generated with the vue-cli.
Did you look at your app.js without your UglifyJsPlugin modifications? Was is minified etc?
Does your app.js contain the sourcemap?
Some configurations inline the sourcemaps to your .js file - which creates a massive .js file
Yeah the sourcemaps will be large for an app that big. It shouldn't affect your production deploy.
|
2025-04-01T06:37:06.133006
| 2018-08-09T18:12:12
|
349234590
|
{
"authors": [
"ankurk91",
"raniesantos"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1364",
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/1735"
}
|
gharchive/issue
|
Just a question. What is the reasoning behind this vue loader config?
Why does laravel-mix set esModule to false?
vue: {
preLoaders: {},
postLoaders: {},
esModule: false
},
I'm asking because setups provided by vue-cli have this set to true. I tend to use require().default in specific situations, and using it becomes inconsistent for laravel-mix because this config makes me drop .default.
esModule: false option is no longer available in vue-loader v14+, laravel mix is using v13.
It has been set to false to keep mix behaviour non breaking
Of course
|
2025-04-01T06:37:06.134846
| 2017-01-31T10:33:41
|
204256969
|
{
"authors": [
"adriaanzon",
"vedmant"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:1365",
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/247"
}
|
gharchive/issue
|
Doesn't compile both less and sass files
I have following configuration:
mix.js('resources/assets/js/app.js', 'public/js').extract(['vue'])
.less('resources/assets/less/theme.less', 'public/css')
.sass('resources/assets/sass/app.scss', 'public/css');
However less files are not compiled, it works fine if I comment out .sass line.
Support for this was added just recently. Update your laravel-mix dependency to 0.6.0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.