added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:38:03.455693
| 2020-08-13T22:41:28
|
678775416
|
{
"authors": [
"dhakim87",
"wasade"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4228",
"repo": "biocore/microsetta-admin",
"url": "https://github.com/biocore/microsetta-admin/pull/44"
}
|
gharchive/pull-request
|
Unit tests for javascript
Requires qunit from npm/node in order to run tests from command line. Installs a node_modules folder containing qunit based on the package-lock.json in the microsetta_admin/tests/js folder.
First build should fail with the new unit tests - hopefully
Seems reasonable!
See travis build in second commit (f639ef3) for example of a js test failure.
To run locally, you'll need to install node/npm, this will then be used to install qunit and all dependencies. You can then run make test, or if you just want the javascript tests, run_js_tests.sh at the root of the repo.
Note that there is a compromise made to enable testing by command line: any testable functions must be retrievable through node. This means declaring them in the node defined module.exports field at the bottom of js files. Since the browser has no concept of module.exports, you must check for existence before setting this field. For a simple example of this, see microsetta_admin/static/js/testable.js.
The exact mechanism we use for setting module.exports is up for debate - if we want to enable node-like behavior in our browser imports, there is a slightly different pattern used in Emperor that makes use of requirejs. The pattern used in testable.js should work so long as all our javascript is expected to be hosted in the browser.
In the future, if you want to update qunit, cd to /microsetta_admin/tests/js/ and run npm install qunit, then commit the package-lock.json and travis will automatically use your new configuration.
Thanks, @dhakim87! Would it be possible to put the text about running and updating into the repository directly (e.g., in the readme or comments in the makefile)?
|
2025-04-01T06:38:03.460996
| 2024-01-26T10:25:05
|
2101963536
|
{
"authors": [
"haesleinhuepf",
"oeway"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4229",
"repo": "bioimage-io/bioengine",
"url": "https://github.com/bioimage-io/bioengine/issues/2"
}
|
gharchive/issue
|
Add license
Hi @oeway ,
great training resource here and great that you were publishing it openly! I was just wondering under which conditions one could reuse the code ansd materials provided here. Would you mind adding a license file? If you're new to licensing and/or wonder which license to use, you can read more in this blob post: https://focalplane.biologists.com/2023/05/06/if-you-license-it-itll-be-harder-to-steal-it-why-we-should-license-our-work/
Thanks!
Best,
Robert
Hi thanks for the heads up, our good old MIT should do!
https://github.com/bioimage-io/bioengine/commit/620440be21799457f7f73395bd9ccf6e8ca7bbea
|
2025-04-01T06:38:03.505091
| 2024-11-12T05:42:33
|
2651106983
|
{
"authors": [
"HShokaku"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4230",
"repo": "biomed-AI/DiffDec",
"url": "https://github.com/biomed-AI/DiffDec/issues/7"
}
|
gharchive/issue
|
question about multi R-group data preprocessing
Hi,
I'm trying to reproduce the data preprocess results.
I followed the tutorial in lib-invent repo to process multi R-group data of crossdock. However, I could only get 17,586 data entries which is much less than the results in you article (about 150k).
My configs for slicing are as followed:
filter_conditions.json
{
"scaffold": [{
"name":"ring_count",
"min": 1
}],
"decoration": [
{
"name":"molecular_weight",
"max": 300
},
{
"name":"hydrogen_bond_acceptors",
"max": 3
},
{
"name":"hydrogen_bond_donors",
"max": 3
},
{
"name":"clogp",
"max": 3
},
{
"name":"rotatable_bonds",
"max": 3
},
{
"name": "heavy_atom_count",
"max": 10,
"min": 1
}
]
}
reaction_based_slicing.json
{
"run_type": "reaction_based_slicing",
"parameters": {
"input_file": "path/to/unsliced/data.smi",
"output_path": "path/to/output/folder",
"output_smiles_file": "path/to/output/file.smi",
"conditions_file": "configs/filter_conditions.json",
"reactions_file": "configs/reaction.smirks",
"max_cuts": 4,
"number_of_partitions": 1000,
"validate_randomization": true
}
}
Would you like to share more details about how to get the 150k multi R-group data of crossdock?
Thanks
I found that reaction based slicing will automatically deduplicate the output data. There are many same molecule in Crossdock train set. However, one molecule may correspond to multiple proteins. Run python -W ignore process_and_prepare.py will finally get the 150k results.
|
2025-04-01T06:38:03.509417
| 2023-12-10T14:50:55
|
2034402491
|
{
"authors": [
"Conaclos",
"eryue0220"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4231",
"repo": "biomejs/biome",
"url": "https://github.com/biomejs/biome/pull/1137"
}
|
gharchive/pull-request
|
feat(style/useConsistentArrayType): add rule
Summary
Implement lint/style/useConsistentArrayType
Fixed: #68
Test Plan
All existing tests has passed.
Thanks for your contribution! This is greatly(大大) appreciated :)
I left suggestions. Feel free to ask details if it is not clear.
Many thanks for your patience reviewing and suggestions.
Sorry for the late reply.
@eryue0220 Have you still some time for this PR?
Yes, I'm still working on this PR.
@Conaclos Sorry for the late response before. And huge thanks to your suggestions and reviews. Merry Christmas.
I think it is ready for merging.
Please run just ready to format/lint the code and generate missing files.
Once CI is passing, we will be able to merge :)
Again. Huge Thanks to @Conaclos for your patience and your suggestion that I can ship this. It's a really wonderful travel.
|
2025-04-01T06:38:03.513798
| 2024-11-10T11:13:18
|
2647124713
|
{
"authors": [
"alex-hh"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4232",
"repo": "bioml-tools/bio-datasets",
"url": "https://github.com/bioml-tools/bio-datasets/issues/63"
}
|
gharchive/issue
|
write a load_example method
assuming a dataset has an id field and an index.
index will be a parquet file with no extension mapping id to shard - then we can download a single shard and retrieve the example
What we need:
a split generator that looks for split-specific index files (train_index or train/index)
index files allow us to subset both parquets and examples
we then add a ds.filter before returning the dataset.
there might be an efficient arrow way to implement the filter
(this could also go directly into yaml but the index file solution is more modular).
|
2025-04-01T06:38:03.522315
| 2021-10-08T16:25:53
|
1021276835
|
{
"authors": [
"jonrkarr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4233",
"repo": "biosimulations/biosimulations",
"url": "https://github.com/biosimulations/biosimulations/pull/3182"
}
|
gharchive/pull-request
|
refactor(platform): added tool to publish example
Indicated in which examples should be published and what ids to use in tools/example-projects.json
This avoids polluting our published projects with many versions of the same project with different simulation tools
Added command-line program to update these published projects
Presently requires a machine-to-machine api client and secret.
I started with machine-to-machine because its easier to setup. I could append this to the GitHub action I created to check the examples and automatically create/update their publication.
|
2025-04-01T06:38:03.526794
| 2016-03-30T22:51:57
|
144742312
|
{
"authors": [
"gregcaporaso"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4234",
"repo": "biota/sourcetracker2",
"url": "https://github.com/biota/sourcetracker2/issues/23"
}
|
gharchive/issue
|
Display verbose progress
From @lkursell on February 12, 2016 22:39
Displaying ST2's progress, like the original, is very helpful for making sure commands are getting executed correctly, especially given the time required for jobs to finish.
Copied from original issue: biota/sourcetracker2_internal#14
From @lkursell on February 15, 2016 16:24
The thing I counted on most from the verbose ST was to make sure that my sinks and sources were properly defined, and were running as I intended. But it also helped me gauge speed to know if some rarefaction level was just never going to work out.
On Feb 15, 2016, at 12:15 AM, Will Van Treuren<EMAIL_ADDRESS>wrote:
Will check in to using click for this. Should be straightforward given that the number of samples is known and each iteration of Gibbs is independent.
For methods that are not Gibb's this might be harder, but we will investigate those as we come to it.
—
Reply to this email directly or view it on GitHub https://github.com/biota/sourcetracker2/issues/14#issuecomment-184107722.
From @wdwvt1 on February 15, 2016 8:15
Will check in to using click for this. Should be straightforward given that the number of samples is known and each iteration of Gibbs is independent.
For methods that are not Gibb's this might be harder, but we will investigate those as we come to it.
|
2025-04-01T06:38:03.571179
| 2021-01-02T20:28:49
|
777520353
|
{
"authors": [
"cara-wong",
"crugwiro",
"wynnset"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4235",
"repo": "bipolarbridges/companion-app",
"url": "https://github.com/bipolarbridges/companion-app/pull/17"
}
|
gharchive/pull-request
|
Google Fit API client. authentication
Followed the instructions here https://github.com/StasDoskalenko/react-native-google-fit/blob/HEAD/docs/INSTALLATION.md#getting-started
Did not pass new GoogleFitPackage(BuildConfig.APPLICATION_ID) in MainApplication.java, auto link takes care if this
Add - to android Manifest.xml
https://developers.google.com/fit/android/get-started - followed this to set up OAuth client on the google console. ( having issues with this because the OAuth client with the SHA1 and package name already exists, not sure how to add the fitness api to an existing client id)
Once you pull the changes (googlefit.tsx, TipCard.tsx), after log in on android, you will be prompted to sign in to a google account.
ERROR
Does not resolve, getting stuck here
Followed the instructions here https://github.com/StasDoskalenko/react-native-google-fit/blob/HEAD/docs/INSTALLATION.md#getting-started
Do not pass new GoogleFitPackage(BuildConfig.APPLICATION_ID) in MainApplication.java, auto link takes care if this
Add <uses-permission android:name="android.permission.ACTIVITY_RECOGNITION"/> to android Manifest.xml
On the GCP, go to OAuth consent screen, and add a test user for the authentication to work
Copy useGoogleFit and Tipcard files to your local
Followed the instructions here https://github.com/StasDoskalenko/react-native-google-fit/blob/HEAD/docs/INSTALLATION.md#getting-started
Do not pass new GoogleFitPackage(BuildConfig.APPLICATION_ID) in MainApplication.java, auto link takes care if this
Add <uses-permission android:name="android.permission.ACTIVITY_RECOGNITION"/> to android Manifest.xml
On the GCP, go to OAuth consent screen, and add a test user for the authentication to work
Copy useGoogleFit and Tipcard files to your local
@crugwiro Can you clarify the testing details for me? Here's what I've done so far:
Pulled down this branch and followed step 3 in your comment above as we talked about (the rest of the steps are already done)
Launched android (after syncing with gradle)
Tried logging in with email/password. After logging in, it prompted me to sign into my google account. After selecting the email I wanted to login with, it got stuck on the loading pop up
Rebuilt the app and tried logging in with google sign, same thing happened - got stuck on a loading pop up.
Am I missing something? (screenshot below for reference) Also still in the process of reviewing the changes file by file, so maybe I will find something from that.
@crugwiro Can you clarify the testing details for me? Here's what I've done so far:
Pulled down this branch and followed step 3 in your comment above as we talked about (the rest of the steps are already done)
Launched android (after syncing with gradle)
Tried logging in with email/password. After logging in, it prompted me to sign into my google account. After selecting the email I wanted to login with, it got stuck on the loading pop up
Rebuilt the app and tried logging in with google sign, same thing happened - got stuck on a loading pop up.
Am I missing something? (screenshot below for reference) Also still in the process of reviewing the changes file by file, so maybe I will find something from that.
This branch has too many changes and merges across different PRs. It is safe to Close/delete.
This branch has too many changes and merges across different PRs. It is safe to Close/delete.
Closing as per @crugwiro
|
2025-04-01T06:38:03.598769
| 2018-08-23T22:20:07
|
353574712
|
{
"authors": [
"birik",
"nickrenfo2",
"sherwinchu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4236",
"repo": "birik/react-week-calendar",
"url": "https://github.com/birik/react-week-calendar/pull/2"
}
|
gharchive/pull-request
|
Fix duplicate key error when creating >10 back-to-back events
Solved an issue when creating >10 events on several days. Now there shouldn't be an issue even when creating one event for each minute for the entire week.
@birik Found this defect as well. It would be great if @nickrenfo2 's fix could be merged to master.
@nickrenfo2 @sherwinchu Change was merged. NPM will be updated soon
The updates in 0.1.2.
@nickrenfo2 Thank you for your commits
|
2025-04-01T06:38:03.610252
| 2022-04-04T20:27:02
|
1192277876
|
{
"authors": [
"clementd-fretlink",
"divarvel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4237",
"repo": "biscuit-auth/biscuit",
"url": "https://github.com/biscuit-auth/biscuit/pull/103"
}
|
gharchive/pull-request
|
Spec for third-party blocks
Based on the haskell PoC implementation, here's a cleaned up spec of 3rd party blocks.
Some implementations details are open to discussion, i'll add comments highlighting them
A commit-by-commit review is advised.
Both biscuit-haskell and biscuit-rust have candidate releases with third-party blocks support
|
2025-04-01T06:38:03.641023
| 2017-12-19T18:22:39
|
283325337
|
{
"authors": [
"morawskim",
"tristanlins"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4238",
"repo": "bit3/faker-cli",
"url": "https://github.com/bit3/faker-cli/pull/9"
}
|
gharchive/pull-request
|
Add bash completion
Faker offer many "formatters" (name, email). I create simple bash completion script for them.
For example:
faker.php [tab][tab]
#will display all available formatters (with some other options) for default locale
faker.php --locale pl_PL [tab][tab]
#will display all available formatters for pl_PL locale
I am sorry, but I will not continue this project. The project is now archived on github and abandoned on packagist.
|
2025-04-01T06:38:03.713063
| 2020-08-19T07:02:47
|
681600323
|
{
"authors": [
"JBaczuk",
"gy0801151351"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4239",
"repo": "bitchan/eccrypto",
"url": "https://github.com/bitchan/eccrypto/issues/69"
}
|
gharchive/issue
|
Is it compatible with bouncycastle?
I am searching an ECIES algorithm which can work fine both in client and server. My server is using Java, and the algorithm provider is bouncycastle. Is there any sample encrypting by javascript and decrypting by java?
If it supports ECIES it should be compatible.
|
2025-04-01T06:38:03.889453
| 2022-10-25T11:32:16
|
1422315065
|
{
"authors": [
"davidjoshuaford",
"ettnoll"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4240",
"repo": "bitfocus/companion-module-bmd-atem",
"url": "https://github.com/bitfocus/companion-module-bmd-atem/issues/214"
}
|
gharchive/issue
|
[Feature request] Preview button presets with One ME program source feedback
For when you don't want or have space for a preview and a program row it's handy to see what source is live on the preview button. Right now you have to add the feedback manually to each button. Would be perfect if there were an option to have preset buttons with both "One ME preview source" and "One ME program source".
Are you looking for these label variables?
$(atem:pvw1_input)
$(atem:pgm1_input)
Check the Variables tab for a list of all available shortcodes.
|
2025-04-01T06:38:03.898873
| 2022-02-14T16:35:06
|
1137578714
|
{
"authors": [
"AHub88",
"Julusian",
"bevanjkay",
"nick-potts"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4241",
"repo": "bitfocus/companion-module-spotify-remote",
"url": "https://github.com/bitfocus/companion-module-spotify-remote/issues/31"
}
|
gharchive/issue
|
[BUG] Spotify - Some Press actions are not working
Is this a bug in companion itself or a module?
[X] I believe this to be a bug in companion
Is there an existing issue for this?
[X] I have searched the existing issues
Describe the bug
The following press actions are not working:
Play
Volume Up
Volume down
Steps To Reproduce
No response
Expected Behavior
No response
Environment (please complete the following information)
- OS:
- Browser:
- Companion Version:
Additional context
No response
Meanwhile, it is also no longer possible to load albums.
Can you try with the latest beta, with a fresh api keys? There was a problem with reaching the api rate limit that have recently been resolved.
What version of companion or the module did loading albums last work for you in?
Same for the other actions? Some idea of when it could have broken would be useful, otherwise it is hard to figure out where to start looking
Acutally I'm using the latest beta -> 2.2.0 (2.2.0+4125-beta-5d1d7f80)
Unfortunately, I can not say exactly when the albums worked. It must have been 1.5 months ago.
The other Keys have never worked.
I have new infos...
The key works with an iPad as audio device.
But not with the browser solution https://open.spotify.com/
Also happening for me, volume not working.
@nick-potts Can you please provide some more information about your environment? What device are you using as the host device for Spotify?
What device are you using as the host device for Spotify?
Windows in my case
|
2025-04-01T06:38:03.902939
| 2020-04-15T20:55:55
|
600582416
|
{
"authors": [
"nodeful",
"northamerican"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4242",
"repo": "bitgapp/eqMac",
"url": "https://github.com/bitgapp/eqMac/issues/212"
}
|
gharchive/issue
|
Audio delay/latency issue
With eqMac enabled, an audio delay forms and gets longer and more noticeable the longer it runs.
After about an hour it becomes unmistakable. If I leave eqMac enabled overnight, it's a several second delay.
This happens with eqMac enabled and an HDMI output source on Catalina.
The same issue would happen with SoundFlower which I had tried using years back to have keyboard audio control of HDMI devices (my TV).
Switching the source to Internal Speakers, then back to eqMac resolves the delay temporarily.
possibly related:
https://arstechnica.com/civis/viewtopic.php?f=19&t=1289815
https://github.com/SakuraG/soundflower/issues/43
I'm moving the discussion around this issue to #225
|
2025-04-01T06:38:03.904882
| 2024-10-13T13:28:28
|
2583968433
|
{
"authors": [
"ben-kaufman",
"cakesoft-swati"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4243",
"repo": "bithyve/bitcoin-keeper",
"url": "https://github.com/bithyve/bitcoin-keeper/issues/5323"
}
|
gharchive/issue
|
Register Signer should allow more sharing options
The register signer screen when clicking on Vault details when signing with an external key only shows QR code, it should have the NFC, File, and Remote link options as well
Verified this issue on dev app v 1.2.18(410)
|
2025-04-01T06:38:03.906027
| 2021-05-05T08:36:49
|
876208884
|
{
"authors": [
"cakesoft-devika",
"cakesoft-nikhita"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4244",
"repo": "bithyve/hexa",
"url": "https://github.com/bithyve/hexa/issues/3288"
}
|
gharchive/issue
|
CloudHistory Page when we click on Backup it will increase current level
CloudHistory Page when we click on Backup it will increase the current level so when user goes to Restore flow having level1 so it will show level 2 on restore and flow breaks
-Verified this issue on Staging app v1.6.5(281)
|
2025-04-01T06:38:03.915762
| 2017-10-04T08:35:44
|
262707026
|
{
"authors": [
"codecov-io",
"josh--newman",
"voxpelli"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4245",
"repo": "bitinn/node-fetch",
"url": "https://github.com/bitinn/node-fetch/pull/350"
}
|
gharchive/pull-request
|
Add an "additionalRequestProperties" option to requests
This is to allow eg. AWS XRay instrumentation, which requires the addition of an "XRaySegment" value to be sent to the http.request() call.
The reasoning for adding additionalRequestProperties is to preserve backwards compatibility – just assigning all unknown properties from init could result in unwanted values being forwarded to http.request().
I am aware that this module wants to mimik the browser fetch() and this property doesn't align with that, but it do align with properties like agent and is useful when eg. there's a need to better trace calls across microservices in a server environment.
Codecov Report
Merging #350 into master will decrease coverage by 0.47%.
The diff coverage is 0%.
@@ Coverage Diff @@
## master #350 +/- ##
==========================================
- Coverage 100% 99.52% -0.48%
==========================================
Files 6 6
Lines 423 425 +2
Branches 133 134 +1
==========================================
Hits 423 423
- Misses 0 1 +1
- Partials 0 1 +1
Impacted Files
Coverage Δ
src/request.js
97.18% <0%> (-2.82%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update d1a3b1e...18704dd. Read the comment docs.
@voxpelli, Did you manage to get X-Ray to play nicely with node-fetch? I tried using the captureHTTPsGlobal function with no luck.
@josh--newman We're running https://github.com/Sydsvenskan/node-fetch/tree/1.x-fork internally to achieve this behavior and will probably move to another module eventually, that better suits our node.js needs. This module is more geared towards modules that should work both in in node.js and the browser (no x-ray there)
@voxpelli thanks for your response! I'm considering reevaluating our use of node-fetch as well.
|
2025-04-01T06:38:03.926204
| 2023-08-02T14:04:54
|
1833258912
|
{
"authors": [
"jonathanm-tkf",
"rolandkakonyi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4246",
"repo": "bitmovin/bitmovin-player-react-native",
"url": "https://github.com/bitmovin/bitmovin-player-react-native/issues/176"
}
|
gharchive/issue
|
We need a couple of values to be added to the analytics helper
Is there an existing issue for this?
[X] I have searched the existing issues
Current Behavior
Currently, we have the following methods from the documentation
https://developer.bitmovin.com/playback/docs/enabling-bitmovin-analytics
// Update the current custom data config.
player.analyticsCollector?.setCustomDataOnce({
customData2: 'Updated custom data field 2',
customData4: 'Updated custom data field 4',
});
we need to add new methods such as
Expected Behavior
player.analyticsCollector.setVideoId('123');
player.analyticsCollector.setTitle('Title of the Video');
player.analyticsCollector.setCdnProvider('Cdn Provider');
Steps To Reproduce
No response
What platform(s) are you experiencing the issue on?
[X] Android
[ ] Android TV / Fire TV
[X] iOS
[ ] tvOS
Player React Native SDK version
0.8.0
Device / Environment
No response
Stream URL (Optional)
No response
Additional information / Code snippets / Screenshots
No response
Hi @jonathanm-tkf!
AnalyticsConfig already offers options to assign those values on player creation.
Could this be used for your use case?
Hi @rolandkakonyi unfortunately it doesn't work for me, as we have the Next Episode functionality, which we don't destroy the player.
I'll comment in more detail:
Currently, we have the player, which has a Next Episode button, once pressed, loads the next video, for performance reasons and to not delete the current player and recreate it we use the load implementation as follows:
player.load({
url: videoData.url,
type: Platform.OS === 'ios' ? SourceType.HLS : SourceType.DASH,
title: videoData.title,
poster: videoData.posterUrl,
...
...
..
after this we need to do the following
if (player.analyticsCollector) {
player.analyticsCollector.setCustomDataOnce({
customData2: videoData.customId,
customData3: videoData.supplier,
});
player.analyticsCollector.setVideoId(videoData.id);
player.analyticsCollector.setTitle(getTitleAnalytics(videoData.dataAnalytics));
player.analyticsCollector.setCdnProvider(videoData.supplier);
}
Do you understand what the idea is? any recommendations? or do we need to destroy the player every time we want to modify this data?
Thanks for your help
Hi @jonathanm-tkf!
I see, you are correct that this is not possible right now without creating a new player instance.
You could implement this in AnalyticsModule yourself or this would be a feature request.
We have further plans with our analytics integration, we can get back to you during next week.
Thanks for the effort unfortunately I have too little time to contribute apologies, it is something we have today and I would not like to change the way of implementation, I look forward to the update, thanks again. If it is not implemented and I can help when I am free I will gladly do so.
Regards.
@jonathanm-tkf we are already working on a solution for this, you can watch #184 for the details.
I will post it here once it is released.
Hi @jonathanm-tkf, we just released v0.9.0 with support for the above use cases.
Now you can call the following API to update source-specific values for the analytics collector:
player.analyticsCollector.addSourceMetadata({
videoId: 'new video ID',
title: 'new video title',
path: 'new path',
cdnProvider: 'new CDN provider' // your new CDN provider,
});
Please see SourceMetadata for all options.
|
2025-04-01T06:38:03.927736
| 2024-02-09T15:27:44
|
2127365235
|
{
"authors": [
"bitPlayerGHActions",
"rolandkakonyi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4247",
"repo": "bitmovin/bitmovin-player-react-native",
"url": "https://github.com/bitmovin/bitmovin-player-react-native/pull/395"
}
|
gharchive/pull-request
|
Update iOS player to 3.56.0
Automated iOS player version update to 3.56.0
Closing in favor of #398
|
2025-04-01T06:38:03.988019
| 2020-10-19T11:31:55
|
724526995
|
{
"authors": [
"juan131",
"koyan"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4248",
"repo": "bitnami/tutorials",
"url": "https://github.com/bitnami/tutorials/pull/26"
}
|
gharchive/pull-request
|
Fix error in nginx configuration
When trying to deploy the helm chart phpfpm
It was failing with:
nginx 11:21:11.85 ERROR ==> Custom server blocks files were found inside '/bitnami/nginx/conf/vhosts'. This configuration is not supported anymore. Please mount your custom server blocks config files at '/opt/bitnami/nginx/conf/server_blocks' instead.
The fix is already applied (and taken from): https://github.com/bitnami/bitnami-docker-php-fpm/pull/124/files
@juan131 I was too hasty to submit it.
While it fixes the original problem, and the pod does not get stuck in a crashLoop, the whole application does not deliver anymore the php. My knowlege of nginx configuration is not good, so I don't know else should be fixed so that the nginx points to the right place.
(when you open the app on the given ip on the browser, it points to the default nginx index.html. If you try to point it to the correct php file, it returns a "not found")
@koyan please the changes I did at https://github.com/bitnami/tutorials/pull/28
That should fix the issue, thanks for reporting it!
|
2025-04-01T06:38:04.005715
| 2020-03-30T00:49:42
|
589925717
|
{
"authors": [
"etang01",
"ismaildude"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4249",
"repo": "bitprj/curriculum",
"url": "https://github.com/bitprj/curriculum/pull/964"
}
|
gharchive/pull-request
|
Ethan twitter new
Issues Closed
Please make one new line for each issue, otherwise not all issues will be accounted for!
Closes #944
Closes #942
...
Changes proposed in this pull request:
*Explanation on rate limiting
*New diagram for rate limiting
*New visual for data analysis
*Simple explanations of the libraries used
@reviewer/kavuong
Please also add your local images to GitHub (as right now you have added the file paths of your local computer). It should go something like this: commit your images locally(add them to your local version of the activity folder) and push to origin on your GitHub desktop. Other than that: Good job! I'll start merging after you are done with that. Also the new path should be "./image_name" and the image should be where the cards that use it are.
|
2025-04-01T06:38:04.010838
| 2023-08-03T14:08:10
|
1835142186
|
{
"authors": [
"lpusok"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4250",
"repo": "bitrise-steplib/steps-xcode-test",
"url": "https://github.com/bitrise-steplib/steps-xcode-test/pull/233"
}
|
gharchive/pull-request
|
Fix fallback to xcodebuild when xcpretty/xcbeautify is unavailable
Checklist
[x] I've read and followed the Contribution Guidelines
[x] step.yml and README.md is updated with the changes (if needed)
Version
Requires a PATCH version update
Context
Before:
Checking log formatter (xcbeautify) version
Failed to install Step dependencies:
installing log formatter failed:
failed to run xcbeautify command:
executing command failed (xcbeautify "--version"):
exec: "xcbeautify":
executable file not found in $PATH
Switching to xcodebuild for output tool
Running the tests...
[16:58:43] $ set -o pipefail && xcodebuild "-workspace" "/Users/lpusok/Develop/go/src/github.com/bitrise-steplib/steps-xcode-test/_tmp/BullsEye.xcworkspace" "-scheme" "BullsEye" "test" "-destination" "id=56F53136-D68E-4D5A-83B3-C35BDC7D9AFD" "-testPlan" "UITests" "-resultBundlePath" "/var/folders/r5/gkvczn3j2tb0m79nwby9fjv80000gq/T/XCUITestOutput3261492831/Test-BullsEye.xcresult" "-xcconfig" "/var/folders/r5/gkvczn3j2tb0m79nwby9fjv80000gq/T/2103298897/temp.xcconfig" | xcbeautify
xcbeautify command failed: executing command failed (xcbeautify): exec: not started
Exit code: -1
After:
Checking log formatter (xcbeautify) version
Checking log formatter failed: failed to run xcbeautify command: executing command failed (xcbeautify "--version"): exec: "xcbeautify": executable file not found in $PATH
Falling back to xcodebuild log formatter
Running the tests...
[16:57:02] $ xcodebuild "-workspace" "/Users/lpusok/Develop/go/src/github.com/bitrise-steplib/steps-xcode-test/_tmp/BullsEye.xcworkspace" "-scheme" "BullsEye" "test" "-destination" "id=56F53136-D68E-4D5A-83B3-C35BDC7D9AFD" "-testPlan" "UITests" "-resultBundlePath" "/var/folders/r5/gkvczn3j2tb0m79nwby9fjv80000gq/T/XCUITestOutput3631683970/Test-BullsEye.xcresult" "-xcconfig" "/var/folders/r5/gkvczn3j2tb0m79nwby9fjv80000gq/T/1898852877/temp.xcconfig"
Resolves: https://bitrise.atlassian.net/browse/BE-880
Changes
Investigation details
Decisions
Fix fallback to xcodebuild when xcpretty/xcbeautify is unavailable.
|
2025-04-01T06:38:04.018722
| 2021-10-18T23:23:39
|
1029693225
|
{
"authors": [
"bitsandbytesdev"
],
"license": "BSD-4-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4251",
"repo": "bitsandbytesdev/bitsandbytesdev.github.io",
"url": "https://github.com/bitsandbytesdev/bitsandbytesdev.github.io/issues/1"
}
|
gharchive/issue
|
Add scroll down arrow on desktop view
Add a scroll down arrow on the desktop view
not only for desktop, but for all devices...
|
2025-04-01T06:38:04.066973
| 2016-08-03T18:36:08
|
169202066
|
{
"authors": [
"theharq"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4252",
"repo": "bitzesty/trade-tariff-frontend",
"url": "https://github.com/bitzesty/trade-tariff-frontend/pull/32"
}
|
gharchive/pull-request
|
Add attribute rel="nofollow" to changes links
What this PR does:
Seems we do not have the tariff in the gov.uk robots file, so this PR adds the rel attribute to nofollow to the atom/changes links to prevent spiders crawling them.
[TARIFF16] Reject Spiders / No follow / No Index
|
2025-04-01T06:38:04.086764
| 2024-02-26T18:49:24
|
2154871420
|
{
"authors": [
"bjing2016",
"jfreeze95"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4253",
"repo": "bjing2016/alphaflow",
"url": "https://github.com/bjing2016/alphaflow/issues/3"
}
|
gharchive/issue
|
Example Input Files Not Working
Great work on the latest version of the paper and thanks for putting this repo out.
I was trying to test the basic inference you outlined using either the ESMFlow or AlphaFlow models and weights and ran into problems at every corner. I'll detail my specific issues below but repos always get increased usage when authors provide at least one full example input line for inference, so if you provide that I'm sure it would help many people checking out your code. Thanks!
Trying ESMFlow Model
mkdir output
mkdir weights
python predict.py --mode esmfold --input_csv splits/atlas_test.csv --weights weights/esmflow_md_distilled_202402.pt --samples 5 --outpdb output/
Output
2024-02-26 12:54:34,511 [---] [INFO] Loading the model
2024-02-26 12:55:16,878 [---] [INFO] Model has been loaded
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 5/5 [00:25<00:00, 5.08s/it]
Traceback (most recent call last):
File "/---/alphaflow/predict.py", line 132, in
main()
File "/---/miniconda3/envs/AlphaFlow/lib/python3.9/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "/---/alphaflow/predict.py", line 126, in main
f.write(protein.prots_to_pdb(result))
File "/---/alphaflow/alphaflow/utils/protein.py", line 163, in prots_to_pdb
prot = to_pdb(prot)
File "/---/miniconda3/envs/AlphaFlow/lib/python3.9/site-packages/openfold/np/protein.py", line 341, in to_pdb
chain_index = prot.chain_index.astype(np.int32)
AttributeError: 'NoneType' object has no attribute 'astype'
Tried with esmflow_pdb_base_202402.pt weights as well...same result.
Trying AlphaFlow Model
Preparing the MSA
python -m scripts.mmseqs_query --split splits/atlas_test.csv --outdir output
COMPLETE: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████| 450/450 [elapsed: 00:02 remaining: 00:00]
SUCCESS!
Running Inference
python predict.py --mode alphafold --input_csv splits/atlas_test.csv --msa_dir output/ --weights weights/alphaflow_pdb_distilled_202402.pt --samples 5 --outpdb output/
2024-02-26 13:17:56,383 [---] [INFO] Loading the model
Traceback (most recent call last):
File "/---/alphaflow/predict.py", line 132, in
main()
File "/---/miniconda3/envs/AlphaFlow/lib/python3.9/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "/---/alphaflow/predict.py", line 78, in main
model = model_class(**ckpt['hyper_parameters'], training=False)
File "/---/alphaflow/alphaflow/model/wrapper.py", line 496, in init
self.model = AlphaFold(config,
File "/---/alphaflow/alphaflow/model/alphafold.py", line 73, in init
self.extra_msa_stack = ExtraMSAStack(
TypeError: init() missing 2 required positional arguments: 'opm_first' and 'fuse_projection_weights'
Thanks again for your assistance. Looking forward to trying out this great work.
Can you check if the OpenFold version is correct? A previous version of the README had the wrong install command --- see https://github.com/bjing2016/alphaflow/issues/2 where the issue sounds similar to what you describe with the ExtraMSAStack.
Yep, that was it. Thanks!
In case anyone asks, the git request for the commit from OpenFold you specify won't work if you have Cuda 12.3, which is what led me to try the latest OpenFold commit which did work with 12.3 before I got Cuda 11.6 setup.
Also, in mmseqs_query line 284 you have a hard-coded iloc[:3] which only allows the first three entries of a csv file to be processed. I ran into that when trying to use the atlas_test.csv file which will then throw an error during inference when it can't find the 4th entry's mmSeq folder.
Thanks, mmseqs_query has been fixed.
|
2025-04-01T06:38:04.093561
| 2024-01-17T15:36:41
|
2086449441
|
{
"authors": [
"bjlittle",
"trexfeathers"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4254",
"repo": "bjlittle/geovista",
"url": "https://github.com/bjlittle/geovista/pull/645"
}
|
gharchive/pull-request
|
Refactor samples, pantry etcetera
🚀 Pull Request
Description
Applying some lessons learned from introducing GeoVista's 'assets' structure to new developers - seeing what was easy to grasp, and what was hard.
If we like this: corresponding changes will be needed in the structure of geovista-data.
Please take a look at the new structure in my branch, to get a feel for it.
Made geovista.cache into a package directory, allowing the inclusion of registry.txt and an explanatory README.md.
Made a geovista.pantry package directory, encompassing several sorts of reusable things that were previously scattered through several root modules:
fetch_coastlines() has moved to the geovista.pantry root
The previous ~geovista.pantry~ has become geovista.pantry.data.
~geovista.samples~ has become geovista.pantry.meshes.
The texture routines previously in ~geovista.cache~ have moved to geovista.pantry.textures.
Hopefully I've caught everything - refactoring can be rather fraught!
Thanks @bjlittle!
@all-contributors please add @trexfeathers for maintenance
|
2025-04-01T06:38:04.097361
| 2020-06-25T19:07:38
|
645788037
|
{
"authors": [
"andrew-platt",
"bjonkman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4255",
"repo": "bjonkman/openfast",
"url": "https://github.com/bjonkman/openfast/pull/12"
}
|
gharchive/pull-request
|
f/Linear -- documentation updates for nodal outputs
I added some updates to the documentation for the nodal outputs. I also added a rough skeleton of documentation for ElastoDyn that will need to be filled out sometime later.
During testing, I also added a few minor changes to the nodal output parsing.
Thanks, Andy!
|
2025-04-01T06:38:04.103094
| 2020-07-08T19:56:17
|
653556551
|
{
"authors": [
"bjorn3",
"zwhitchcox"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4256",
"repo": "bjorn3/rustc_codegen_cranelift",
"url": "https://github.com/bjorn3/rustc_codegen_cranelift/issues/1060"
}
|
gharchive/issue
|
[arm] error: couldn't load codegen backend
Hi, I'm trying to follow the build instructions, but when I get to test.sh, I get this error:
error: couldn't load codegen backend "/home/pi/codegen/rustc_codegen_cranelift/target/release/librustc_codegen_cranelift.so": "/home/pi/codegen/rustc_codegen_cranelift/target/release/librustc_codegen_cranelift.so: undefined symbol: __register_frame"
Are you using glibc or another libc? Just guessing.
That symbol should be provided by libunwind: https://github.com/bjorn3/rustc_codegen_cranelift/blob/eb5ce4e92ae8d512804279fda1101032c7ec9f28/src/debuginfo/unwind.rs#L136
Not sure if this answers your question, but this is the output of ldd --version
ldd --version
ldd (Debian GLIBC 2.28-10+rpi1) 2.28
Copyright (C) 2018 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
Written by Roland McGrath and Ulrich Drepper.
pi@raspberrypi:~/codegen/rustc_codegen_cranelift
Ah, you are using an arm system. While the error is unrelated, please note thatthe AArch64 backend of Cranelift is still missing some things necessary for cg_clif (mostly 128bit int support) and arm32 is completely missing.
Ok, I see, so this is a lost cause, and I should close the issue?
btw, I just installed libunwind-dev, and I got the same error
You can keep it open. Once Cranelift implements the necessary features I do want to get cg_clif fully functioning on AArch64.
AArch64 support is now almost complete. It only needs a couple of changes to Cranelift that have already landed on main support to fix the remaining tests. I'm not sure why __register_frame wasn't found for you. If you still have this issue you can build with --no-unstable-features to disable the JIT.
|
2025-04-01T06:38:04.113560
| 2020-10-23T04:20:18
|
727894338
|
{
"authors": [
"benforeva",
"bkeepers",
"marnen"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4257",
"repo": "bkeepers/dotenv",
"url": "https://github.com/bkeepers/dotenv/pull/416"
}
|
gharchive/pull-request
|
Fix template (-t) handling of export declarations
Fixed export commands being repeated in templated values.
Can we keep this open? I like to use export syntax in my .env files, and having to go through the template and replace =export with = is a bit of a drag. If I can help get this ready for merge, let me know and I'll be happy to.
Commenting in the hope that this will be kept open and fixed.
Changes look great. Sorry for the delay. I will merge now and push out a new release soon.
|
2025-04-01T06:38:04.121581
| 2022-05-06T16:19:45
|
1228059705
|
{
"authors": [
"hazarjast",
"jericsmith504"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4258",
"repo": "bkerler/netgear_telnet",
"url": "https://github.com/bkerler/netgear_telnet/issues/13"
}
|
gharchive/issue
|
Updated telnetenabled on NBR750
Hello sir! Hope all is well with you these days. I see Netgear has changed the 'magic packet' for telnetenabled on my new router NBR750 and so sadly this 'telnet-enable2.py' is no longer working for it. Was hoping you might have a chance to take a peek at the updated binary/libraries for it so that maybe we could fork an updated version of the python script for NBR750 and newer routers.
As always happy to send some pizza your way for your valuable time. I have zipped up binary along with library dependencies in the download link that follows along with a .txt inside that gives strings, strace, and ldd output which hopefully helps. Cheers!
http://paste.c-net.org/RaquelBuffalo
(SHA-256 .zip checksum: 517f4422d3b6ac36d20bfaedb5e80d094b834927e4901d9e15f4bf96c3482430)
Worked with @bkerler on updating this telnet enabler for the NBR750 and he found that the final command to created the hashed password in the magic packet had to be updated to '...hexdigest().lower()'; basically lowercase instead of uppercase used in the LBR20. However, even after this was done and the magic packet could be successfully sent, the telnet daemon still would not launch so it seems Netgear has changed something even deeper and/or otherwise broken the telnet enabler daemon.
I spun my wheels on it for a few weeks but ultimately decided upon a workaround instead. At least on the NBR750 initial stable release OEM firmware, Netgear has brought back the NVRAM parameter 'telnetd_enable' which by default is set to '0' to disable it. However, if connected via serial console one can set it to enabled with the command 'fnvram set telnet_enable=1' followed by 'fnvram commit'. After reboot telnet daemon will be running so you can telnet into the device. From there you can either continue to use telnet or enable SSH instead (which I recommend).
To enable telnet easily to start with for those without access to a serial console I've taken a config backup which captures the 'telnetd_enable=1' parameter value so that anyone on the OEM initial release firmware can restore it and gain telnet access. As part of the config backup it also overwrites things like 'admin' and wifi passwords along with wifi SSID but those can be changed post-restore. All the details including the required config backup file are located in the thread below with instructions. Hope this is helpful.
https://wirelessjoint.com/viewtopic.php?p=24894
Closing as per the workaround in my last comment.
@bkerler pinged me end of October 2022 and indicated that there was an additional change in the python script required that was identified and that I should not download the updated script to test. Unfortunately, other work piled up and I am now just getting back to testing and validation of this on the various firmware that Netgear have released for the NBR750 since the last time I tested. Will post the result here momentarily.
I have successfully tested the updated 'telnet-enable2.py' script from this repo on the NBR750 for the following firmware versions:
V<IP_ADDRESS>_1.5.50
V<IP_ADDRESS>_1.5.63
V<IP_ADDRESS>_1.5.64
It is confirmed functional for all of these now. Thanks so much for your expertise and effort to make this work :)
Confirmed script is functional on V<IP_ADDRESS>_1.5.66
|
2025-04-01T06:38:04.182126
| 2018-10-18T18:51:13
|
371663319
|
{
"authors": [
"Blackbaud-BobbyEarl",
"codecov-io"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4259",
"repo": "blackbaud/skyux-builder",
"url": "https://github.com/blackbaud/skyux-builder/pull/489"
}
|
gharchive/pull-request
|
Using the newly created chromeDriverVersionFinder.
Issue: https://github.com/blackbaud/skyux2/issues/2124
Codecov Report
Merging #489 into master will not change coverage.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #489 +/- ##
=====================================
Coverage 100% 100%
=====================================
Files 54 54
Lines 1653 1664 +11
Branches 245 246 +1
=====================================
+ Hits 1653 1664 +11
Flag
Coverage Δ
#builder
100% <100%> (ø)
:arrow_up:
#runtime
100% <ø> (ø)
:arrow_up:
#srcapp
100% <ø> (ø)
:arrow_up:
Impacted Files
Coverage Δ
cli/e2e.js
100% <100%> (ø)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update f9d7447...fbae359. Read the comment docs.
|
2025-04-01T06:38:04.198436
| 2016-04-26T13:03:53
|
151128669
|
{
"authors": [
"Nrated",
"divinity76"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4260",
"repo": "blackdtools/Blackd-Proxy-CLASSIC",
"url": "https://github.com/blackdtools/Blackd-Proxy-CLASSIC/issues/80"
}
|
gharchive/issue
|
Cavebot sometimes incorrectly detects GM's when going down floors
there is some weird bug wherein this error will occur:
mychar closed with this message: Client closed - condition onGMcloseConnection was activated: GM Tsar
when blackd proxy have seen GM Tsar when leaving depot, and cavebot has onGMcloseConnection, it may sometimes (completely incorrectly!!) believe that GM Tsar is nearby when the cavebot is changing floor down!
Whoaaaaah I just logged in to my e-mail account and from what I can see there's a lot of new things here! Keep it up mate
^ i started playing again ^^
anyway, i found out that it's actually a bug in this OT, it will sometimes send info about players which is like 100 SQM away when walking up/down floors. like here, i am very very far away from the GM in question, yet the server sends me info about everyone in DP every time i go down a floor. i have no idea why. weird OT custom code bug.
' we already knew his ID + include some info
tempID = FourBytesDouble(packet(pos + 2), packet(pos + 3), packet(pos + 4), packet(pos + 5))
AddID_HP idConnection, tempID, packet(pos + 6) 'update hp
nameofgivenID = GetNameFromID(idConnection, tempID)
|
2025-04-01T06:38:04.207620
| 2019-12-31T01:14:51
|
544057248
|
{
"authors": [
"MrsFlux"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4261",
"repo": "blackflux/lambda-monitor",
"url": "https://github.com/blackflux/lambda-monitor/pull/1524"
}
|
gharchive/pull-request
|
[Gally]: master <- dev
Automatically created by Git-Ally
:tada: This PR is included in version 2.1.5 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:38:04.215035
| 2022-03-27T22:40:20
|
1182687471
|
{
"authors": [
"drabbit17",
"rlouf"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4262",
"repo": "blackjax-devs/blackjax",
"url": "https://github.com/blackjax-devs/blackjax/issues/185"
}
|
gharchive/issue
|
Docs and examples using refactored APIs while most recent release still using the old ones
Bug Description
I was trying to reproduce the Logistic Regression example using version 0.3.0 and got the following error when running line rmh_sampler = blackjax.rmh(logprob_fn, sigma=jnp.ones(M) * 0.7).
I also tried to run the other examples in the Introduction notebook as well as the one in the README without success.
It took me a while to realise that the APIs in the last release 0.3.0 have been considerably refactored with !159 and that all the docs are using the new APIs.
I think that it would be nice if you could specify in the README as well as in the main doc that the examples are all using a still to be released version and that to make them work you need to install directly from a local clone of the project.
Versions
BlackJAX 0.3.0
Python 3.8.10 (default, Nov 26 2021, 20:14:08)
[GCC 9.3.0]
Jax 0.2.28
Jaxlib 0.1.76
Thank you for raising the issue. I am going to do even better than this and release a new version of the package!
Done!
|
2025-04-01T06:38:04.220952
| 2017-06-13T15:53:56
|
235603054
|
{
"authors": [
"blacklabelops",
"marton78"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4263",
"repo": "blacklabelops/jira",
"url": "https://github.com/blacklabelops/jira/issues/34"
}
|
gharchive/issue
|
Can't connect to Atlassian Marketplace due to DNS issues
Sorry if this is a general Docker issue and not due to your images, but JIRA can't connect to Marketplace. I guess it's due to name resolution not working from inside the container:
# docker exec -it jira /bin/bash
bash-4.3$ nslookup google.com
nslookup: can't resolve '(null)': Name does not resolve
nslookup: can't resolve 'google.com': Try again
Any idea on how to fix this? I'm using your docker-compose.yml.
PS:
Docker version 17.05.0-ce, build 89658be
docker-compose version 1.13.0, build 1719ceb
Please check this page: https://docs.docker.com/engine/userguide/networking/default_network/configure-dns/
Yes, I know this page, but it doesn't speak about docker-compose. I have changed your docker-compose.yml to include the setting dns: <IP_ADDRESS> and verified that docker inspect shows this address as dns, but nslookup still fails. Maybe it's related to this issue, where users claim the problem is due to the line options ndots:0 injected into resolv.conf. Indeed the container's resolv.conf looks like:
search mydomain.com
nameserver <IP_ADDRESS>
options ndots:0
Does name resolution work for you with your provided docker-compose.yml?
I verified:
$ curl -O https://raw.githubusercontent.com/blacklabelops/jira/master/docker-compose.yml
$ docker-compose up -d
$ docker-compose exec jira
$ nslookup google.com
Name: google.com
Address 1: <IP_ADDRESS> muc11s04-in-f14.1e100.net
Address 2: 2a00:1450:4016:807::200e muc11s04-in-x0e.1e100.net
Your docker-demon is not able to configure the network bridge correctly. You container's are not able to connect to the internet. Can't help you with more.
Thanks, @blacklabelops. The issue was entirely unrelated to Docker: our internal DNS server didn't accept recursion from Docker's subnet. I needed to add
allow-recursion { <IP_ADDRESS>; <IP_ADDRESS>/8; <IP_ADDRESS>/12; };
to bind.conf. Confusion arose from the line nslookup: can't resolve '(null)': Name does not resolve which is apparently unique to alpine, other containers don't show it. Now the line still appears, but name resolution works:
# docker exec -it jira nslookup google.com
nslookup: can't resolve '(null)': Name does not resolve
Name: google.com
Address 1: <IP_ADDRESS> ams15s30-in-f14.1e100.net
Address 2: 2a00:1450:400e:807::200e ams15s30-in-x0e.1e100.net
Thanks, closing.
|
2025-04-01T06:38:04.229880
| 2022-03-17T03:00:54
|
1171871949
|
{
"authors": [
"AB1908",
"s-blu",
"spasticginger"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4264",
"repo": "blacksmithgu/obsidian-dataview",
"url": "https://github.com/blacksmithgu/obsidian-dataview/issues/956"
}
|
gharchive/issue
|
Empty value in bracketed inline data overwriting default value in table
What happened?
This error occurred on both deesktop & mobile versions of Obsidian
Frontmatter & unbracketed key-value pairs performed as expected, however an empty bracketed inline data overwrites a table's default value with blank. Screenshot of source & preview mode attached.
issue can be recreated :
Test1:
Test2::
[Test3::]
DQL
Table
default(Test1, "👽") AS "1",
default(Test2, "🛸") AS "2",
default(Test3, "🐮") AS "3"
WHERE Type = "Test"
JS
No response
Dataview Version
0.4.26
Obsidian Version
0.13.33
OS
Windows
In Discord, was asked to also run ' = this` query. Results attached
Also, was asked to reassign issue to @AB1908 but I can't seem to find a way to do that (using mobile GitHub)
Test1 and Test2 evaluate to undefined and null. Test2 is not indexed at all as you can see here:
default, per the docs, changes default values for values that are null, and also appears to work for undefined. However, Test3 has an empty string "", which is not null, which is why the field is not populating as you expect it to.
Ah that makes sense now! I didn't realize the brackets would automatically make it a blank string, whereas the other formats don't.
Thank you for your help!!
Admittedly I was using the brackets primarily just to be able to target the data with CSS. Is there a way to target the Test2 format? If not, I totally understand, styling kind of goes beyond the point of reading & working with raw data.
I'm not aware of how to apply CSS to inline fields. Let me poke around and get back to you.
Sorry for the late check on this but any progress? If not, I'll dig it up again.
Hello,
sorry to come back to you so late. I currently go through older issues to see which are stale or already solved.
Theres a FR in #713 to be able to style non-bracketed inline fields. If I get you right, that's what you want to have? If this is the only open point on this issue and if it's okay, I'd like to close the issue in favor or #713.
That the bracketed inline field is set to an empty string as value is intended behaviour, as far as I can tell - in fact, it is the only way I am aware of to use an empty value.
|
2025-04-01T06:38:04.325295
| 2024-10-19T08:19:58
|
2598891885
|
{
"authors": [
"Norrox",
"Starkium"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4267",
"repo": "blazium-engine/blazium",
"url": "https://github.com/blazium-engine/blazium/issues/74"
}
|
gharchive/issue
|
Default view on scenes when importing to blazium 4.3 default to wireframe
Tested versions
v4.3.stable.custom_build [ce7311f34]
System information
Windows 11, Ryzen 9 7900X, 7900XTX, 64G Ram
Issue description
Other need to check this, or if this was something fishy with my project...
Majority of my scenes defaulted to wireframe when imporing from godot to blazium
Steps to reproduce
Minimal reproduction project (MRP)
is this still an issue?
|
2025-04-01T06:38:04.328073
| 2024-08-01T13:15:43
|
2442433828
|
{
"authors": [
"bldambtn"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4268",
"repo": "bldambtn/WhoDoYouWorkFor",
"url": "https://github.com/bldambtn/WhoDoYouWorkFor/issues/3"
}
|
gharchive/issue
|
AC-3: Roles
AC-3: Roles
User Story
AS A business owner
I WANT to be able to view and manage the departments, roles, and employees in my company
SO THAT I can organize and plan my business
Acceptance Criteria
GIVEN a command-line application that accepts user input
WHEN I choose to view all roles
THEN I am presented with the job title, role id, the department that role belongs to, and the salary for that role
To fulfill the acceptance criteria for the viewAllRoles function in your command-line application, you need to retrieve the necessary information from your database and present it to the user in a formatted manner. Here's a step-by-step guide on how you can achieve this:
Retrieve Roles Data: Query your database to fetch all roles along with their associated information such as job title, role ID, department ID, and salary.
Format Data: Once you have the roles data, format it in a way that displays the job title, role ID, department name (instead of ID), and salary for each role.
Display Data: Present the formatted roles data to the user in a clear and organized manner, such as printing it in a table format.
|
2025-04-01T06:38:04.361632
| 2016-04-20T10:06:32
|
149723338
|
{
"authors": [
"jokeyrhyme"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4269",
"repo": "blinkmobile/blinkmrc.js",
"url": "https://github.com/blinkmobile/blinkmrc.js/issues/2"
}
|
gharchive/issue
|
alternatives to appdirectory dependency
https://github.com/LinusU/node-application-config
https://github.com/sindresorhus/env-paths
|
2025-04-01T06:38:04.400122
| 2017-02-21T16:09:49
|
209192397
|
{
"authors": [
"RomiValladares",
"blipinsk"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4270",
"repo": "blipinsk/FlippableStackView",
"url": "https://github.com/blipinsk/FlippableStackView/issues/27"
}
|
gharchive/issue
|
How to make views appear from the bottom?
I want to make the views appear from the bottom rather than from the top. Is that possible?
Check my comment on the issue #3
|
2025-04-01T06:38:04.408317
| 2019-08-25T21:12:31
|
484976492
|
{
"authors": [
"dlamkins",
"greaka"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4271",
"repo": "blish-hud/Blish-HUD",
"url": "https://github.com/blish-hud/Blish-HUD/pull/106"
}
|
gharchive/pull-request
|
Contexts + CdnInfoContext & Gw2ClientContext
Added base structure for new ContextsService. Also added two contexts CdnInfoContext and Gw2ClientContext which demonstrate detecting the Client version of the application using the Mumble Link service to detect the current build ID and then the asset CDNs to check which client we are using based on that current build ID.
This provides some of the potential implementation currently being discussed in #62.
Additionally, if accepted, this contains contexts that would allow us to more easily solve #50.
Resolved bug with CdnInfo where parsedSuccessfully could be true if the number of values was not 5. https://github.com/blish-hud/Blish-HUD/pull/106#discussion_r318001999
Reduced complexity of "Is______ClientType" function. https://github.com/blish-hud/Blish-HUD/pull/106#discussion_r318009487 & https://github.com/blish-hud/Blish-HUD/pull/106#discussion_r318009731
Fixed bugs in CdnInfoContext as well as Gw2ClientContext.
Contexts can now be unloaded.
Gw2MumbleService now has an event "BuildIdChanged"
OverlayService now, in addition to TacO, checks to see if the running client is the standard client or the Chinese client (moves the CornerIcons over if it is the Chinese client).
@greaka While I think the primary focus of this PR is the Contexts implementation, I think it'd be good to have Lei test the artifact that is created from this build and ensure the Blish HUD CornerIcon moves over as expected on a real Chinese version of the client. I did a good amount of testing with Fiddler to fake the build ID that is returned and was able to confirm it should work. Also ensured that it would fallback gracefully if something fails with the request (and logs are much more detailed with why it failed, now).
If you like this latest implementation, I can finish writing up the XML docs for it and mark it ready for a true review.
Once we get this merged, we can start adding things like the FestivalContext, which I am fairly excited about. 🎉
@greaka Please re-review the implementation when you have a chance. I believe I have implemented everything now and have put in XML documentation where appropriate. I've also added a State member to Contexts which indicates if the Context is loading, ready, etc.
To prevent anybody from unregistering a Context, I added a ContextHandle which can be used to expire a Context. The ContextHandle associated with a Context type is returned when you call RegisterContext. The full workflow would look something like:
// Register the context when the module is loaded
var myContextHandle = GameService.Contexts.RegisterContext(new ExampleContext());
// Expire context when the module is unloaded
myContextHandle.Expire();
Also, I received confirmation that this implementation has fixed Lei's problem! 🙂
https://cdn.discordapp.com/attachments/568470477543178261/617894300327477249/jt_2019-09-02_09-30-59.png
I have multiple nitpicks in case the contexts are meant to be thread safe.
Having them threadsafe is probably a nice to have for now though
Agreed!
|
2025-04-01T06:38:04.442819
| 2022-06-29T11:49:19
|
1288601197
|
{
"authors": [
"cmalex23"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4272",
"repo": "blocknative/web3-onboard",
"url": "https://github.com/blocknative/web3-onboard/issues/1107"
}
|
gharchive/issue
|
[Issue]: Wallet addresses shouldn't be lowercase
Is your request related to a problem?
The last update made the wallet address lowercase by default, which changes the behavior of things like Blockies which now have different colors than the ones in MetaMask for example.
Feature Description
No response
Alternative Solutions
Maybe don't call .toLowerCase() on the addresses, let the user to that when needed. Or was there a specific reason you made this change?
Anything else?
No response
Hm actually this seems fine on desktop Metamask, but on their mobile app, Blockies have different colors.
It may not be related to this package so we may close it.
|
2025-04-01T06:38:04.450339
| 2018-04-13T22:16:10
|
314265729
|
{
"authors": [
"friedger",
"larrysalibra"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4273",
"repo": "blockstack/blockstack-android",
"url": "https://github.com/blockstack/blockstack-android/issues/12"
}
|
gharchive/issue
|
As a developer, I want to better understand which callbacks run in which threads
In most callbacks it is necessary to user runOnUIThread.
This should be documented or the SDK should call the callbacks from the UI thread already.
Thanks @friedger - i made a note of this in the comments.
Going to close this issue. Please open a new issue if you think think it would be better to set these callbacks up to run on the UI thread and let me know if you'd like to take a stab at it.
|
2025-04-01T06:38:04.476714
| 2019-03-08T19:43:22
|
418930147
|
{
"authors": [
"moxiegirl",
"yknl"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4274",
"repo": "blockstack/blockstack.js",
"url": "https://github.com/blockstack/blockstack.js/issues/619"
}
|
gharchive/issue
|
Installation of Blockstack 19.0.0 includes a blockstack-18.3.0-rc1.tgz file
Installing blockstack 19.0.0 comes with a blockstack-18.3.0-rc1.tgz file. Was confusing to me when I listed the module contents.
npm ls shows the right version.
We should remove the confusing tgz file.
$ ls -al node_modules/blockstack/
total 6328
drwxr-xr-x 23 manthony wheel 736 Mar 8 11:38 .
drwxr-xr-x 148 manthony wheel 4736 Mar 8 11:38 ..
-rw-r--r-- 1 manthony wheel 200 Oct 26 1985 .babelrc
drwxr-xr-x 3 manthony wheel 96 Mar 8 11:38 .circleci
-rw-r--r-- 1 manthony wheel 118 Oct 26 1985 .eslintignore
-rw-r--r-- 1 manthony wheel 1395 Oct 26 1985 .eslintrc
-rw-r--r-- 1 manthony wheel 242 Oct 26 1985 .flowconfig
drwxr-xr-x 17 manthony wheel 544 Mar 8 11:38 .nyc_output
-rw-r--r-- 1 manthony wheel 8353 Oct 26 1985 CHANGELOG.md
-rw-r--r-- 1 manthony wheel 2137 Oct 26 1985 CONTRIBUTING.md
-rw-r--r-- 1 manthony wheel 1082 Oct 26 1985 LICENSE
-rw-r--r-- 1 manthony wheel 3746 Oct 26 1985 README.md
-rw-r--r-- 1 manthony wheel 2412148 Oct 26 1985 blockstack-18.3.0-rc1.tgz
-rw-r--r-- 1 manthony wheel 889 Oct 26 1985 bower.json
drwxr-xr-x 4 manthony wheel 128 Mar 8 11:38 dist
drwxr-xr-x 4 manthony wheel 128 Mar 8 11:38 docs
-rw-r--r-- 1 manthony wheel 4545 Oct 26 1985 docs-button.png
-rw-r--r-- 1 manthony wheel 761104 Oct 26 1985 docs.json
-rw-r--r-- 1 manthony wheel 686 Oct 26 1985 documentation.yml
drwxr-xr-x 3 manthony wheel 96 Mar 8 11:38 flow-typed
drwxr-xr-x 18 manthony wheel 576 Mar 8 11:38 lib
-rw-r--r-- 1 manthony wheel 5274 Mar 8 11:38 package.json
drwxr-xr-x 7 manthony wheel 224 Mar 8 11:38 tests
$ npm ls blockstack
/private/tmp
└──<EMAIL_ADDRESS>
This should be fix in v19.1.0
@yknl yo blockstack gave me this 5 minutes ago. that's what we use to build the initial hello world
manthony at booboo in /tmp/test-download/node_modules/blockstack
$ ls
CHANGELOG.md bower.json docs.json package.json
LICENSE dist documentation.yml tests
README.md docs flow-typed
blockstack-18.3.0-rc1.tgz docs-button.png lib
I don't see this file anymore in the latest versions. The app generator should have been updated as well.
|
2025-04-01T06:38:04.481499
| 2020-08-04T09:07:57
|
672634320
|
{
"authors": [
"aulneau",
"jasperjansz"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4275",
"repo": "blockstack/docs.blockstack",
"url": "https://github.com/blockstack/docs.blockstack/issues/686"
}
|
gharchive/issue
|
Was this page helpful? section improvements
Align the edges with the content
Icons draw too much attention because they're too large. Should be 28px instead of 42px
Icons missing background, attached svgs!
Happy.svg
@jasperjansz these are react components, the background is missing only on accident when I was making it work with dark mode. Could you design a dark mode version of the faces I can use? Thanks!
|
2025-04-01T06:38:04.483532
| 2021-05-21T22:52:12
|
898648067
|
{
"authors": [
"314159265359879",
"razpurge"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4276",
"repo": "blockstack/explorer",
"url": "https://github.com/blockstack/explorer/issues/446"
}
|
gharchive/issue
|
recuperacion de token
There is a possibility of recovering this transfer that you do without the memo for not knowing how to transfer
99,00 STX
SP34FXH035DMS7BDZCHF17Q0P0QXPT713BZZGBER4
SP1P72Z3704VMT3DMHPP2CB8TGQWGDBHD3RPR9GZS
0xf6a911512fd3c55afd8e85a91864db0d75fb529d582369d469b4f7dbe4dc96bf
0.00018 STX
#15805
Hace 19 horas
0xa847f231c4261882fa4f737b6d14f2c3bb3ef17de3a70a52b0033b929499cce3
This is a transaction to Binance, without a memo. That means they can not automatically credit it to your account but you can contact them with these details and I am sure they can sort it out for you. Nobody here can do anything about it though.
This issue can be closed.
|
2025-04-01T06:38:04.490623
| 2020-11-24T13:01:55
|
749687834
|
{
"authors": [
"markmhx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4277",
"repo": "blockstack/ux",
"url": "https://github.com/blockstack/ux/issues/689"
}
|
gharchive/issue
|
Integrate identity, latest transaction and assets into home modal
See screen design
Identity
If the user hasn't set a username for the currently selected address, the address is shown (needs design)
If user has set a username, the username shows primarily with the address secondarily (needs design)
In both cases, there are options to both copy the address to clipboard and visit the address page on the Explorer (needs design)
Latest transaction
If the user hasn't had a transaction within the past 48 hours, this section is hidden entirely
If they've had one or more transaction within the past 48 hours, the latest is shown in this section. The image, ticker and name should be set as defaults for any of the tokens involved in the transaction.
The amount listed should be positive if the token was received and negative if sent.
NFTs should show "1" or "-1" whereas FTs should show the exact amount transferred.
Tokens (fungible tokens)
If the user has no tokens, they should be directed to buy STX with link to CoinMarketCap (see design)
If the user has one or more tokens, they should be listed with default image, ticker and name values
Collectibles (non-fungible tokens)
If the user has no collectibles, they should simply see "You don't own any collectibles" message.
If the user has one or more collectibles, they should be listed with default image, ticker and name values and no value on the right (since they're all implicitly "1")
@jasperjansz it appears the identity area of the home modal needs some small updates here?
@hstove what's the best way to test multiple assets in the UI here?
@hstove what's the best way to test multiple assets in the UI here?
|
2025-04-01T06:38:04.502397
| 2015-04-22T03:20:49
|
70009613
|
{
"authors": [
"bijugs"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4278",
"repo": "bloomberg/chef-bach",
"url": "https://github.com/bloomberg/chef-bach/issues/137"
}
|
gharchive/issue
|
Upgrade JMXTrans
JMXTrans community cookbook points to an older version (2012) of JMXTrans software to install. We need to use the latest version of the software which can fix some of the stats data issues seen in Graphite. Also it will help us get support from the JMXTrans community.
Version 2.0 of the jmxtrans-cookbook is available for testing. The default recipe does the install from the tar.gz file published by the jmxtrans project and install_ubuntu uses the deb package. Also included is remove-ver1 to remove jmxtrans installed using the previous version of the cookbook.
Fixed as part of PR #556.
|
2025-04-01T06:38:04.531343
| 2024-09-26T07:31:01
|
2549751579
|
{
"authors": [
"royvardhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4279",
"repo": "bluealloy/revm",
"url": "https://github.com/bluealloy/revm/pull/1808"
}
|
gharchive/pull-request
|
feat: extract statetest models/structs to standalone crate
Closes: #1787
Cool! lgtm with one nit.
Done! Ci green.
|
2025-04-01T06:38:04.532396
| 2016-07-14T07:22:37
|
165494371
|
{
"authors": [
"Ceda",
"mmagnusek"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4280",
"repo": "blueberryapps/blueberry_rails",
"url": "https://github.com/blueberryapps/blueberry_rails/issues/46"
}
|
gharchive/issue
|
Add rails env to Airbrake
Airbrake.configure do |config|
config.host = 'http://errors.blueberry.cz'
config.project_id = -1
config.project_key = 'API_KEY'
config.environment = Rails.env
config.ignore_environments = %w(development test)
end
https://github.com/airbrake/airbrake-ruby#blacklist_keys
|
2025-04-01T06:38:04.533950
| 2015-08-10T16:34:34
|
100109727
|
{
"authors": [
"bbc-jenkins",
"lihkin213"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4281",
"repo": "blueboxgroup/ursula",
"url": "https://github.com/blueboxgroup/ursula/pull/1129"
}
|
gharchive/pull-request
|
Keystone ldap
Updates added for deploying openldap, setting up password policies, and configuring keystone. By default openldap is not deployed, it needs to be enabled explicitly.
Can one of the admins verify this patch?
"ok to test" to accept this pull request for testing
"test this please" for a one time test run
"add to whitelist" to add the author to the whitelist
|
2025-04-01T06:38:04.535585
| 2016-11-28T22:16:27
|
192132224
|
{
"authors": [
"nirajdp76",
"paulczar"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4282",
"repo": "blueboxgroup/ursula",
"url": "https://github.com/blueboxgroup/ursula/pull/2300"
}
|
gharchive/pull-request
|
wip: experimental CentOS support
demo: ursula --vagrant envs/example/allinone-centos site.yml
this is just enough to get nova, glance, keystone, neutron seemingly fuctioning
correctly on CentOS.
have disabled heat, lbaas, etc. also disabling logging/monitoring/etc.
retest
retest
retest
|
2025-04-01T06:38:04.537374
| 2021-08-30T14:47:46
|
982875491
|
{
"authors": [
"bluecmd"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4283",
"repo": "bluecmd/fortigate_exporter",
"url": "https://github.com/bluecmd/fortigate_exporter/pull/136"
}
|
gharchive/pull-request
|
doc(readme): Matrix chat room
Add a chat room for discussions
:tada: This PR is included in version 1.14.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:38:04.545977
| 2016-10-08T07:08:24
|
181809765
|
{
"authors": [
"CompositionCloud",
"bluejamesbond",
"ericmguimaraes",
"zuraba"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4284",
"repo": "bluejamesbond/TextJustify-Android",
"url": "https://github.com/bluejamesbond/TextJustify-Android/issues/127"
}
|
gharchive/issue
|
last line when textFormat="formatted"
Hello,
The last line is automatically justified when using textFormat="formatted". I've seen in another issue here that you suggest to use the JustifiedSpan. The problem with that is I cannot know in advance how long the last line will be, and the span must consist of the entire line (so setting it just to the last character/word won't help). When using textFormat="plain" it works well. I've attached 2 screenshots, so you can see the difference.
Please let me know if there's something I got wrong. If you have any idea of how to solve it, I would be extremely thankful.
Goni
|
@gkrishnan
+1
any update??
|
2025-04-01T06:38:04.554448
| 2020-06-10T20:42:35
|
636535142
|
{
"authors": [
"Pushkal-G",
"joel-bluedata"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4285",
"repo": "bluek8s/kubedirector",
"url": "https://github.com/bluek8s/kubedirector/issues/330"
}
|
gharchive/issue
|
enforce selectedRoles
selectedRoles isn't really useful until we get choice-driven config variation modelling, but we should make sure that it works as expected for now so that things don't break when config choices are implemented.
We do have validation that any selected role has a role configuration in the kdapp. What we don't have is awareness in the cluster setup that only selected roles should be paid attention to.
Hi @joel-bluedata , I am new to open source. Could I please take it up and start working?
|
2025-04-01T06:38:04.559154
| 2017-03-06T02:52:00
|
212011716
|
{
"authors": [
"Jofairden",
"Kimi-Arthur",
"bluemagic123"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4286",
"repo": "bluemagic123/tModLoader",
"url": "https://github.com/bluemagic123/tModLoader/issues/134"
}
|
gharchive/issue
|
Allow translation in mods
Since Terraria will have full translation starting from 1.3.5, it will be great if tModLoader supports a similar pattern and make the switch of language the same way as the vanilla one.
Thanks!
I see as why jopo added the 'Far in Future' label. This will be quite a large undertaking and not to mention, this is actually already possible for mods, but with their own implementation.
Resolved with the recent updates to 1.3.5 (from 57418dd6a3e7abbc472dc0403e5ef415f80124fe to 00240dafc313bf708947a7719ba7d7d1d55a4d7a)
|
2025-04-01T06:38:04.570652
| 2022-06-17T02:26:28
|
1274413999
|
{
"authors": [
"ES-Alexander"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4287",
"repo": "bluerobotics/BlueOS-Water-Linked-DVL",
"url": "https://github.com/bluerobotics/BlueOS-Water-Linked-DVL/issues/4"
}
|
gharchive/issue
|
[Needs confirmation] GPS origin should be wrapped
This forum post found that setting the GPS origin in the Water Linked DVL interface with a -220 longitude didn't work.
@Williangalvani suspects that it may require a value between -180 to 180. If confirmed, this package should wrap the value before displaying to the user and/or before sending to the autopilot to ensure correct functionality.
Longitude wrapping added in #11
|
2025-04-01T06:38:04.576556
| 2023-09-04T23:13:34
|
1880949003
|
{
"authors": [
"gaearon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4288",
"repo": "bluesky-social/social-app",
"url": "https://github.com/bluesky-social/social-app/pull/1374"
}
|
gharchive/pull-request
|
Fix caret jumps for web composer
Fixes https://github.com/bluesky-social/social-app/issues/1300
Fixes caret jumping to the end when inserting an emoji in the middle of the post using our picker
Fixes caret jumping to the end when blurring out of the composer and focusing it back again (maybe controversial)
See https://github.com/bluesky-social/social-app/issues/1300#issuecomment-1705753534 for an explanation. TLDR:
We shouldn't be jumping the caret when the user picks an emoji. They might be in the middle of the post anyway.
We should not be reading mutable .length in order to "jump" the caret to the end — it appears that our text editing library does not guarantee whether the new value is flushed yet by the time onFocus fires.
Although we could have fixed that by using focus('end') instead of focus().setTextSelection(...), we probably shouldn't be jumping the caret to the end at all on focus anyway — since it's not what you actually want e.g. when inserting an emoji in the middle of the text. So I removed that line altogether unless we have some motivation to keep it. I haven't found a case where it's needed yet.
This might need a review from someone familiar with #1241 and #1254.
This needs more work. @ansh pointed out that we need to jump the cursor to the end when the user opens the composer on somebody else's profile. This is because we want to place it after the mention.
This needs more work. @ansh pointed out that we need to jump the cursor to the end when the user opens the composer on somebody else's profile. This is because we want to place it after the mention.
Pushed another commit. This ensures that if we press Compose on someone's profile, the cursor will appear after the handle.
|
2025-04-01T06:38:04.580953
| 2024-06-04T18:28:44
|
2334132466
|
{
"authors": [
"gaearon",
"haileyok"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4289",
"repo": "bluesky-social/social-app",
"url": "https://github.com/bluesky-social/social-app/pull/4356"
}
|
gharchive/pull-request
|
add types for desktopFixedHeight to List
Let's stop requiring @ts-ignore for this prop in our List implementation
Oh, actually need to look closer. In Views.web.tsx it's typed as boolean | number
I never implemented it for web actually. I think it's maybe used for native?
I'm not sure what it's really supposed to be doing. I think feeds use it.
Yea, the FlatList_INTERNAL takes both, but that makes the prop name a little confusing. I'll have to look through later to see why it gets used on native or if it even does.
I see that desktopFixedHeightOffset is a prop for Feed, but it never gets used either as far as I can tell?
Will leave number as well for now, but made a note to revisit this and remove number if possible. Seems fine, though want to give it a more solid pass before assuming that's true.
|
2025-04-01T06:38:04.585266
| 2024-06-20T15:46:22
|
2364764052
|
{
"authors": [
"danielballan"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4290",
"repo": "bluesky/databroker",
"url": "https://github.com/bluesky/databroker/pull/814"
}
|
gharchive/pull-request
|
Factor out bluesky-tiled-plugins package
This creates a second Python package in the databroker repository, bluesky-tiled-plugins, with the "special client" objects BlueskyRun, BlueskyEventStream, and CatalogOfBlueskyRuns, as well as the custom query objects PartialUID, ScanID, and TimeRange. Quoting the README:
For a user wishing to connect to a running Tiled server and access Bluesky data,
this package, along with its dependency tiled[client], is all they need.
The databroker package is only required if the user wants to use the legacy
databroker.Broker API.
This means it is no longer necessary to install databroker in the client environment unless they have legacy databroker.Broker code.
To be clear, the server environment still needs databroker.mongo_normalized, the Tiled Adapter for MongoDB with Bluesky document collections.
This is a backward-compatible change. Databroker now has a dependency on bluesky-tiled-plugins and has shim modules that expose the moved objects at the original locations within the databroker package.
Closes #812
I have registered the publish-pypi.yml workflow in this repository as a "pending" trusted publisher such that merge this PR should create the bluesky-tiled-plugins package on PyPI.
|
2025-04-01T06:38:04.587082
| 2024-09-20T22:21:34
|
2539766250
|
{
"authors": [
"prjemian"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4291",
"repo": "bluesky/ophyd",
"url": "https://github.com/bluesky/ophyd/pull/1211"
}
|
gharchive/pull-request
|
TYPO should be lower case to be consistent with others in this module
Somehow, this missed testing. A long time ago.
tetramm = TetrAMM("8idTetra:QUAD1:", name="tetramm")
for axis in "x y".split():
for attr_name in "offset offset_calc scale".split():
getattr(tetramm, f"position_{attr_name}_{axis}").kind = "config"
AttributeError: position_scale_y
We can get by without this change (which might break existing usage) with:
for attr_name in tetramm.component_names:
attr = getattr(tetramm, attr_name)
if attr_name.startswith("current_"):
for ch_name in attr.component_names:
getattr(attr, ch_name).kind = "config"
elif attr_name.startswith("position_"):
attr.kind = "config"
But for sure, all these components need to be kind="config".
|
2025-04-01T06:38:04.631494
| 2015-06-29T11:12:55
|
91775620
|
{
"authors": [
"bmhatfield",
"ealekseev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4293",
"repo": "bmhatfield/riemann-sumd",
"url": "https://github.com/bmhatfield/riemann-sumd/pull/22"
}
|
gharchive/pull-request
|
Change status code description from 'warn' to 'warning'.
For compatibility with alerta dashboard (http://alerta.io) this change is needed as status 'warn' it treats as 'unknown'.
http://nagios.sourceforge.net/docs/3_0/pluginapi.html also describes it as 'WARNING', not 'WARN'.
Thanks for doing this! I didn't realize I'd made this mistake :-)
|
2025-04-01T06:38:04.646662
| 2024-01-05T04:47:42
|
2066722477
|
{
"authors": [
"bn222",
"vrindle",
"wizhaoredhat"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4294",
"repo": "bn222/cluster-deployment-automation",
"url": "https://github.com/bn222/cluster-deployment-automation/pull/95"
}
|
gharchive/pull-request
|
Adds kind field and ability to create Microshift ISO
PR is complete:
generates an iso using os-builder
uses redfish api to boot it into a system
@bn222 this script is currently working consistently. I have tested it out. After investigation, I figured out that the issue I was running into was a space problem which is why the build was failing. PTAL. I will clean up the style checks and address your comments to get this into CDA. I have also included a small microshift.yml which I used to test my configuration on CDA.
Please make sure that the linter checks pass while I take a look.
@bn222 @wizhaoredhat PTAL I have addressed all the comments
A lot better than your previous drafts, Good Job. Please keep in mind modular/readable/reusable code!
A lot better than your previous drafts, Good Job. Please keep in mind modular/readable/reusable code! I would add more commentary to explain things better.
Added some more comments to my functions
@bn222 PTAL
I've tested the PR. It works but it has two more issues.
The cleanup uses composer-cli but it is installed later so there is a mistake in the ordering
Your code returns before the installation completes making it difficult to use this in automation since there is no way to tell when it's really done.
https://github.com/bn222/cluster-deployment-automation/blob/main/clustersConfig.py#L103
This is where I expect the kubeconfig to be
https://github.com/bn222/cluster-deployment-automation/blob/main/clustersConfig.py#L103
This is where I expect the kubeconfig to be
@vrindle @bn222
This is a bit more nuanced since it is possible to have multiple kubeconfigs. We might need to change it such that it is a folder of kubeconfigs or we can concatenate into one file
wdym multiple kubeconfigs? 1 kubeconfig per cluster, I would assume.
wdym multiple kubeconfigs? 1 kubeconfig per cluster, I would assume.
Current schema:
clusters:
- name: "microshift-cluster"
kind: "microshift"
kubeconfig: "/root/kubeconfig.whichclusterdoibelong"`
masters:
- name: "m1"
type: "physical"
node: "m1"
ip: "<IP_ADDRESS>"
bmc_ip: "<IP_ADDRESS>"
bmc_user: "root"
bmc_password: "calvin"
- name: "m2"
type: "physical"
node: "m2"
ip: "<IP_ADDRESS>"
bmc_ip: "<IP_ADDRESS>"
bmc_user: "root"
bmc_password: "calvin"
I expect:
clusters:
- name: "microshift-cluster"
kind: "microshift"
kubeconfig: "/root/kubeconfig.1"`
masters:
- name: "m1"
type: "physical"
node: "m1"
ip: "<IP_ADDRESS>"
bmc_ip: "<IP_ADDRESS>"
bmc_user: "root"
bmc_password: "calvin"
- name: "microshift-cluster"
kind: "microshift"
kubeconfig: "/root/kubeconfig.2"`
masters:
- name: "m1"
type: "physical"
node: "m1"
ip: "<IP_ADDRESS>"
bmc_ip: "<IP_ADDRESS>"
bmc_user: "root"
bmc_password: "calvin"
assert len(masters) == 1 or not microshift
I expect:
clusters:
- name: "microshift-cluster1"
kind: "microshift"
kubeconfig: "/root/kubeconfig.1"`
masters:
- name: "m1"
type: "physical"
node: "m1"
ip: "<IP_ADDRESS>"
bmc_ip: "<IP_ADDRESS>"
bmc_user: "root"
bmc_password: "calvin"
- name: "microshift-cluster2"
kind: "microshift"
kubeconfig: "/root/kubeconfig.2"`
masters:
- name: "m1"
type: "physical"
node: "m1"
ip: "<IP_ADDRESS>"
bmc_ip: "<IP_ADDRESS>"
bmc_user: "root"
bmc_password: "calvin"
assert len(masters) == 1 or not microshift
Ah I see, ok with that scheme then it should be ok.
assert len(masters) == 1 or not microshift
@wizhaoredhat @bn222 With the way that this code is currently implemented we check to see if kind microshift we want len(masters) == 1 so I believe that this should be in line with the schema that was presented.
|
2025-04-01T06:38:04.670581
| 2017-07-10T21:27:56
|
241856433
|
{
"authors": [
"boazsegev",
"cdkrot"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4295",
"repo": "boazsegev/facil.io",
"url": "https://github.com/boazsegev/facil.io/issues/14"
}
|
gharchive/issue
|
Possible bug in start/stop routines
Played with "echo" demo, as found on "Getting started" page (using stable branch with debug symbols).
All runs smoothly, however when I run this demo under valgrind and ^C it, the following errors are reported:
==7444== Memcheck, a memory error detector
==7444== Copyright (C) 2002-2015, and GNU GPL'd, by Julian Seward et al.
==7444== Using Valgrind-3.12.0.SVN and LibVEX; rerun with -h for copyright info
==7444== Command: ./a.out
==7444==
* Listening on port 3000
Server is running 1 worker X 1 thread, press ^C to stop
* 7444 is running.
^C==7444== Invalid read of size 4
==7444== at 0x12115D: defer_perform_in_fork (defer.c:394)
==7444== by 0x112DD6: facil_run (facil.c:1069)
==7444== by 0x1105A8: main (e2.c:53)
==7444== Address 0x5404cc0 is 0 bytes after a block of size 0 alloc'd
==7444== at 0x4C2DBC5: calloc (vg_replace_malloc.c:711)
==7444== by 0x121022: defer_perform_in_fork (defer.c:363)
==7444== by 0x112DD6: facil_run (facil.c:1069)
==7444== by 0x1105A8: main (e2.c:53)
==7444==
==7444== Invalid read of size 4
==7444== at 0x121196: defer_perform_in_fork (defer.c:397)
==7444== by 0x112DD6: facil_run (facil.c:1069)
==7444== by 0x1105A8: main (e2.c:53)
==7444== Address 0x5404cc0 is 0 bytes after a block of size 0 alloc'd
==7444== at 0x4C2DBC5: calloc (vg_replace_malloc.c:711)
==7444== by 0x121022: defer_perform_in_fork (defer.c:363)
==7444== by 0x112DD6: facil_run (facil.c:1069)
==7444== by 0x1105A8: main (e2.c:53)
==7444==
* 7444 cleanning up.
* (7444) Stopped listening on port 3000
--- Completed Shutdown ---
Probably there is an error there and you should have a look.
I have looked over the defer.c's defer_perform_in_fork, and for me it seems, that the main culprit is defer.c:382
pids_count++;
This makes the loop under 'finish:' to fall over the allocated memory, deleting this line fixes the valgrind error.
Hope this helps :)
Thanks 🙏🏻 🎉👍🏻👏🏻👏🏻
I re-read that piece of code and I have no idea what that line was doing there... 😂 I guess it was a leftover from a previous implementation.
Thank you very much for spotting it. 👍🏻
Thank you for your work :)
Out of curiosity, why do you use hand-written spinlocks, and not mutexes?
You're welcome :-)
why do you use hand-written spinlocks, and not mutexes?
It's a combination of performance testing and my own foolishness.
When I started the project, some of my design choices were a bit foolish and resulted in higher lock contention. Some of the lock contention issues couldn't be avoided (i.e. the defer library's queue access).
The original locking used mutexes and I noticed that mutexes were slower than I would like them to be. Something about the rescheduling mechanism was taking a long time to wake the threads up. Conditional variables weren't any better...
I tested it against spin locks using stdatomic.h . The spinlocks were noticeably faster for the defer queue (where contention can't be avoided).
But stdatomic.h wasn't available on Linux 14.04 (I was deploying on Heroku) so I just ended up writing the thing myself.
|
2025-04-01T06:38:04.673147
| 2024-08-18T17:00:01
|
2472060417
|
{
"authors": [
"CodeForHunger",
"bobbyiliev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4296",
"repo": "bobbyiliev/introduction-to-sql",
"url": "https://github.com/bobbyiliev/introduction-to-sql/pull/88"
}
|
gharchive/pull-request
|
Update README.md
Added new Section called Other Resources which contains SQL Tutorials and Interview questions Free resources...
Please let me know if any changes required.
Thank you!
Thank you but I think that this is not needed at the moment!
|
2025-04-01T06:38:04.704974
| 2015-04-24T20:48:29
|
70790805
|
{
"authors": [
"bmf",
"rwaldron"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4297",
"repo": "bocoup/johnny-five.io",
"url": "https://github.com/bocoup/johnny-five.io/issues/44"
}
|
gharchive/issue
|
Missing Getting Started with Johnny Five and JY MCU Bluetooth Serial Port Module
I was unable to locate this article https://github.com/rwaldron/johnny-five/wiki/Getting-Started-with-Johnny-Five-and-JY-MCU-Bluetooth-Serial-Port-Module or anything talking about the JY-MCU module. Not sure if this was by design. If not, how can I help to get it up there?
Also - Great job to everyone contributing to this project. This is an excellent resource and it's much appreciated.
It's here: http://johnny-five.io/api/
Currently, those links go back to the repo, but I plan on doing a formal migration in very near future
That works for me! Thank you for your help.
Awesome :)
|
2025-04-01T06:38:04.743477
| 2024-09-10T16:02:56
|
2516867844
|
{
"authors": [
"bogumilchilinski"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4298",
"repo": "bogumilchilinski/dynpy",
"url": "https://github.com/bogumilchilinski/dynpy/issues/909"
}
|
gharchive/issue
|
Scope and time scheduling of the meeting for overview of activities performed during development of dynamic systems (issue #908 related)
The goal is to schedule the meeting in Teams calendar on 2024.09.10 at 20:30.
Meeting scheduled.
|
2025-04-01T06:38:04.745277
| 2019-10-10T15:43:33
|
505356074
|
{
"authors": [
"flaschbier"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4299",
"repo": "bojzi/sound-machine-electron-guide",
"url": "https://github.com/bojzi/sound-machine-electron-guide/issues/13"
}
|
gharchive/issue
|
electron-prebuilt is deprecated
Following this blog: https://medium.com/developers-writing/building-a-desktop-application-with-electron-204203eeb658
$ npm install --save-dev electron-prebuilt
npm WARN deprecated<EMAIL_ADDRESS>electron-prebuilt has been renamed to electron. For more details, see http://electron.atom.io/blog/2016/08/16/npm-install-electron
@bojzi I will work around errors as long as I can and report issues here. If you are not interested in further issues regarding the blog or you are not going to fix them anyway, just let me know to save me that work. Thanks.
Anyone alive here around?
|
2025-04-01T06:38:04.791812
| 2018-02-27T03:44:50
|
300483613
|
{
"authors": [
"canavandl"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4300",
"repo": "bokeh/jupyterlab_bokeh",
"url": "https://github.com/bokeh/jupyterlab_bokeh/pull/22"
}
|
gharchive/pull-request
|
Philippjfr/register kernel 2
fixes #10
supercedes #21
I'm pointing out that this PR causes us to expose the Jupyter kernel to viewers of a push_notebook plot. This should be revisited in the future.
|
2025-04-01T06:38:04.840812
| 2022-12-28T22:48:54
|
1513219532
|
{
"authors": [
"Jengah",
"jhunschejones",
"tongueroo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4301",
"repo": "boltops-tools/jets",
"url": "https://github.com/boltops-tools/jets/issues/636"
}
|
gharchive/issue
|
Ruby 3.x support
Summary
Ruby 3.x has been out for more than 2 years now, so I was surprised when I went to run jets new today and couldn't because there's no support for modern Ruby versions. 😞
I understand priorities need to be balanced when maintaining a project of this size, but it feels like it's time to bump this one up the list a little if it isn't already on top 😊 Thanks for helping build and maintain this awesome library. I'm excited to see it continue to push Ruby forward in the serverless age.
Main reason Ruby 3 is not support yet is because AWS Lambda doesn’t support it yet. It’s a bummer.
RE: Thanks for helping build and maintain this awesome library. I'm excited to see it continue to push Ruby forward in the serverless age.
Thanks for the kind words.
Main reason Ruby 3 is not support yet is because AWS Lambda doesn’t support it yet. It’s a bummer.
RE: Thanks for helping build and maintain this awesome library. I'm excited to see it continue to push Ruby forward in the serverless age.
Thanks for the kind words.
Thanks for the quick response @tongueroo! Apologies, I did not realize lambda doesn't support 3.x yet 😅 Hopefully this closed issue can serve as an answer to future folks who have this same question at least. 🙏🏻
Seems like this could probably be re-opened with the recent news of container base image support of Ruby 3.2 in Lambda: https://github.com/aws/aws-lambda-base-images/issues/54#issuecomment-1486974882
@Jengah Got me excited. Sadly, its only for the lambda container base image. It's not yet rolled out to AWS Lambda itself officially. The AWS console still shows Ruby 2.7 only.
I checked 2 regions: us-east-1 and us-west-2. Seems like AWS is getting closer ready. Guessing they updated the lambda container image in preparation for it. Will look at adding 3.2 support once it's officially released.
Also, saw that AWS also released python 3.1 recently. So Ruby 3.2 hopefully is close 🤞
Sorry for the false excitement. As the comment mentions the AWS managed framework runtime should be available within 90 days of the announcement, with 2.7 support lasting 6 months past the General Availability release of the Ruby 3.2 runtime for some roadway to get migrated.
I was more pointing out that preliminary testing could be done using a the supported base container as a custom runtime, but the official managed runtime should be out in the next 60 days or so and may not be worth the effort.
Awesome. Appreciate the heads up!
Done in https://github.com/boltops-tools/jets/pull/654 🎉 Blog Post: Jets v4 Release: Ruby 3.2 Support
Looks like AWS is ghost-testing Ruby 3.2 support for AWS Lambda. Ruby 3.2 does not show up in the AWS console, but you're able to deploy with the Ruby 3.2 runtime via CloudFormation. Was able to deploy a jets v4 app with ruby 3.2 successfully without the use of custom runtime. 🎉
|
2025-04-01T06:38:04.845372
| 2024-11-11T19:46:54
|
2650241527
|
{
"authors": [
"boly38"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4302",
"repo": "boly38/botEnSky",
"url": "https://github.com/boly38/botEnSky/pull/105"
}
|
gharchive/pull-request
|
Fix #100 bird identify or plugin unexpected error must be reported
Fix #100 bird identify or plugin unexpected error must be reported
Coverage after merging identify-500-must-be-reported into main will be77.96%
Coverage Report for Changed FilesFileStmtsBranchesFuncsLinesUncovered Linessrc/services PluginsCommonService.js66.02%72%50%66.51%100–101, 110–123, 132–135, 154, 159, 161–164, 191, 196, 200–201, 207–209, 38–40, 46, 46–59, 63–66, 69–73, 76–81, 84–89, 92–99
|
2025-04-01T06:38:04.854965
| 2023-08-14T14:07:57
|
1849854811
|
{
"authors": [
"manifestori",
"puerco"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4303",
"repo": "bom-squad/protobom",
"url": "https://github.com/bom-squad/protobom/issues/71"
}
|
gharchive/issue
|
Support annotations / properties for components / sboms
As you know, CDX and SPDX use different to tag/label/annotate components.
SPDX uses Annotations for adding additional data to SPDX Element, SPDX2.x has Document and Package/snippet/file... level annotations. While SPDX3 uses flat Annotations list with ref to and SPDX element.
CDX 1.x has Property, which is a narrowed-down version of annotations introduced in CDX 1.5.
I've looked over past work for conversions, which yielded:
https://github.com/spdx/cdx2spdx/blob/08ec34f11b15dd747410d13c1fc4d11645a20a4b/src/main/java/org/spdx/cdx2spdx/CycloneSpdxConverter.java#L565C60-L565C60
As I figured, they used the opinionated way to convert Properties to Annotations in SPDX (this is a CDX2SPDX tool), forcing Properties as Annotations related to the package/file/document/etc.
Since we have the luxury of having it both ways, we should add Annotations to Nodes and the Bom(document); we should also structure annotation so it could easily convert from/to property.
When unserializing SPDX: we do not have "Properties," only annotations. We would need to support Node-level properties. The problem is that we won't have a way to know how to parse "key" "value" unless. We can default to "annotation1", "annotation2" etc., with the value from "statement."
When serializing SPDX: if we'd have a specific "key" [if we pulled it from CDX] we would encode it alongside the value into a statement ("key=value") and add protobom as the annotator.
When serializing CDX we have no issues when serializing to annotations (document level) and properties (from annotations with properties)
When we unserialize, we simply load document level as global annotation and properties (as node-level annotation).
Wdyt?
I have opened an issue for SPDX3 at https://github.com/spdx/spdx-3-model/issues/477. It appears that name could be a suitable source for "key", and storing it would enable lossless conversion from SPDX to CDX and vice versa.
SPDX->CDX->SPDX
annotation -> document annotation (cdx1.5 only) / node annotations as properties rebuilding the entire Element as properties list, using the Name prop value as "key" and statement as "value"
"properties": [
{
"name": "annotator",
"value": "annotator value"
},
{
"name": "comment",
"value": "comment value"
},
{
"name": "annotationType",
"value": "other"
},
{
"name": "annotationDate",
"value": "...."
}
]
Going back to SPDX would include rebuilding the annotation object from those props.
There is an edge case where multiple annotations refer to the same key. To handle this, property names need to be scoped and prefixed either by index (in SPDX 2.x) or by SPDXID in SPDX 3.0.
Ah this is a common hack that abuses both standards a little bit. Properties in CDX != Annotations in SPDX. There is really no equivalent between them, it was one of the points raised during discussions between the SPDX/CDX compatibility talks and CDX offer to add annotations to their spec.
My understanding is that annotations were devised as a way to comment on the document while properties are a way to capture additional data. So, in the purest sense, I would not store them in the same place, but since it is a hack that people use frequently I'd support that as an option at the conversion layer.
Ah awesome, reviewing the 1.5 spec annotations are now in:
https://cyclonedx.org/docs/1.5/json/#annotations_items_bom-ref
Yup, the issue with SPDX is the lack of key-value structure that can properly capture data. In the spdx-3 thread I referenced, Im in contact with maintainers about adding a new properties field to Annotation to avoid abusing statement value for kv store. That would make conversation easier going forward.
Basically, SPDX annotation could be translated to CDX1.5 annotations with the exception of signature, and ofc since CDX annotation does not contain properties and has a separate field for that. we could safely assume that ANY document-level annotation is either CDX Annotation or CDX Properties, making all other annotations (element annotation) convert into CDX component properties
Not sure why this is completed. reopening.
|
2025-04-01T06:38:04.899513
| 2021-12-13T14:25:31
|
1078595551
|
{
"authors": [
"corcoja",
"morarucostel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4306",
"repo": "boomerang-io/roadmap",
"url": "https://github.com/boomerang-io/roadmap/issues/300"
}
|
gharchive/issue
|
Update the documentation with the changes required for "Redesign the eventing library to support multiple Streams and Consumers"
Hey @corcoja , can you please take a look and update the necessary documentation part of the changes done in #255 ?
You can create a new branch from useboomerang.io repo for this change.
Hey @morarucostel! Do you want me to update the boomerang-io/lib.eventing repo docs, the documentation on useboomerang.io website or both?
Hey @corcoja , I believe both of them should be updated to reflect the latest.
The action with higher priority, imo is the useboomerang.io one.
|
2025-04-01T06:38:05.025288
| 2023-03-03T10:40:41
|
1608394308
|
{
"authors": [
"crftwrk",
"justinkruit",
"kreincke"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4307",
"repo": "bootscore/bootscore",
"url": "https://github.com/bootscore/bootscore/pull/414"
}
|
gharchive/pull-request
|
Switch from medium featured images to an own bootScore size
There seems to be a weakness concerning dealing with Featured Images. I've described the effect in two posts: https://fodina.de/image-d/ and https://fodina.de/image-e/. If you want to see the difference, use your desktop machine and try this
Go to https://fodina.de/ and look at the post list. The square featured images are sharp.
Insert the word 'Picture' into the search box and look at that list (delivered by the unmodified archive.php). On large Screens, everything blurs.
I am going to describe the technical solution in a longer statement, posted in discussions.
Personally I feel like this change is more up to the developer, instead of in the theme by default.
Also what I usually do is call a bigger size, like large, and make it square using the Bootstrap ratios. That way I have more control over the size if it needs to change depending on screen size.
possible. But the more large pictures you have to distribute over the network, the slower the site.
However, I think that such a valuable work as bootScore should not show blurred and distorted images by default. There's never a second chance for the first impression. Yes, the user could fix it himself, but why do we want to burden him with something that is very little effort on our side: it costs only 1 line more, and 4 changed values to improve it.
As @justinkruit said, image sizes should be the developers job. Because we follow straight the WordPress standard and changing default WP image sizes will produce more questions than answers like here https://github.com/bootscore/bootscore/discussions/399.
However, I understand the thing with the super large screens. But:
Bootstrap is mobile first, means designed for smaller screens.
thumbnail or medium are the default WP preview image sizes .
Changing to larger sizes will load larger images on mobile as well, bad for SEO.
XXL screens have often a larger screen resolution. Means the image is larger than the original size is. Of course, it's not sharp anymore.
What you can do:
Use . get_the_post_thumbnail(null, 'large') . instead of medium.
Shorten the excerpt text or write a custom excerpt to lower card (and image) height.
Use a different loop template https://github.com/bootscore/bs-loop-templates which not crop the featured preview image. For example this one https://bootscore.me/archives/equal-height-sidebar-right/.
Many thanks for your answers. Unfortunately, I cannot agree with you in this case. A tool creating bad optical effects, by default, will perhaps lose its advantage: Again you never get a second chance to make a first impression. If you have templates, which can better deal with the default image sizes of WordPress, then bootScore should perhaps take one of them as 'index.php'
However, my opinion is nothing more than my opinion. Hence, if you don't like my solution, there's no need to integrate this PR. And yes, I will test the other bs-loop-templates later on.
with best regards KR
We can generally rethink and improve list view in loop. But let's do this carefully and later.
I've just summarized our discussion: https://fodina.de/image-i/ Fell free to contact me, if you think the post contains unfair remarks. I wanted to describe the case, not to blame anyone. Again and again and again: I appreciate your work very much! KR
|
2025-04-01T06:38:05.027859
| 2017-03-20T06:26:15
|
215338496
|
{
"authors": [
"krtschmr"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4308",
"repo": "bootstrap-ruby/rails-bootstrap-forms",
"url": "https://github.com/bootstrap-ruby/rails-bootstrap-forms/issues/319"
}
|
gharchive/issue
|
using helper inside the form-builder
undefined method `team?' for #BootstrapForm::FormBuilder:0x007fe9c2c91590
BootstrapForm::FormBuilder.class_eval do
if manager_namespace?
end
if team_namespace?
end
end
apparently those namespace_methods aren't known.
they actually are defined in the BaseController for that namespace which is pretty much the ApplicationController for our Backoffice
class Manager::BaseController < ActionController::Base
def team_namespace?
#. ...
helper_method :team_namespace?
end
anybody know how to get our own helpers included?
ok, that was easy fix
include NamespaceHelper and then define the methods inside that module
|
2025-04-01T06:38:05.058089
| 2022-06-30T02:56:43
|
1289513311
|
{
"authors": [
"DaveVdE",
"borisdj",
"proc01",
"tincann"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4310",
"repo": "borisdj/EFCore.BulkExtensions",
"url": "https://github.com/borisdj/EFCore.BulkExtensions/issues/867"
}
|
gharchive/issue
|
System.InvalidOperationException: 'When 'UseTempDB' is set then BulkOperation has to be inside Transaction. Otherwise destination table gets dropped too early because transaction ends before operation is finished.'
System.InvalidOperationException: 'When 'UseTempDB' is set then BulkOperation has to be inside Transaction. Otherwise destination table gets dropped too early because transaction ends before operation is finished.'
Step to reproduce
try
{
using (TransactionScope scope = new TransactionScope(TransactionScopeAsyncFlowOption.Enabled))
{
var core = new ZZZContext();
...
await core.AuvyBulkUpdateAsync(objects);
scope.Complete();
}
}
catch (Exception)
{
throw;
}
Take a look at Test:
https://github.com/borisdj/EFCore.BulkExtensions/blob/4bb9fd3d4b28ad2c43009d2469b7758557692fd5/EFCore.BulkExtensions.Tests/EFCoreBulkTest.cs#L325-L334
In my opinion, this is a bug.
The use of a System.Transaction will allow the database provider to create a transaction automatically, so the check that's built into TableInfo is not correct.
I've solved this by opening the database connection explicitly before using BulkUpdateAsync.
In my opinion, this is a bug.
The use of a System.Transaction will allow the database provider to create a transaction automatically, so the check that's built into TableInfo is not correct.
I've solved this by opening the database connection explicitly before using BulkUpdateAsync.
@DaveVdE Thank you! Our code was also using an ambient transaction in combination with UseTempDb and SetOutputIdentity, which confusingly kept throwing the "When 'UseTempDB' is set then BulkOperation has to be inside Transaction" exception. Opening the db connection before calling the bulk operation solved my issue.
@borisdj Looking at the exception condition, would it make sense to add || Transaction.Current != null to this expression? https://github.com/borisdj/EFCore.BulkExtensions/blob/03934e648db21b790f903b008d0deafb0a71cd1e/EFCore.BulkExtensions/TableInfo.cs#L126C21-L126C21
|
2025-04-01T06:38:05.074546
| 2015-05-01T18:06:37
|
72486057
|
{
"authors": [
"Kelvin-Ng",
"aciccarello",
"chbrown",
"drudru",
"hsrobmln",
"lucasmciruzzi",
"omencat",
"sethx",
"vote539",
"vvakame"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4312",
"repo": "borisyankov/DefinitelyTyped",
"url": "https://github.com/borisyankov/DefinitelyTyped/issues/4249"
}
|
gharchive/issue
|
node.d.ts: interface Global / NodeJS.Global breaks current Node+TypeScript
TypeScript complains of missing names in the new interface Global type, from 5235188 (5 days ago).
tsc --module commonjs --target ES5 --declaration my_incredible_new_mvc_framework.ts
node/node.d.ts(198,26): error TS2304: Cannot find name 'DataView'.
node/node.d.ts(212,21): error TS2304: Cannot find name 'Map'.
node/node.d.ts(221,21): error TS2304: Cannot find name 'Set'.
node/node.d.ts(231,25): error TS2304: Cannot find name 'WeakMap'.
Those are all ES6 types. The current Node.js / v8 version implements them, but they aren't native types, as far as the TypeScript compiler is concerned. Perhaps we are getting ahead of ourselves here?
My system:
$ node --version
v0.12.2
$ tsc --version
message TS6029: Version 1.5.0-beta
@chbrown please watching this PR. https://github.com/borisyankov/DefinitelyTyped/pull/4101
Is this fixed or there is a temporary fix to implement?
I'm still getting those same errors:
node.d.ts(198,26): error TS2304: Cannot find name 'DataView'.
node.d.ts(212,21): error TS2304: Cannot find name 'Map'.
node.d.ts(221,21): error TS2304: Cannot find name 'Set'.
node.d.ts(231,25): error TS2304: Cannot find name 'WeakMap'.
@lucasmciruzzi I simply reverted the breaking commit in my DefinitelyTyped fork and it's been working great: https://github.com/chbrown/DefinitelyTyped/commit/17f99f1
Thanks for the reply @chbrown! I'm using DTSM so I don't get updates from forks, I only get them from the borisyankov repo. Shouldn't this fix be on the main repo?
@vvakame thanks for the headsup on https://github.com/borisyankov/DefinitelyTyped/pull/4101 .
with typescript 1.5.2 ES5 target output resolves the DataView error. Map, Set, WeakMap are still present, but compile appears to work fine.
+1 to merge changes from chbrown to borisyankof repo.
@chbrown Thank you for the quick fix, this solved my issue and works for now!
I updated the SHA in my tsd.json file to f0aa5507070dc74859b636bd2dac37f3e8cab8d1 and ran tsd reinstall -o to revert back to the prior version.
Thanks @chbrown - I switched to your repo until this gets fixed.
Why is this fix not merged with the repo yet? @chbrown @vvakame
master/HEAD are solved this issue.
I'm still using the chbrown@17f99f1 version because the borisyankov version still trows errors with "Map", "Set" and "WeakMap" when target is ES5 ....
@lucasmciruzzi no harm in that, but after @vvakame's comment, I tried out node.d.ts from DefinitelyTyped/master, which, combined with the<EMAIL_ADDRESS>that was released ~5 days ago, compiles just fine.
Are you getting those errors even with the latest tsc?
Yep, I'm running TypeScript v1.5.3
Fixed! ... it seems like it was a problem with the packages cache on my PC. Thanks a lot :)
I don't get any error when I run tsc in command line directly. However, I do get error TS2304: Cannot find name 'DataView'. when I use grunt. GruntFile.js:
...
typescript: {
base: {
src: ['src/**/*.ts'],
dest: 'build',
options: {
module: 'commonjs',
target: 'es5'
}
}
}
...
Using https://github.com/borisyankov/DefinitelyTyped/commit/f0aa5507070dc74859b636bd2dac37f3e8cab8d1 as stated by @aciccarello works.
TypeScript version is 1.5.3. Any ideas?
@Kelvin-Ng it should be an issue with the grunt plugin. It surely uses its own TypeScript instead of the globally installed. Check the settings of the plugin you are using for an option to set the TypeScript instance to use.
PS: If you can, move away from grunt to gulp, the plugin is far better: https://www.npmjs.com/package/gulp-typescript and has a "typescript" option to set a custom version of it.
@lucasmciruzzi
Thank you! I have moved to gulp and gulp-typescript works great. Thanks for your recommendation and I like gulp more than grunt now.
As of yesterday this is now fixed in grunt-typescript.
https://github.com/k-maru/grunt-typescript/issues/105
|
2025-04-01T06:38:05.076012
| 2015-07-25T00:51:56
|
97176368
|
{
"authors": [
"chrootsu",
"vvakame"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4313",
"repo": "borisyankov/DefinitelyTyped",
"url": "https://github.com/borisyankov/DefinitelyTyped/pull/5071"
}
|
gharchive/pull-request
|
lodash: added _.toPlainObject() method
https://lodash.com/docs#toPlainObject
@chrootsu thanks mate!
|
2025-04-01T06:38:05.078328
| 2015-09-24T00:17:04
|
108032268
|
{
"authors": [
"marknadig",
"stephenlautier",
"vvakame"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4314",
"repo": "borisyankov/DefinitelyTyped",
"url": "https://github.com/borisyankov/DefinitelyTyped/pull/5983"
}
|
gharchive/pull-request
|
ng-dialog fix IDialogOpenResult.closePromise
create IDialogClosePromise type to be returned by IDialogOpenResult.closePromise.
ng-dialog/ng-dialog.d.ts
to author(@stephenlautier). could you review this PR?
:+1: or :-1:?
check list
[ ] pass the Travic-CI test?
Looks good :+1:. Thanks @marknadig
@vvakame looks good.
thanks mate!
|
2025-04-01T06:38:05.082727
| 2019-01-04T15:33:21
|
395966645
|
{
"authors": [
"matt-adigital",
"roelvanhintum"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4315",
"repo": "born05/craft-twofactorauthentication",
"url": "https://github.com/born05/craft-twofactorauthentication/issues/22"
}
|
gharchive/issue
|
Stable Release Date
Not really an issue, more of a question, but any ideas of when a stable release might happen for the Craft 3 version?
We want to use this for a client of ours but aren't too keen on using software which is still in beta as we'd like to avoid as many bugs as possible. Looks like a great plugin though, good work!
@matt-adigital Thanks! We're doing some in depth security tests in the coming week. Once those clear, i'll release it as stable. Shouldn't take more than two weeks.
|
2025-04-01T06:38:05.090731
| 2023-09-05T12:50:26
|
1881916967
|
{
"authors": [
"MattiJarvinen",
"jorgealgaba",
"robertIsaac"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4316",
"repo": "borremosch/cobertura-merge",
"url": "https://github.com/borremosch/cobertura-merge/issues/40"
}
|
gharchive/issue
|
can't install in windows
in windows it try for some reason to locate python package then fail
full error
yarn install cobertura-merge
yarn install v1.22.19
error `install` has been replaced with `add` to add new dependencies. Run "yarn add cobertura-merge" instead.
info Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command.
PS C:\Users\aa.risaac\repos\Dashboard> yarn add cobertura-merge -D
yarn add v1.22.19
[1/4] Resolving packages...
warning cobertura-merge > xml2json ><EMAIL_ADDRESS>This version has been deprecated in accordance with the hapi support policy (hapi.im/support). Please upgrade to the latest version to get the best features, bug fixes, and securi
ty patches. If you are unable to upgrade at this time, paid support is available for older versions (hapi.im/commercial).
warning cobertura-merge > xml2json ><EMAIL_ADDRESS>This version has been deprecated in accordance with the hapi support policy (hapi.im/support). Please upgrade to the latest version to get the best features, bug fixes, and securi
ty patches. If you are unable to upgrade at this time, paid support is available for older versions (hapi.im/commercial).
warning cobertura-merge > xml2json > joi ><EMAIL_ADDRESS>This version has been deprecated in accordance with the hapi support policy (hapi.im/support). Please upgrade to the latest version to get the best features, bug fixes, and
security patches. If you are unable to upgrade at this time, paid support is available for older versions (hapi.im/commercial).
warning cobertura-merge > xml2json > joi ><EMAIL_ADDRESS>This module has moved and is now available at @hapi/topo. Please update your dependencies as this version is no longer maintained an may contain bugs and security issues.
warning cobertura-merge > xml2json > joi > topo ><EMAIL_ADDRESS>This module has moved and is now available at @hapi/hoek. Please update your dependencies as this version is no longer maintained an may contain bugs and security iss
ues.
[2/4] Fetching packages...
warning Pattern ["object-assign@latest"] is trying to unpack in the same destination "C:\\Users\\aa.risaac\\AppData\\Local\\Yarn\\Cache\\v6\\npm-object-assign-4.1.1-2109adc7965887cfc05cbbd442cac8bfbb360863-integrity\\node_module
s\\object-assign" as pattern ["object-assign@^4","object-assign@^4.0.1","object-assign<EMAIL_ADDRESS>This could result in non-deterministic behavior, skipping.
[3/4] Linking dependencies...
warning " ><EMAIL_ADDRESS>has incorrect peer dependency "<EMAIL_ADDRESS><=13.0.0 || ^13.0.0".
warning " ><EMAIL_ADDRESS>has incorrect peer dependency "<EMAIL_ADDRESS><=13.0.0 || ^13.0.0".
warning "@angular-devkit/build-angular ><EMAIL_ADDRESS>has incorrect peer dependency<EMAIL_ADDRESS>warning " ><EMAIL_ADDRESS>has incorrect peer dependency "@angular/cli@>= 13.0.0 < 14.0.0".
warning " ><EMAIL_ADDRESS>has unmet peer dependency "<EMAIL_ADDRESS>warning " ><EMAIL_ADDRESS>has unmet peer dependency "<EMAIL_ADDRESS>warning " ><EMAIL_ADDRESS>has unmet peer dependency<EMAIL_ADDRESS>warning " ><EMAIL_ADDRESS>has unmet peer dependency<EMAIL_ADDRESS>warning " ><EMAIL_ADDRESS>has unmet peer dependency "mocha@>=7".
[4/4] Building fresh packages...
error C:\Users\aa.risaac\repos\Dashboard\node_modules\node-expat: Command failed.
Exit code: 1
Command: node-gyp rebuild
Arguments:
Directory: C:\Users\aa.risaac\repos\Dashboard\node_modules\node-expat
Output:
C:\Users\aa.risaac\repos\Dashboard\node_modules\node-expat>if not defined npm_config_node_gyp (node "C:\Program Files\nodejs\node_modules\npm\bin\node-gyp-bin\\..\..\node_modules\node-gyp\bin\node-gyp.js" rebuild ) else (node "
" rebuild )
gyp info it worked if it ends with ok
gyp info using<EMAIL_ADDRESS>gyp info using<EMAIL_ADDRESS>| win32 | x64
gyp ERR! find Python
gyp ERR! find Python Python is not set from command line or npm configuration
gyp ERR! find Python Python is not set from environment variable PYTHON
gyp ERR! find Python checking if "python3" can be used
gyp ERR! find Python - "python3" is not in PATH or produced an error
gyp ERR! find Python checking if "python" can be used
gyp ERR! find Python - "python" is not in PATH or produced an error
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python39\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python39\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python39\python.exe
gyp ERR! find Python - "C:\Program Files\Python39\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python39-32\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python39-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python39-32\python.exe
gyp ERR! find Python - "C:\Program Files\Python39-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files (x86)\Python39-32\python.exe
gyp ERR! find Python - "C:\Program Files (x86)\Python39-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python38\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python38\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python38\python.exe
gyp ERR! find Python - "C:\Program Files\Python38\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python38-32\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python38-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python38-32\python.exe
gyp ERR! find Python - "C:\Program Files\Python38-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files (x86)\Python38-32\python.exe
gyp ERR! find Python - "C:\Program Files (x86)\Python38-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python37\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python37\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python37\python.exe
gyp ERR! find Python - "C:\Program Files\Python37\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python37-32\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python37-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python37-32\python.exe
gyp ERR! find Python - "C:\Program Files\Python37-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files (x86)\Python37-32\python.exe
gyp ERR! find Python - "C:\Program Files (x86)\Python37-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python36\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python36\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python36\python.exe
gyp ERR! find Python - "C:\Program Files\Python36\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Users\aa.risaac\AppData\Local\Programs\Python\Python36-32\python.exe
gyp ERR! find Python - "C:\Users\aa.risaac\AppData\Local\Programs\Python\Python36-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files\Python36-32\python.exe
gyp ERR! find Python - "C:\Program Files\Python36-32\python.exe" could not be run
gyp ERR! find Python checking if Python is C:\Program Files (x86)\Python36-32\python.exe
gyp ERR! find Python - "C:\Program Files (x86)\Python36-32\python.exe" could not be run
gyp ERR! find Python checking if the py launcher can be used to find Python 3
gyp ERR! find Python - "py.exe" is not in PATH or produced an error
gyp ERR! find Python
gyp ERR! find Python **********************************************************
gyp ERR! find Python You need to install the latest version of Python.
gyp ERR! find Python Node-gyp should be able to find and use Python. If not,
gyp ERR! find Python you can try one of the following options:
gyp ERR! find Python - Use the switch --python="C:\Path\To\python.exe"
gyp ERR! find Python (accepted by both node-gyp and npm)
gyp ERR! find Python - Set the environment variable PYTHON
gyp ERR! find Python - Set the npm configuration variable python:
gyp ERR! find Python npm config set python "C:\Path\To\python.exe"
gyp ERR! find Python For more information consult the documentation at:
gyp ERR! find Python https://github.com/nodejs/node-gyp#installation
gyp ERR! find Python **********************************************************
gyp ERR! find Python
gyp ERR! configure error
gyp ERR! stack Error: Could not find any Python installation to use
gyp ERR! stack at PythonFinder.fail (C:\Users\aa.risaac\AppData\Roaming\nvm\v18.16.0\node_modules\npm\node_modules\node-gyp\lib\find-python.js:330:47)
gyp ERR! stack at PythonFinder.runChecks (C:\Users\aa.risaac\AppData\Roaming\nvm\v18.16.0\node_modules\npm\node_modules\node-gyp\lib\find-python.js:159:21)
gyp ERR! stack at PythonFinder.<anonymous> (C:\Users\aa.risaac\AppData\Roaming\nvm\v18.16.0\node_modules\npm\node_modules\node-gyp\lib\find-python.js:228:18)
gyp ERR! stack at PythonFinder.execFileCallback (C:\Users\aa.risaac\AppData\Roaming\nvm\v18.16.0\node_modules\npm\node_modules\node-gyp\lib\find-python.js:294:16)
gyp ERR! stack at exithandler (node:child_process:427:5)
gyp ERR! stack at ChildProcess.errorhandler (node:child_process:439:5)
gyp ERR! stack at ChildProcess.emit (node:events:513:28)
gyp ERR! stack at ChildProcess._handle.onexit (node:internal/child_process:289:12)
gyp ERR! stack at onErrorNT (node:internal/child_process:476:16)
gyp ERR! stack at process.processTicksAndRejections (node:internal/process/task_queues:82:21)
gyp ERR! System Windows_NT 10.0.19044
gyp ERR! command "C:\\Program Files\\nodejs\\node.exe" "C:\\Program Files\\nodejs\\node_modules\\npm\\node_modules\\node-gyp\\bin\\node-gyp.js" "rebuild"
gyp ERR! cwd C:\Users\aa.risaac\repos\Dashboard\node_modules\node-expat
gyp ERR! node -v v18.16.0
gyp ERR! node-gyp -v v9.3.1
gyp ERR! not ok
info Visit https://yarnpkg.com/en/docs/cli/add for documentation about this command.
You need to read https://github.com/nodejs/node-gyp#on-windows
This is how nodejs handles building native binaries.
You need to read https://github.com/nodejs/node-gyp#on-windows
This is how nodejs handles building native binaries.
can you publish binary like other libraries usually do so we don't have to
Hi @robertIsaac, you need to install python and have it on your path to fix it :)
Works after installing python
|
2025-04-01T06:38:05.101650
| 2019-09-05T03:01:07
|
489501864
|
{
"authors": [
"AndrejMitrovic",
"Geod24"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4317",
"repo": "bosagora/agora",
"url": "https://github.com/bosagora/agora/issues/300"
}
|
gharchive/issue
|
Make the UTXO set re-entrant
This also needs to make sure it's memory-efficient. A simple (but naive) approach is to use a new Set for each time we return a utxoFinder delegate. But this would cause a lot of allocations.
Related to #194
Moving this out of Full Node and into Validator and removing it from the current kanban, because it's something we can think about doing at a later point. The re-entrancy might not be an issue right now since we use fibers, and we don't yield anywhere in the utxoFinder delegate. In theory it should just work.
This hasn't been needed for CoinNet as well, so moving out of the milestone.
|
2025-04-01T06:38:05.108372
| 2023-01-23T01:30:00
|
1552379118
|
{
"authors": [
"bosley"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4318",
"repo": "bosley/sauros",
"url": "https://github.com/bosley/sauros/issues/98"
}
|
gharchive/issue
|
import boxing
Right now imports dump everything into the global space. Perhaps we should change this to create a box and then stuff everything in that box kind of like "use" does.
This would give the advantage of sectioning off everything imported.
Need to keep in mind that we don't import things multiple times right now. Need to determine a way to cache imported boxes so upon a second import we can just give a shared pointer to the importer with access to the box in their own env
We already have a caching mechanism ... it IS env.... just do this. it will make sense
Realistically we should merge use and import into one command. Using some FS stuff we can determine if we will import a file or a directory. Then we can ensure that the directory is a pkg or not. Single files would be loaded just like the source_files in pkgs.
This is a great idea.
After deep investigations I've decided that this is not behavior that I want to create at this point in time
|
2025-04-01T06:38:05.109312
| 2021-04-16T07:32:37
|
859549296
|
{
"authors": [
"TonyRahme",
"haehn"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4319",
"repo": "bostongfx/cs480student",
"url": "https://github.com/bostongfx/cs480student/pull/142"
}
|
gharchive/pull-request
|
Assignment 5
Forgot to upload to github and pull request it.
I did not have enough time to do bonus.
100!
|
2025-04-01T06:38:05.112094
| 2015-07-20T08:51:00
|
96017530
|
{
"authors": [
"k-bx",
"kylebrandt"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4320",
"repo": "bosun-monitor/bosun",
"url": "https://github.com/bosun-monitor/bosun/issues/1180"
}
|
gharchive/issue
|
"Shouldn't be here" error upon query
My metric linux.net.sockets.used for one of hosts is shown fine for 3h-long data, but when I choose 4h interval I get an error:
opentsdb: {"start":"4h-ago","queries":[{"aggregator":"sum","metric":"linux.net.sockets.used","rateOptions":{},"tags":{"host":"app3"}}]}: Shouldn't be here
What is the meaning of this?
Thank you.
Haven't seen this one in a long time. IIRC it is what OpenTSDB returns when you query something that had duplicated datapoints (Same tagset, Same Timestamp, different values.
If this is the case, you have two options: Ensure there are no duplicates (you will only find out at query time, not write time) or enable tsd.storage.fix_duplicates in the OpenTSDB config.
If it is something else it is an error returned by OpenTSDB, so you that would be the place for further troubleshooting / bug filing.
Thanks, closing for now then, will reopen if I'll see the error again and these won't help.
@kylebrandt I think 2 instances of scollector were running on a problematic host
|
2025-04-01T06:38:05.129676
| 2015-04-24T20:31:36
|
70786909
|
{
"authors": [
"coveralls",
"jamesls",
"kyleknap"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4321",
"repo": "boto/botocore",
"url": "https://github.com/boto/botocore/pull/527"
}
|
gharchive/pull-request
|
Update and simplify botocore loaders to new dir layout
This reworks and greatly simplifies the loaders used in botocore.
This is part paying off technical debt as well as standardizing
on a shared common format.
In terms of the new layout, the loaders docstring discusses how it
looks like, but we now support the directory layout used by the
other AWS SDKs.
In terms of the loaders themselves:
Remove any fuzzy API version matching, you either can specify an
exact API version or you can get the latest API version.
Move up the service type names ('service-2', 'paginators-1', etc)
out of the loader. The session/client objects should know about
those details, but not the loader.
Remove all file extensions from the loader class. This actually
allows pluggable loader types besides JSON. This wasn't
actually possible previously.
cc @kyleknap
Coverage decreased (-0.62%) to 93.84% when pulling 68be91d8480a1d83fe0a92db7692a86ff6ecd951 on jamesls:simple-loaders into 60e72b22bdebfa83e9b964aca232c7cad572212e on boto:develop.
There's actually a few changes I'm going to need to make based on updating boto3, specifically with list_available_services and determine_latest_version.
The problem is that both of these assume that the directory structure is sufficient to extract meaningful data about what is or isn't available (in fact, I even comment on this in this PR: https://github.com/boto/botocore/pull/527/files#diff-f8c1b99f0b4538cf77c2052b498c9630R219).
We actually need to be able to specify the type name (services-2, paginators-1, resources-1) when asking about the latest available and the list of all services. For example, the latest available service for services-2 may not necessarily mean that there's a resources-1 model for that same API version. Similarly listing all the available resources may differ we only care about resources (i.e boto3).
A concrete example of breakage is doing loader.load_service_model('ec2', 'resources-1') in boto3. It will try to find the latest API version for EC2 and assuming it can load a resource for it, which it currently cannot so it will raise an exception.
To fix this I'll be adding these changes:
Modify determine_latest_version and list_available_services to require a type_name param, just like load_service_model.
Update the file loader interface (currently only JSONFileLoader is implemented) to add an exists() method that will return True/False if the resource exists.
@kyleknap I've updated with the necessary changes for boto3 as well as the latest model updates from develop. Should be ready now for another look. Thanks.
Coverage decreased (-0.06%) to 94.39% when pulling 12cc84bcc717b0d4cfd1302b99d8975526e30ccb on jamesls:simple-loaders into ad1d6e298fd46c9b5994d1a073d40bb1ca838726 on boto:develop.
https://github.com/boto/boto3/pull/104 is an example of how this new loader API can be used. Now boto3 simply needs to add its internal directory to the search path and then specify the type name when dealing with services that must contain a resource definition.
Coverage decreased (-1.06%) to 93.4% when pulling 56d4f48b718489cbea5ba89abd680067770bebe0 on jamesls:simple-loaders into ad1d6e298fd46c9b5994d1a073d40bb1ca838726 on boto:develop.
It looks good. Looks like you updated the code after my pep8 comment through the pr-check script. Had a couple of comments about tests and docstrings, but nothing too significant :ship:
Looks like you also need to rebase on develop and pull in the latest model changes, which should only be dynamodb (not sure if you rebased after my ec2 pr).
The PR should have been rebased on develop as of the latest release a few hours ago. I'll double check this is the case.
Hmm the merge button is grey right now. Usually means that I need to rebase.
Probably because I merged in the _retry bugfix PR.
@kyleknap pushed a commit to extract out all the test boilerplate. I think the tests are much easier to read now.
Also rebased against develop. Looks like it's because of the _retry PR I merged, which makes sense. Any changes to anything in botocore/data/ in develop will require a rebase to merge cleanly:
$ git rebase develop
First, rewinding head to replay your work on top of it...
Applying: Update and simplify botocore loaders to new dir layout
Applying: Move services to standard directory structure
Using index info to reconstruct a base tree...
M botocore/data/aws/_retry.json
Falling back to patching base and 3-way merge...
Applying: Fix pr-check issues
Applying: Fix docstring
Applying: Move out boiler plate for dir patching to helper method
Recent commits look good. Those tests look much cleaner. :ship: again.
Coverage increased (+0.06%) to 94.52% when pulling f5d2c33311fd99c613d2ee30c001501000c88ae1 on jamesls:simple-loaders into 5eb99b9e09702057ba5772d86af2d7751d75a459 on boto:develop.
|
2025-04-01T06:38:05.239303
| 2017-11-12T01:22:11
|
273190463
|
{
"authors": [
"bartvde",
"brambow"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4322",
"repo": "boundlessgeo/sdk",
"url": "https://github.com/boundlessgeo/sdk/issues/734"
}
|
gharchive/issue
|
LayerList layers prop behavior
Say I have two layers: an OSM basemap, and a geojson point layer. I've given each of these layers a metadata property called 'isBasemap', and set it to true for the OSM layer and false for the geojson.
I'm trying to exclude the basemap from the layer list component by passing an array of objects with the following shape:
[{id: 'layer_id}]
and only populating that array with layers where isBasemap is false.
However, passing that array to the LayerList component as the layers prop doesn't seem to work, as I'm still seeing the OSM layer in the layer list.
I believe the issue is because the layers prop in the LayerList component is being mapped to the Redux store: state.map.layers. Unless I'm missing something, this prevents using a custom list of layers to show in the LayerList component.
Good point, we recently introduced a metadata prop for hiding layers from the layerlist, bnd:hide-layerlist so in your case you could set those to true for the layers you do not want in the layerlist.
id: 'my_layer_id',
metadata: {
'bnd:hide-layerlist': true
}
would that work for your use case?
I don't think it is already in a release though.
thanks for the feedback, that is good to hear
@brambow we just released 2.2.0 now https://github.com/boundlessgeo/sdk/releases/tag/v2.2.0 please check it out at your convenience and let us know if the new metadata works for you
It works. Thanks!
|
2025-04-01T06:38:05.271207
| 2024-01-30T23:00:44
|
2108951694
|
{
"authors": [
"bouzidanas",
"calvinchai",
"marcinsoftem"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4323",
"repo": "bouzidanas/streamlit-code-editor",
"url": "https://github.com/bouzidanas/streamlit-code-editor/issues/12"
}
|
gharchive/issue
|
Callback for change
I read through the documentation and found there is no (simple) way to let the editor save user input automatically. Is it possible to have a callback function such as on_change which exists now in all streamlit official widgets?
I agree.
BTW is there an simple way to submit content by using an external button?
import streamlit as st
import code_editor
output = code_editor.code_editor("print('Hello world!')")
if st.button("Show output"):
st.write(output)
I have heard a few similar requests.
To be honest, I have been avoiding this issue simply because it seems like it all leads to the same solution which comes with an issue that I think will make the behavior of the editor feel more janky. However, Streamlit now supports partial reruns so maybe now is the time to add the debounced update feature.
Let me give some background that might illuminate why this feature is missing. One of the things I didnt like about the other ace editor component when I tried using it (after learning about it) is that it would rerun the app/script after almost every keypress. And I think you had to set a prop to avoid this. The frequent refreshes made the whole experience so bad that I saw it as a bad thing as it promoted bad UI experiences.
In order for the streamlit script to get data from the code component, the script had to be rerun with this new data. The same would be true with a callback function. But with fragments, the code editor and dependent elements can be the only thing that is rerun and that can perhaps minimize some of the jerkiness.
I think I will add a prop that will allow for a debounced response which will enable auto-updating the dictionary with current contents. I think I will also try and provide an option for updating when the editor loses focus which might prove to be a great compromise.
Thanks for the feedback!
I have some good news to share. I got this feature up and running in version 0.1.4
See here for more: https://discuss.streamlit.io/t/new-component-streamlit-code-editor-a-react-ace-code-editor-customized-to-fit-with-streamlit-with-some-extra-goodies-added-on-top/42868/16
Works great! Kudos for the great job and amazing response time.
and amazing response time.
You caught me in a moment where I had some time and I already knew how to implement the feature cause I did it before a while ago. So the stars aligned lol. Unfortunately, its been a month long wait for @calvinchai
Thank you for the great work and timely response!
|
2025-04-01T06:38:05.308752
| 2023-10-20T18:14:22
|
1954810873
|
{
"authors": [
"bowen-xu",
"maxeeem"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4324",
"repo": "bowen-xu/PyNARS",
"url": "https://github.com/bowen-xu/PyNARS/pull/38"
}
|
gharchive/pull-request
|
use truth in hash function of Sentence
@bowen-xu I looked at the code and it seems the core of the issue is the hash of Task not taking truth value into account. This simple change achieves what you described in the issue but I wonder if it could have some unintended consequences? Can you think of the situation when elsewhere in the code we would want two sentences to compare same regardless of their truth values? Then we could try a different fix.
@maxeeem I found there's something "magic" by adding self.truth to the tuple.
If two input tasks have the different truth-values,for sure, the hash-value are different. Ex.
A. %1.0;0.9%
A. %1.0;0.8%
However, I thought if the two truth-values are the same, for example, input two sentences
A. %1.0;0.9%
A. %1.0;0.9%
The hash-values of the tasks should have been the same, as a result, the beliefs table would have still contains only one item. But in fact, it works; the beliefs table contains the two statements with the same truth-value.
I didn't think of that before. I guess this is because the python object of the first truth-value %1.0; 0.9% and that of the second have different ids. When computing the hash-value, it takes the id of an object into account.
Can we record the premises of a task, can use them when computing the hash value?
For example, we can store its (direct) premises into a list if it is derived via inference, so the length of the list is usually 2. If it is an input task, then the premises list is empty. The hash-value is influenced by the input_id of the task, as well as the hash-values of the premises.
Another solution might be adding all elements of the evidential base into that tuple when computing hash-value, though this approach might be more costly.
Can we record the premises of a task and use them when computing the hash-value?
For example, we can store its (direct) premises into a list if it is derived via inference, so the length of the list is usually 2. If it is an input task, then the premises list is empty. The hash-value is influenced by the input_id of the task, as well as the hash-values of the premises.
Another solution might be adding all elements of the evidential base into that tuple when computing hash-value, though this approach might be more costly.
I didn't think of that before. I guess this is because the python object of the first truth-value %1.0; 0.9% and that of the second have different ids. When computing the hash-value, it takes the id of an object into account.
Huh, interesting. I'm so used to value types that I forget sometimes that everything is an object in python. What do you think of just implementing a hash function in Truth that will only take the values into account and disregard the id?
I pushed up a commit, see if this works. I tested on your examples and when truth values are the same it doesn't store a duplicate belief.
What do you think of just implementing a hash function in Truth that will only take the values into account and disregard the id?
That looks better. Then there would be an another issue. When inputing two tasks with the same truth-value, actually we hope them both to be stored in the table, right?
What do you think of just implementing a hash function in Truth that will only take the values into account and disregard the id?
That looks better. Then there would be an another issue. When inputing two tasks with the same truth-value, actually we hope them both to be stored in the table, right?
We do? Wouldn't we only store one since they are both input tasks?
We do? Wouldn't we only store one since they are both input tasks?
According to the technical report 3.1.0
Beliefs/desires in a concept-level table: multiple versions of a belief with the same content but different truth-values compete each time a belief is requested. Bag is not used here because each request is one-time. Multiple versions of beliefs are kept because overlapping evidence may prevent the version with the highest confidence from being used.
Well, it looks like that statements in the beliefs table have different truth-values.
But I doubt on this. Even thought two input tasks have the same truth-value but different evidential base, should they be both kept in the table?
But I doubt on this. If two input tasks have the same truth-value but different evidential base, should they be both kept in the table?
What is the evidential base of an input task? Just a serial number? For derived tasks it is formed from the evidential bases of the parents, but for input... If we treat each input task with identical truth values as distinct then wouldn't the belief table quickly fill up with identical beliefs?
But I doubt on this. If two input tasks have the same truth-value but different evidential base, should they be both kept in the table?
What is the evidential base of an input task? Just a serial number? For derived tasks it is formed from the evidential bases of the parents, but for input... If we treat each input task with identical truth values as distinct then wouldn't the belief table quickly fill up with identical beliefs?
The evidential base of each input task is a set with only one identical element.
You may be right. The table may be filled up quickly.
The same problem exists in OpenNARS 3.0.4. So we can tentatively adopt this design, though it needs further discussion.
|
2025-04-01T06:38:05.332128
| 2016-08-17T16:11:32
|
171698287
|
{
"authors": [
"GolubevV",
"tas50"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4325",
"repo": "boxcutter/windows",
"url": "https://github.com/boxcutter/windows/issues/80"
}
|
gharchive/issue
|
7-zip is not uninstalled correctly.
Currently, the 7-zip is getting installed as part of Windows boxes provisioning logic, by the script named vmtool.bat.
It is supposed to be completely and correctly removed at a later stage by the dedicated script uninstall-7zip.bat, however it does not perform the correct uninstall of the 7-zip from the registry (and it is still present in the "Programs and Features" in Control panel).
This result in some unpredictable and hard to track behavior, when 7-zip is getting installed on the created box as part of integration tests, for example, for Chef cookbooks.
Here is the corresponding message from the packer\boxcutter log:
virtualbox-iso: ==> Uninstalling 7zip
virtualbox-iso: ==> WARNING: Directory not found: "C:\Users\vagrant\AppData\Local\Temp\sevenzip"
It happens due to the fact, that the following piece of code does not get executed:
msiexec /qb /x "%SEVENZIP_PATH%"
as the msi file and its parent folder are getting cleaned up by previous script, named clean.bat, which performs complete remove of all folders from the %TEMP% folder, including the download folder of 7-zip.
The possible solution can be to exclude the sevenzip folder from the cleanup, so that it can be correctly uninstalled.
I've like to move away from using 7zip altogether. We can unzip the files directly via Powershell, which would avoid the whole mess of installing / uninstalling 7zip multiple times. Look for that fix in the future one I get a nice powershell based pattern
I've built several hosts today and all of them correctly removed 7zip so I'm going to close this issue at now. If you're still having issues with the latest code in master feel free to open it back up
|
2025-04-01T06:38:05.355677
| 2024-01-14T14:14:19
|
2080745838
|
{
"authors": [
"scala-steward"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4326",
"repo": "bpholt/java-time-literals",
"url": "https://github.com/bpholt/java-time-literals/pull/117"
}
|
gharchive/pull-request
|
Update sbt-scalajs, scalajs-compiler, ... to 1.15.0
About this PR
📦 Updates
org.scala-js:sbt-scalajs
org.scala-js:scalajs-compiler
org.scala-js:scalajs-library
org.scala-js:scalajs-library_2.13
org.scala-js:scalajs-test-bridge
org.scala-js:scalajs-test-bridge_2.13
from 1.13.2 to 1.15.0
📜 GitHub Release Notes - Version Diff
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scala-js" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "org.scala-js" }
}]
labels: library-update, early-semver-minor, semver-spec-minor, commit-count:1
Superseded by #128.
|
2025-04-01T06:38:05.357501
| 2016-10-02T09:47:12
|
180504476
|
{
"authors": [
"bpietraga",
"coveralls"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4327",
"repo": "bpietraga/dollar_tracker",
"url": "https://github.com/bpietraga/dollar_tracker/pull/56"
}
|
gharchive/pull-request
|
Add stock prices chart
Connected to #44
Coverage remained the same at 89.495% when pulling 0bcfb725c50e362d63893744f1e94a3ffccdeb64 on add-js-stock-graphs into 49a6b702e1b31aa5dd74707f0f7cb1e05c36d0a8 on master.
|
2025-04-01T06:38:05.380825
| 2021-12-24T18:13:17
|
1088469795
|
{
"authors": [
"bradtaniguchi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4328",
"repo": "bradtaniguchi/nx-workspace-template",
"url": "https://github.com/bradtaniguchi/nx-workspace-template/issues/18"
}
|
gharchive/issue
|
[Enhancement] Look into reafactoring features to use createFeature
Is your enhancement request related to a problem? Please describe.
with ngrx 13, there is a new createFeature function that can be used to cut down on boilerplate.
Describe the solution you'd like
Refactor existing ngrx utils to leverage this.
Additional context
see: https://ngrx.io/guide/store/feature-creators
Not going to this right now.
|
2025-04-01T06:38:05.390908
| 2023-06-08T10:03:19
|
1747539391
|
{
"authors": [
"alili",
"bramblex"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4329",
"repo": "bramblex/niva",
"url": "https://github.com/bramblex/niva/issues/67"
}
|
gharchive/issue
|
Can we use a UI for configuration?
For those new to the process, configuring settings directly through json can be quite complex and may require referring to documentation. Perhaps a more user-friendly approach would be to offer a visual configuration method similar to that of vscode.
对于新手来说,直接通过json配置比较复杂,需要参考文档,也许我们可以提供类似vscode那样的可视化方式配置。
Thank you for your advice. I will discuss with other developers how to implement a user-friendly GUI configuration editor.
I am a front-end engineer and a fan of yours at zhihu. If you need any help with the development of this feature, I can contribute some code
|
2025-04-01T06:38:05.419256
| 2024-04-07T16:31:58
|
2229842892
|
{
"authors": [
"Shigeto-Amatake",
"benibargera",
"branislavhesko"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4330",
"repo": "branislavhesko/segment-anything-ui",
"url": "https://github.com/branislavhesko/segment-anything-ui/issues/5"
}
|
gharchive/issue
|
where is default_path variable?
I could not find in config.py.
Hi Branislav. I'm a bit confused about that as well. As well as what value we're supposed to put in there? I'd really like to use your app :) Thanks in advance!
Hello.
If you think default checkpoint for SAM path, it is called from here:
https://github.com/branislavhesko/segment-anything-ui/blob/496bef9ac7571edd9080def9816a8efc1f882093/segment_anything_ui/config.py#L35
|
2025-04-01T06:38:05.773502
| 2018-04-21T02:31:06
|
316454229
|
{
"authors": [
"AdityaAnand1",
"Brianvdb",
"NiekAkerboom",
"liuatgit",
"nzackoya"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4331",
"repo": "bravobit/FFmpeg-Android",
"url": "https://github.com/bravobit/FFmpeg-Android/issues/31"
}
|
gharchive/issue
|
Upgrade to FFmpeg 4.0
https://ffmpeg.org/index.html#pr4.0
:tada:
I'm waiting for 4.0 too.
@AdityaAnand1 @liuatgit You guys can build it yourself, which I highly recommend. Most FFmpeg binaries you find online have modules enabled that you might never use which will increase the size of it immensely.
@AdityaAnand1 @liuatgit
Good news! FFmpeg 4.0 is now supported by this library since version 1.1.4
dependencies {
implementation 'nl.bravobit:android-ffmpeg:1.1.4'
}
It is still 3.0.1
|
2025-04-01T06:38:05.812293
| 2018-06-20T21:33:13
|
334257681
|
{
"authors": [
"brechtsanders",
"webern"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4332",
"repo": "brechtsanders/xlsxio",
"url": "https://github.com/brechtsanders/xlsxio/issues/29"
}
|
gharchive/issue
|
xlsxio fails in the presence of xmlnamespace qualification
The same proprietary file that I mentioned in #28 also includes namespace qualification on the elements in several of its xml files. For example
<?xml version="1.0" encoding="utf-8"?>
<x:sst xmlns:x="http://schemas.openxmlformats.org/spreadsheetml/2006/main">
<x:si>
<x:t>NPANXX</x:t></x:si><x:si>
<x:t>JUR</x:t></x:si><x:si>
<x:t>CARRIERGRP</x:t></x:si><x:si>
<x:t>DETAILED_CARRIERGRP</x:t></x:si><x:si>
<x:t>ST</x:t></x:si><x:si>
<x:t>LATA</x:t></x:si><x:si>
...
xlsxio assumes the author of an xlsx file will not do this, but it is valid for them to do so.
I hack-fixed this in my fork by changing every comparison like this:
if ((XML_Char_icmp(name, X("sst")) == 0) || (XML_Char_icmp(name, X("x:sst")) == 0)) {
But this is just a quick workaround, since any namespace is possible, I just needed to solve for a particular vendor that was using x.
A proper solution would be to only compare the element name's ending substring, i.e. (pseudocode)
int len = "sst'.length();
if( element.name.last(len) == "sst")
This way xlsxio could just ignore namespaces all-together without failing in their presence.
I believe a seperate compare function to either match completely or match after the last colon should do the trick. When I have some time I can look into this.
Could you send me a complete example .xlsx file?
Would a similar fix be required in the get_expat_attr_by_name() function?
Hi,
I was just looking back on this issue that was never closed.
XLSX I/O relies on Expat to process the XML, which is why namespaces aren't checked by the library itself, as Expat normally takes care of that.
But if you still have this issue and you can send an example .xlsx file I would be glad to take a look at the issue.
Regards
Brecht
I’m several projects away from where I was when I wrote this. Feel free to
close!
Thanks.
Matt
On March 19, 2020 at 11:55:32 AM, Brecht Sanders<EMAIL_ADDRESS>wrote:
Hi,
I was just looking back on this issue that was never closed.
XLSX I/O relies on Expat to process the XML, which is why namespaces aren't
checked by the library itself, as Expat normally takes care of that.
But if you still have this issue and you can send an example .xlsx file I
would be glad to take a look at the issue.
Regards
Brecht
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/brechtsanders/xlsxio/issues/29#issuecomment-601357559,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ABPYNFA37QDDX55CB75YW2LRIJTCJANCNFSM4FGAC5HA
.
|
2025-04-01T06:38:05.821256
| 2020-04-06T17:28:38
|
595293512
|
{
"authors": [
"aweary",
"brendanzab"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4333",
"repo": "brendanzab/codespan",
"url": "https://github.com/brendanzab/codespan/issues/227"
}
|
gharchive/issue
|
codespan-reporting and codespan-lsp compatibility
I've been using codespan and codespan-reporting for error reporting for my CLI. I'm now working on an LSP extension and want to use codespan-lsp to covert to the proper LSP types, but it seems like codespan-reporting doesn't play well.
For example, I have a custom file system abstraction that implements the Files trait but codespan-lsp utilities like byte-span-to-range expect a reference to codespan::Files. They also expect a Span but codespan-reporting's Label uses Range<usize>.
Are there plans to align the packages one way or the other, or should I just roll my own diagnostics object and have compatibility layers for codespan-reporting and codespan-lsp?
Thanks for your issue!
I've been trying hard to de-emphasise codespan, and emphasizing codespan-reporting, because I've found it rather challenging to create a one-size-fits all thing for handling files. This was part of the work I did in inverting the dependency between those two crates. Eventually I'd like to deprecate codespan, if I'm honest.
codespan-lsp currently depends on codespan though, so sadly you are tied to using codespan if you want these conversions. It would be cool if there would be some way to make those conversions easier to implement for custom file system abstractions though! I'm still not sure where they fit though…
@brendanzab would you be open to including some LSP utilities in codespan-reporting? It might make sense to have codespan_reporting::lsp to mirror the codespan_reporting::term backend. I'd be happy to help here if you're interested.
Ohhh, that is a good idea, yes! Now I have a better handle on the lsp-types version ranges this could also be helpful too - it was a constant pain to keep having to update them.
I'd definitely be open to collaborating on this! Feel free to chat on Gitter or the the #langdev channel on the rust community discord if you like!
|
2025-04-01T06:38:05.840020
| 2024-02-14T18:34:41
|
2134942543
|
{
"authors": [
"brettonw",
"reedstrm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4334",
"repo": "brettonw/Remove-Home-Assistant-Device",
"url": "https://github.com/brettonw/Remove-Home-Assistant-Device/pull/1"
}
|
gharchive/pull-request
|
implement remove-by-model-name. Make backups of configs
I had several hundred iBeacon devices to get rid of, all with long hex-string names, so I extended your code to do model names as well. Minimal change w/o restructuring completely. In the spirit of it being a power-user tool, but even power users can sometimes shoot themselves in the foot, I provided one level of toe-armor: it makes a backup of the three config files.
Looks good, though I haven't had occasion to use it. Does it still work for basic removal?
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.