added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:54:48.823804
| 2017-05-03T12:13:00
|
225965616
|
{
"authors": [
"bolbass",
"chaorace",
"e1ectron",
"legdrop"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13467",
"repo": "CQUI-Org/cqui",
"url": "https://github.com/CQUI-Org/cqui/issues/517"
}
|
gharchive/issue
|
Missing gossip on screen
Good afternoon, firstly, thank you very much for such a wonderful modification, which I use with the acquisition of the game.
But there is a problem - gossip on the screen about AI does not appear at all, just notifications about my studied technologies of culture and science.
How can I solve this problem? Or tell me, which file or string I should cut to get rid of the gossip removal function.
I'm playing the Russian version of the game.
Thank you very much!
This may be an issue with the heuristic used for intercepting gossips. I did my best to render it language neutral, but there's only so much I can achieve as an English-only speaker.
Hey @e1ectron, does this issue affect you as well?
Sorry, I can't check it now and in the near future (don't have access to actual game version).
Hey, @chaorace. Test it on a 1.0.4 (<IP_ADDRESS>) Mac version on Russian and English.
For a hour, I haven't seen any gossips. So problem is not the language.
May be platform? @legdrop, you play on Win or Mac?
Windows 10, 64 bit.
I think @legdrop wants to read popup gossips that appears on vanilla on center of the screen. If so you need to remove line 35 in UI/Panels/StatusMessagePanel.lua that was added on this commit f04d0b5e0ff387a80420ccb0009865df134fc524
@bolbass
Yes, you understood everything correctly, thank you all very much!
I do not have the option now to run the game, but please check, I selected the correct code to delete?
Don't delete it :)
The commit I mentioned above was already fixed in CQUI so my last comment is wrong.
The last commit that affects gossips is c8963130252a8cb90546423564eb6c22911e64ea
The result of this commit is that you can choose gossips you want to see or disable on CQUI settings that are above the minimap. Have you tried changing it?
Thanks a lot, it works!
Okay... so did we ever figure out if this issue was caused by CQUI not correctly showing gossips?
Yes, there is at least one type of gossips that is missing on gossips settings screen (a gossip when other civs enter new era for example).
This type of gossips is under "City>>X has counquered Y" settings. We should fix it and can close this issue as soon as it's enough to unchek "Trim gossip messages" to see all the gossips.
|
2025-04-01T04:54:48.896393
| 2016-09-27T08:03:37
|
179422868
|
{
"authors": [
"bluikko",
"mike-csis"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13469",
"repo": "CSIS/EnrollmentStation",
"url": "https://github.com/CSIS/EnrollmentStation/issues/16"
}
|
gharchive/issue
|
Typo in "Terminate (WILL revoke)"
There is a typo on the termination confirmation dialog:
"This will terminate the Yubikey, wiping the PIN, PUK, Management Key and Certificates. This will also revoke the certificiate. Proceed?"
Note "certificiate" and not "certificate".
Thx :)
|
2025-04-01T04:54:49.057127
| 2023-12-13T02:29:00
|
2038841662
|
{
"authors": [
"LachlanMurphy",
"Pandabear1125"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13470",
"repo": "CU-Robotics/firmware",
"url": "https://github.com/CU-Robotics/firmware/pull/21"
}
|
gharchive/pull-request
|
feature-robo-control
Moved around some items in the main file and created the control_manager.hpp/cpp. We are now good to start working on different control methods, though we do also need to think about how to integrate the yaml into the control manager file.
Marking this as closed. The main robot control code was merged with feature-controllers. This is now obsolete.
|
2025-04-01T04:54:49.079783
| 2024-10-16T17:25:58
|
2592568461
|
{
"authors": [
"NoraLoose"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13472",
"repo": "CWorthy-ocean/roms-tools",
"url": "https://github.com/CWorthy-ocean/roms-tools/pull/165"
}
|
gharchive/pull-request
|
Fix mask at boundaries
Closes #163.
The failing tests are not introduced by this PR. I opened a separate issue #166.
|
2025-04-01T04:54:49.119309
| 2019-01-20T07:32:35
|
401071074
|
{
"authors": [
"Cadene",
"ekagra-ranjan"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13473",
"repo": "Cadene/pretrained-models.pytorch",
"url": "https://github.com/Cadene/pretrained-models.pytorch/pull/124"
}
|
gharchive/pull-request
|
Updated fbresnet to accept different sized images
The update allows fbresnet to process images larger or smaller than prescribed image size (224x224). Will be useful while finetuning or testing on different resolution images.
Thanks @ekagra-ranjan
Sorry for the late merged. I had to manually evaluate the model on imagenet.
|
2025-04-01T04:54:49.133794
| 2023-10-24T23:03:43
|
1960228294
|
{
"authors": [
"Dinopony",
"RadzPrower",
"nbrochu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13474",
"repo": "CalDrac/hintMachine",
"url": "https://github.com/CalDrac/hintMachine/pull/75"
}
|
gharchive/pull-request
|
Initial Implement of Super Mario Bros. 3
I have managed to create an implementation for Super Mario Bros. 3 (SMB3).
Quests include:
Score - Collect 100,000 points
Levels - Complete 3 levels
Worlds - Collect a hint token for each world completed
Streak - Get a bonus hint token if you complete 10 levels in a row without dying
There are still cooldown timers on Levels and Streak at the moment as a precaution since there was an issue of values staying in memory too long on death in certain cases resulting in multiple points for those quests. I developed a method of avoiding that issue in code, but have left the timers for the time being as it is beta testing. If we see no reports of messages regarding the cooldown, it should be safe to remove it eventually.
I also included a good many safeguards into the code to avoid means of cheesing the system (entering and exiting pipes on the map) and addressing specific scenarios (warp whistles do not count towards world completion for instance).
Don't forget to add a README entry.
Also a nitpick, and I don't know how @Dinopony feels about this, but I think we've tended to avoid abbreviations in game connector names so far. I personally think there should be no ambiguity when looking at a connector name. SuperMarioBros3Connector is crystal clear. SMB3Connector you have a good sense it's Super Mario Bros 3 but it could be something else when you open it.
Good point, especially in regard to Super Mega Baseball. I've expanded those out to the full names.
Also, didn't realize we needed to add our own entries to the README table, so good to know going forward.
Yup indeed, having full names is encouraged because it could get obscure really quickly if we all used abbreviations 😉
I'll try to find time to test it in the next few days, thanks for the contrib 👍
Tested the connector. Everything working as expected on my end. Good work.
|
2025-04-01T04:54:49.164745
| 2021-09-11T18:57:45
|
993915679
|
{
"authors": [
"Sabine-Justilien",
"davidnixon",
"upkarlidder"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13475",
"repo": "Call-for-Code-for-Racial-Justice/Five-Fifths-Voter",
"url": "https://github.com/Call-for-Code-for-Racial-Justice/Five-Fifths-Voter/issues/203"
}
|
gharchive/issue
|
Child Care resources
Background on the problem the feature will solve/improved user experience
Marginalized communities often bear the heaviest burden when trying to exercise their right to vote especially when it comes to childcare. Parents either have to rely on family members to take care of their children during this period or take the child with them. This can be tiring for a parent especially if there are limited toilets, lack of shade, or long wait times.
Describe the solution you'd like
The YMCA offers numerous childcare throughout their centers. As a national organization there is greater opportunity for parents to reach out for child care during voting periods. The YMCA is a proponent of civic engagement as well as community wellness so we can point users in need of childcare to their resources. They have a 'Find your Y' mapping tool that we can either pull into 5/5 or point users to the external link.
Tasks
Investigate how to incorporate the 'Find your Y' location tool as a resource for parents/guardians
Decide if this map can be pulled into the 5/5 solution or should be pointed to via a link on 5/5
Acceptance Criteria
A user can locate the 'childcare' resource in the resources tab
A user can view the 'Find your Y' tool and select the location they need to find childcare in
A user can connect with/ contact the selected Y and follow the process for getting childcare support
From our scrum 09.15 meeting, the task is to replace the
https://ymcaofcoastalga.org/news/2018/10/17/ymca-news/ymca-works-to-help-make-it-easier-for-parents-to-vote/
with
https://www.ymca.org/find-your-y?distance=175&lat=26.0517448&lng=-80.13727569999999&geolocation_geocoder_address=33004&type=branch
That string is in the ui/src/views/SupportPage/SupportPage.vue file.
Probably it's a straight search/replace.
|
2025-04-01T04:54:49.174250
| 2020-11-19T10:10:19
|
746433064
|
{
"authors": [
"Miqhtiedev",
"SkillBeatsAll"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13476",
"repo": "CalmGuild/CalmBot",
"url": "https://github.com/CalmGuild/CalmBot/pull/61"
}
|
gharchive/pull-request
|
Merge master into Production
Summary
Merges master branch into production
We are not ready for 1.1 yet.
soon tho
|
2025-04-01T04:54:49.189802
| 2021-01-05T16:04:28
|
779224726
|
{
"authors": [
"CamDavidsonPilon",
"mekomlusa"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13477",
"repo": "CamDavidsonPilon/lifelines",
"url": "https://github.com/CamDavidsonPilon/lifelines/issues/1196"
}
|
gharchive/issue
|
Updates to the model serialization example
First of all, thanks for this awesome package - it's really helpful for newbies as me to quickly get started in the field of survival analysis. I've trained a CoxPHFitter model and would like to save it to the disk for future inference. I was able to find the following codes from this file but looks like there's a typo below: (see my comments next to those lines)
from dill import loads, dumps
from pickle import loads, dumps
s_cph = dumps(cph)
cph_new = loads(s_cph)
cph.summary # shouldn't it be cph_new.summary?
s_kmf = dumps(kmf)
kmf_new = loads(s_kmf)
kmf.survival_function_ # shouldn't it be kmf_new.summary?
The codes above only serialize binary objects in memory. It would be helpful if the following example codes could be added, too, to help others save their trained models to local path:
import pickle
with open('/path/my.pickle', 'bw') as f:
pickle.dump(cph, f) # saving my trained cph model
with open('/path/my.pickle', 'rb') as f:
cph_new = pickle.load(f)
cph_new.summary # should produce the same output as cph.summary
I can submit a merge request for this, if anybody finds it interested.
I love this suggestion (and agree that it's a typo) - feel free to send a PR and I'll merge it
I love this suggestion (and agree that it's a typo) - feel free to send a PR and I'll merge it
|
2025-04-01T04:54:49.199265
| 2024-03-28T20:24:03
|
2214064466
|
{
"authors": [
"jatkinson1000"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13478",
"repo": "Cambridge-ICCS/FTorch",
"url": "https://github.com/Cambridge-ICCS/FTorch/pull/103"
}
|
gharchive/pull-request
|
Add opt options rebase
This is an updated version of #78 rebased onto main after the GPU changes by @jwallwork23
@ElliottKasoar's comment on the original PR:
Resolves #73
Adds flags in all(?) functions that operate on tensors (tensor creation, model loading, forward) to optionally disable autograd, which should improve performance for inference.
Also adds a similar flag to set evaluation mode for the loaded model.
Evaluation mode
Contrary to my initial comments in #73, from testing evaluation mode does appear to be preserved, both between saving and loading TorchScript, and when applied to the loaded model.
In most cases evaluation mode is therefore likely to already be set, but I think it's useful to have the option to change it, particularly if FTorch may be extended to facilitate training (#22).
NoGradMode
Enabling or disabling gradients is more complicated, as it defined via a context manager, which only appears to define the behaviour within its own scope, and so it seems necessary to enable/disable gradients before every code block that operates on tensors (similar to the Python equivalent with torch.no_grad():).
InferenceMode
No changes are currently included, but it would be good to support InferenceMode too eventually, as it should provide further performance benefits over NoGradMode.
However, it has stricter requirements, and the mode was only added (as a beta) in PyTorch 1.9, so we would need to be careful if we want to support older versions.
Model freezing
No changes are currently included, and less directly applicable to the main FTorch library, although there are still interactions e.g. freezing the model can allow InferenceMode to be enabled when loading the model.
Freezing is currently the "default" when tracing in pt2ts.py, but not for scripting, despite potentially improving performance.
Freezing appears to (sometimes) introduce numerical errors when saving the reloading (differences ~10^-6), and can seem to lead to issues loading with Forpy too.
(For more general explanation of autograd/evaluation mode, see autograd mechanics).
Note: I've also removed the old, commented out torch_from_blob function.
Hi @ElliottKasoar after the work @jwallwork23 did there were some conflicts with your PR in #78.
I have done my best to rebase your work, but please could you take a look and see if everything seems in order to you?
@TomMelt You originally reviewed this PR and approved, but if you could take a quick glance and re-review it would be appreciated - since @jwallwork23 restructured the order of functions in the files and added additional arguments in the same place as @ElliottKasoar some of the merge conflicts got a little hairy so I may have missed the odd thing!
Before we merge we need to:
[ ] Add some docs for these new options
[ ] Add an example showing the use of these new options?
Or perhaps not given we are not training yet and just use the defaults...?
Added a note to the FAQ about eval and no_grad settings.
A detailed example will perhaps wait until these are used as part of #111 since for now they are the sensible defaults for running inference.
Squashing and merging shortly.
|
2025-04-01T04:54:49.215222
| 2017-07-19T13:42:36
|
244045267
|
{
"authors": [
"Malstrom",
"Schwad",
"coorasse"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13479",
"repo": "CanCanCommunity/cancancan",
"url": "https://github.com/CanCanCommunity/cancancan/issues/432"
}
|
gharchive/issue
|
WIKI error Page 'Role Based Authorization'
In your wiki you purpose to do this:
rails generate migration add_role_to_users role:string
But the column type needs to be an integer for work not string then:
rails generate migration add_role_to_users role:integer
Howdy @malstrom thanks for bringing this up. It's a large wiki, could you copy in the link to the specific page you're referring to?
Why does it have to be an integer?
I wrote it in first my post, role must be integer
As per my comment above: why cannot it be a string? Thanks for clarifying.
I close for stale. Feel free to reopen.
|
2025-04-01T04:54:49.241365
| 2017-11-22T14:37:51
|
276090313
|
{
"authors": [
"coveralls",
"jmaupetit"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13480",
"repo": "Candihub/pixel",
"url": "https://github.com/Candihub/pixel/pull/84"
}
|
gharchive/pull-request
|
WIP: Integrate Workflow management app
Core features
[x] Add viewflow new dependency
[x] Configure viewflow
[x] Add django-debug-toolbar dependency for development
[x] Create a submission process
[x] Add tests for the submission process
[x] Add tests for the download task
Fixes
[x] Add support for help text in form fields
[x] Fix content layout (sticky footer & width)
Coverage remained the same at 100.0% when pulling ad995455c214ff36768ddf70d737bb4c0838626e on add-viewflow-and-upload into 0dbb55365ecaa00016b2c45e54e26a02c4dadc02 on master.
|
2025-04-01T04:54:49.244501
| 2022-05-03T19:59:46
|
1224563580
|
{
"authors": [
"alecKent",
"danaspiegel",
"elibosley"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13481",
"repo": "Cannabis-Labeling-API/universal-cannabis-api",
"url": "https://github.com/Cannabis-Labeling-API/universal-cannabis-api/issues/9"
}
|
gharchive/issue
|
404 response to /regulator/{id}/ when not all ids could be mapped.
I am curious about the logic for using 404 as the response if not all ids were able to be attached.
404 is defined as "The server has not found anything matching the Request-URI".
If I am reading the docs correctly in the case that not all ids could be mapped the request URI /regulator/{id}/ is valid and a matching resource was found, the issue exists within the array of ids to map.
Since the /regulator/{id}/ transaction doesn't seem to follow ACID I think the request should return a 200 response code since it is possible that some ids were mapped successfully.
Then it is on the caller to check if the failures array has data / exists at all.
Thoughts on this?
This is a really great point. I am happy to change to a 200 with a failure array.
Nice. I think a 200 would result in the least amount of confusion.
That way 404 can be reserved for when the requested resource is not found.
#13
Closing issue as this will be handled with issue #13
|
2025-04-01T04:54:49.268463
| 2016-03-21T22:15:47
|
142486955
|
{
"authors": [
"tayzlor"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13482",
"repo": "Capgemini/kubeform",
"url": "https://github.com/Capgemini/kubeform/issues/46"
}
|
gharchive/issue
|
Add support for ubernetes-lite
Requires a 1.2 upgrade. See https://github.com/kubernetes/kubernetes/blob/release-1.2/docs/proposals/federation-lite.md
Should be available in aws after https://github.com/Capgemini/kubeform/pull/64 https://github.com/Capgemini/kubeform/pull/77
|
2025-04-01T04:54:49.309398
| 2024-06-19T08:45:09
|
2361755833
|
{
"authors": [
"PJUllrich",
"mcrumm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13484",
"repo": "CargoSense/absinthe_client",
"url": "https://github.com/CargoSense/absinthe_client/issues/17"
}
|
gharchive/issue
|
Release 0.1.1 on Hex
Hey @benwilson512, it seems like you bumped the version to 0.1.1 but that version isn't on Hex. Would you mind releasing it there as well, please? Thank you!
Done, thanks! :)
|
2025-04-01T04:54:49.312849
| 2017-06-19T01:45:50
|
236761343
|
{
"authors": [
"benwilson512",
"kdxu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13485",
"repo": "CargoSense/ex_aws",
"url": "https://github.com/CargoSense/ex_aws/pull/430"
}
|
gharchive/pull-request
|
Update deps version {hackney, poison}
Hi. I'm using dialyzer it depends on ex_aws.
I found error when getting error building a plt of certifi,
And It caused by an old version of hackney depends on an old version of certifi.
Summery
Updated hackney latest version.
Updated poison latest version.
Backgrounds
current version of hackney has dialyzer warnings. (https://github.com/benoitc/hackney/issues/409)
now latest version of hackney, solved dialyzer bugs.
Hey! I appreciate the PR, but I'm not really sure I see the need to force people to ugprade. I already support the latest versions of hackney and poison, all this PR does is remove support for earlier versions.
I really do appreciate your desire to help, but I think it will be easier for me to just update dialyze, and leave the other dependencies as is.
Thank you reviewing for this PR.
Now I know these backgrounds of Poison and apologize for my fault.
So, you may close this pull request 🙇
Thanks!!
|
2025-04-01T04:54:49.361211
| 2021-12-25T16:40:47
|
1088626277
|
{
"authors": [
"SvicidalBug"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13486",
"repo": "Cartmanishere/zippyshare-scraper",
"url": "https://github.com/Cartmanishere/zippyshare-scraper/issues/35"
}
|
gharchive/issue
|
"redirected more than 3 times" during download link verification
This tool seems to have stopped working for me recently. Links can still be parsed, but when verifying download links, it says that it redirected more than 3 times. The links are not saved in links.txt. Additionally, the 'Last download' date on Zippyshare is not updated.
Here is my terminal output when running a single link:
Enter URLs (leave blank to stop): https://www25.zippyshare.com/v/dCDCcbJk/file.html
Enter URLs (leave blank to stop):
[*] pattern_1 has failed for link: https://www25.zippyshare.com/v/dCDCcbJk/file.html
[*] Trying next pattern
[*] pattern_2 has failed for link: https://www25.zippyshare.com/v/dCDCcbJk/file.html
[*] Trying next pattern
[*] pattern_3 has failed for link: https://www25.zippyshare.com/v/dCDCcbJk/file.html
[*] Trying next pattern
[*] pattern_4 has failed for link: https://www25.zippyshare.com/v/dCDCcbJk/file.html
[*] Trying next pattern
[*] 1/1 links parsed https://www25.zippyshare.com/d/dCDCcbJk/5/VPN-White-Paper.pdf
[*] Verifying download links...
[*] https://www25.zippyshare.com/d/dCDCcbJk/4/VPN-White-Paper.pdf redirected more than 3 times
[*] All download links saved at links.txt
Turns out I'm just dumb and running an older version of the tool. Git cloned the latest version and it works fine.
|
2025-04-01T04:54:49.367946
| 2015-02-19T15:07:20
|
58219827
|
{
"authors": [
"Cartofante",
"Kartones",
"juanignaciosl",
"rafatower"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13487",
"repo": "CartoDB/cartodb",
"url": "https://github.com/CartoDB/cartodb/pull/2351"
}
|
gharchive/pull-request
|
Some software extends dbf standard with extra encoding files, fixing #449
@Kartones, @rafatower CR this, please. I didn't know about those extra encoding files and it's pretty common (and useful for us). For example, QGIS save it. It's a workaround for DBase encoding support (I realised looking at the supported encodings table of dbf gem.
Frontend tests were OK :+1: (details)
Frontend tests were OK :+1: (details)
:+1: good research!
+1 nice!
Frontend tests were OK :+1: (details)
Frontend tests were OK :+1: (details)
Frontend tests were OK :+1: (details)
@Kartones could you CR this again? QGIS allows many exotic encodings not supported by PG, so I had to filter them and code has changed.
+1
|
2025-04-01T04:54:49.408109
| 2023-01-06T02:54:50
|
1521801296
|
{
"authors": [
"CasualX",
"tremwil"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13488",
"repo": "CasualX/pelite",
"url": "https://github.com/CasualX/pelite/pull/267"
}
|
gharchive/pull-request
|
Use correct base address for 32-bit ASLR images
Fixes #266.
Adds a method image_base(&self) -> Va to the Pe trait:
For a PeFile, returns the preferred virtual address, i.e. self.optional_header().ImageBase.
For a PeView, it returns the true virtual address of the image. By default, this is the same as the above, except when constructing via PeView::module (uses base directly) or via from_bytes_and_base (where it is provided by the user).
The return value of this method is used for VA/RVA conversions in the Pe trait, replacing self.optional_header().ImageBase.
Unfortunately PeView was meant to be also usable if you just copied the memory of a running image (eg. from another process) where the base of the image ptr is not actually the real virtual address of the image. I've always had a nagging feeling that ignoring it and relying on ImageBase being 'corrected' was not reliable and now it's coming back to bite me :)
This will probably require an extra field in PeView containing the actual image base and trying not to break APIs.
I see. In this case, adding the extra field as you said and only setting it to self.image.as_ptr() as Va when the PeView is constructed using PeView::module should not break any existing APIs, correct?
Also, I guess a new (unsafe?) PeView constructor taking a AsRef<[u8]> and the base address would be useful in case one encounters issue #266 after copying the memory of a running image from another process.
PeView should work when copying the memory of a running image now. However, I'm not familiar with the Serde library, so I'm not sure how to serialize the PeView's image_base field without potentially breaking backwards compatibility for some formats (e.g. sequential ones). Is it fine to assume that some formats may not be compatible no matter what and just add a new serialize_field call at the end of serialize_pe that writes the image base?
Thanks for the update! I'm a bit distracted with other projects right now, I'll try to take a look at it tomorrow or during next week.
I've changed the API to set the base address after construction (I'm not a fan of a bunch of _with constructors, Rust could really use default parameter values...)
|
2025-04-01T04:54:49.464494
| 2022-09-20T14:15:34
|
1379499203
|
{
"authors": [
"stephen-rowan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13489",
"repo": "Catalyst-Auditing/Community-Governance-Oversight-Coordination",
"url": "https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/106"
}
|
gharchive/issue
|
F8 Community Governance Oversight Meeting 3 - Thursday 8th September 2022 - 1400 UTC
Agenda - F8 Community Governance Oversight Meeting 3
Scheduled date/time of meeting : 1400 UTC, Thursday 8th September 2022
Present:
[x] Allison Fromm
[ ] George Lovegrove
[x] Kenric Nelson
[ ] Matthias Sieber
[x] Phil Khoo
[ ] Stephen Whitenstall
[x] Tevo Saks
[x] Thorsten Pottebaum
[x] Vanessa Cardui
[x] Treasury
[x] Andre Diamond
[ ] Miroslav Rajh
Proposed Agenda
Matters arising/action points
Pay Allison $75 for TH Slides - ?
Pay Stephen $500 for PM - Aug 2022 yes
Pay Kenric remainder of the research budget - Aug 2022 yes
Last meeting we were looking to add something to F9 IOG survey (Phil) - did we? - in theory yes but depends on form, scope, data protection
We missed adding something to Danny’s newsletters - tut
We agreed to discuss async the process to build public register of parameters - didn’t happen yet
CGO Treasury
Andre & Miroslav - after what has been paid, what is left?.
https://docs.google.com/spreadsheets/d/1v8EySgaWqoYlOVxHAjKKufj24x8CPn-no33EUPZP8sw/edit?usp=sharing
will rec 6860 $ on Mon
Can pay for all TH slides and project management
Partially pay Treasury management
Items left to pay for CGO8 - Reports, Retros, Surveys, Prep of F9 proposal - there’s not one… Should this be F10 proposal?? Town Hall Slides, Treasury Management. Project Management
All meeting attendance is already paid
should we develop a process for paying deliverables? Splitting equally is easier (Allison)
Tevo - we estimate delivery time of how much tasks cost. Everyone who contributes shares how much time they spent, if no information, the estimated cost will be shared equally between participants of specific task.
Vanessa - Should we split distributions between high level epics/projects/themes or we take reward each deliverable differently.
Phil - community distribution model - equal split and can choose to redistribute your share to others
Kenric - baseline = split evenly, then specific tasks - bonus pool? for those who put in extra effort. Decide thisusing Phil’s method.
Phil - easier to just pay everyone equal and then reallocate
Phil - this doesn’t change what’s already allocated
Andre - pay treasury in full this time, and then divide the rest between all of us Rename it “deliverables”
Leave it with Treasury until such time as deliverables are done? AGREED
Need to set aside money for closing report too
F7 proposal final payment - deliverables payments is what is outstanding
Pay Stephen for close out report and project management
$840 in wallet now
$6,366 to come on Monday
Town Hall Slides
Allison - Updates on Town Hall Slides
https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/87
Continuing
Will stay with every other week cadence
Will continue on recap of history, what’s happening in governance, and then parameter changes
CGO Project Board Review
Board:
https://github.com/orgs/Catalyst-Auditing/projects/3/views/1
F8 CGO Scope & Deliverables
5.1 Challenge Setting
What is the scope of F8 Challenge Setting oversight ?
Issue :
https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/88
We’ve agreed George and Phil are working on this, but what exactly are its deliverable(s) likely to be?
Do we need to consider how to maintain (and be seen to maintain) independence of oversight, given George is so strongly involved with working to change the Challenge-Setting process? George - conflict of interest is sthg he’s aware of. Meeting w Kriss on Fri - there will be changes soon. Whatever approach they want to take, will communicate it w CGO. Survey to gain community feedback on what IOG want to do? Tevo: this is best oversight we could have - let’s not add barriers. George : I do the work, oversight comes from everyone else.
What is the deliverable?? A survey? The proposals George has in? Analysis of discussion e.g in Telegram? To be discussed next meeting after George’s meeting with Kriss. Communicate that this discussion has been had, and the oversight will be to engage with or survey the community - sentiment analysis.
Thorsten - who is enacting changes to the challenge setting process? Also what is the intent, and how do we measure if it is achieved? A survey probably won’t cover that.
Kenric - not every community change HAS to be done by vote.Conversations, surveys etc are already happening - that’s also a decentralised process.
Allison - agree, it’s not always about a vote. How much surveying, conversation etc is enough to create legitimacy?? to protect the implementer from accusations of a decision that is not representative
Thorsten - agree, it’s not clearly defined how decision making is done if not via a vote. It becomes hard to determine if what is happening is within the parameter of controlling decision making process. Who decides if a vote is needed or not.
Phil - Dripdropz ATH last night - building a petition module to their voting structure, which fills some of the role we are talking about. An interesting possibility.
F10 looks likely to be another odd one: developer-heavy, and excluding much of the other work people do. Should we be additionally looking to maintain oversight of what sort of effect this is having on the community, the work that is done and the proposals that are getting funded?
5.2 Catalyst Circle Oversight
What is the scope of F8 Catalyst Circle Oversight ?
Issue :
https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/89
We agreed that those working on this are Tevo (general), Kenric (compensation), Vanessa (CC Oversight) - but what exactly will its deliverable(s) be?
We’ve discussed the ATHs at the end of June regarding planning for CCv4 - things have moved on a lot since then, but do we want to write up anything specific about those? Including Kenric’s points about remuneration?
CCv4 Election will be on DripDropz - what oversight can we / should we have? Do we want to have any oversight of the election design? e.g. what is the effect of doing it via DripDropz, of 1-wallet-1-vote, plutocratic voting power, etc)? And/or do we want to interrogate the resultsof the vote (in terms of what kinds of people voted, who felt “invited”, or anything else)?
Should we oversee the process that got us to this election format (e.g. was there enough community consultation?) It’s already being done by Oversight of Catalyst Circle - is it enough?
Is there scope for us to have oversight of whether the process is working and is fair, or is excluding anyone? (e.g. is there enough time for people to write platform statements; is it OK that no support is being offered to help people stand, no education sessions, etc)
The above is a LOT of oversight. Remuneration for Circle is emerging as quite a big issue, so should we make that our focus? A survey on this could canvass community opinion on how the remuneration affects whether people stand for election, etc.
Kenric - concerns about the structure of the election. CCv3.2 decided to redesign Circle, not just prepare for the election.
Tevo: what are the roles and functions in catalyst - is Circle an official role? - There are functional groups, and they are not represented? (we don’t know that before voting)
Kenric - do we need to put together a report on what we think the problems are? The point of good processes = good outcomes. Kenric thinks we now have a poor outcome - gone from well defined roles (which were not perfect) to no design (? This is not necessarily accurate - the approach of “self-defined” IS a design)
Phil: as an oversight group we shouldn’t have a group opinion? Identify the parameters that have been changed or adjusted for this to happen? could look at this sort of via parameter changes process - it’s a sub-parameter that needs looked at separately
Vanessa - can we assess the outcome when it hasn’t happened yet in the sense that we haven’t had the election?
Phil - yes we could add it to TH slides
George: Observe what is changing and look at the process they took. Pros and cons?
Phil: is Circle becoming a decision making body? V- not yet
Phil- we could use TH slides and newsletter to help inform people about what has been decided
Our oversight ends at after Circle voting results are shared
What about making our focus the remuneration
Kenric - no
Time sensitive tho - decision on pay will not come immediately
Tevo - I think statements from oversight will not have impact
5.3 d-Reps
The White Paper is progressing well - Kenric to give a quick update?
dReps “Review Paper” is nearing completion. Plan to submit to public by the end of Sept. Three sections: literature review, voting power, community agreements
Tevo’s “participating dRep” perspective - what will be the deliverable?
5.4 Governance parameters
What is the scope of F8 Governance parameters ?
Issue : https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/90
We have agreed that Allison, Vanessa, Thorsten and Matthias are working on this- and Andre? (to maintain the register of changes)? And Phil?
We agreed that one deliverable will be creating, publicising and maintaining a publicly- accessible register of parameter changes (including identifying who is actually making decisions, and info on the conditions on which community can take over each parameter)? So what do we need to do to launch that? And how will it be maintained? (Andre?)
We also suggested defining broad categories of parameters - should that come later, when we see what parameters there are?
Could the deliverables also include suggesting a process that IOG or anyone else should follow to launch a parameter change - particularly., an idea of how to determine what is sufficient community consultation?
We have discussed a decision-making tree (is decision urgent, etc) using info from Jeremy’s Miro https://miro.com/app/board/uXjVOsELu0U=/ - to categorize parameters - Tevo / Phil. Are we proceding with ths? and is it part of the above proposed process?
Tevo: Location of parameter changes - what it affects - who takes the decision and who is affected by it
Phil: do we focus on parameters that have “caused waves”, and ones that are announced during this proposal?
Allison: yes, narrow it down to the ones which are changing now. Possible test case - SSI proposal. Conflict of interest?
Phil: If you’re creating a proof of concept it’s not really a parameter change till it affects everyone. If it’s optional, an experiment, it’s not a parameter change
Tevo: We will send out the google form - but which parameters we report on, we will discuss.
V will share google form in Discord - will put it in next slides
Phil: community participation doesn’t yield that many results so let’s be realistic
Tevo -suggest an emphasis on propsal assessor/VPA parameters. The way we have scaled is ineffective. To be discussed.
We also said that we would follow up on the F7 meeting with Harris - particularly, he was going to share info on IOG’s current process re parameter changes. How will we progress this?
Allocating pay for deliverables
In the light of the above discussion - how do we want to do this? Equal share for all involved, as with F7 project, or something else?
CGO Proposal Reporting
F7 CGO Proposal Reporting
https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/91
F8 CGO Proposal Reporting
https://github.com/Catalyst-Auditing/Community-Governance-Oversight-Coordination/issues/92
Summary of CGO meeting for Catalyst Weekly Newsletter
Executive, Legislative, and Judicial functions (Kenric)
This topic can be removed - there isn’t time in these meetings - Kenric
2 meetings ago, we said we’d find a time to discuss (async or informally) the below, that Kenric raised. Don’t think it happened, and don’t want to lose it - so can we quickly decide where and how this discussion could happen?
I’m (Kenric) not sure where to put this, but I suggest having a discussion about the Executive, Legislative, and Judicial functions of governance. The voting processes of Catalyst map clearly to a Legislative function. What are the Executive and Judicial functions? Perhaps the approved projects are an executive function. Perhaps the Circle’s identification and discussion of problems is a judicial function. Clarity about this mapping might help define these different roles, and thereby the form and expectations of these roles.
AOB?
|
2025-04-01T04:54:49.469612
| 2016-09-15T17:43:00
|
177240390
|
{
"authors": [
"kevinzhow"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13490",
"repo": "CatchChat/Yep",
"url": "https://github.com/CatchChat/Yep/issues/449"
}
|
gharchive/issue
|
Buddybuild crash report on build #520
Buddybuild detected a crash from<EMAIL_ADDRESS>View Full Crash Details
Crash report from Unknown user
|
2025-04-01T04:54:49.496405
| 2022-09-05T08:51:38
|
1361613558
|
{
"authors": [
"ArnaudLigny",
"magentix"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13492",
"repo": "Cecilapp/Cecil",
"url": "https://github.com/Cecilapp/Cecil/pull/1455"
}
|
gharchive/pull-request
|
fix: end tag for main element
Changes proposed in this pull request:
Fix missing end tag for the main element
Thanks a lot @magentix 👍
@all-contributors please add @magentix for bug and code.
|
2025-04-01T04:54:49.537206
| 2021-04-24T05:16:18
|
866638619
|
{
"authors": [
"ChrisMacNaughton",
"coveralls"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13494",
"repo": "CentauriSolutions/EyeDP",
"url": "https://github.com/CentauriSolutions/EyeDP/pull/225"
}
|
gharchive/pull-request
|
A User is technically optional on an audit event
Closes #223
Coverage remained the same at 84.235% when pulling ba7125d59df83853147cad4af2a5dec482ad7b92 on bug/224 into 14711e05549bf8908656f938f7578068766d0b2a on master.
|
2025-04-01T04:54:49.543070
| 2022-02-21T19:45:53
|
1146170578
|
{
"authors": [
"Sindica",
"q131172019",
"sonyafenge",
"zmn223"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13495",
"repo": "CentaurusInfra/arktos",
"url": "https://github.com/CentaurusInfra/arktos/pull/1377"
}
|
gharchive/pull-request
|
scale-up mizar support
What type of PR is this?
/kind feature
What this PR does / why we need it:
Start arktos-network-controller
create default network object (flat or mizar) of for system tenant
kubeup support mizar
Which issue(s) this PR fixes:
Fixes #
Special notes for your reviewer:
Tested on scenarios blow:
Kubeup (scale up) with mizar: successfully except known pods issue (metrics-server, cordons... pods failure)
$ kubectl get pods -AT | grep mizar
system default mizar-daemon-pzmzn<PHONE_NUMBER>949531555 1/1 Running 0 43m
system default mizar-daemon-s777f<PHONE_NUMBER>091307921 1/1 Running 0 43m
system default mizar-daemon-xnxfb<PHONE_NUMBER>105442661 1/1 Running 0 43m
system default mizar-operator-5c97f7478d-dcbv8<PHONE_NUMBER>661054885 1/1 Running 0 43m
Kubeup (scale up with default network): successfully
Kubeup (scale up with default network) + kubemark ( scale up with default network): successfully
Does this PR introduce a user-facing change?:
https://github.com/CentaurusInfra/arktos/wiki/Mizar-Arktos-Integration-Release-2022-0130-Test-Plan
https://github.com/CentaurusInfra/arktos/wiki/Mizar-Arktos-Integration-Release-2022-0130-Test-Plan -- The tests for pods were all passed on 2022-02-24.
/lgtm
/approve
|
2025-04-01T04:54:49.550921
| 2016-09-23T20:31:39
|
178967215
|
{
"authors": [
"abought",
"samanehsan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13496",
"repo": "CenterForOpenScience/exp-addons",
"url": "https://github.com/CenterForOpenScience/exp-addons/pull/145"
}
|
gharchive/pull-request
|
Rating Form UI changes [LEI-213]
Refs: https://openscience.atlassian.net/browse/LEI-213
Changes
Increase spacing of radio buttons in questions 2, 4, & 6
Fix scales in questions 2, 6 & 9
Done responding to your comments/questions @abought! Let me know if I should change the rating labels to map to option values instead of indices 🍂
Done responding to your comments/questions @abought! Let me know if I should change the rating labels to map to option values instead of indices 🍂
I'll leave this slightly to your discretion- but maintaining two distinct sets of numbers associated with each item is slightly confusing, and seems like it could make the code harder to maintain. Somewhere at the back of my mind, this seems to hint that the code is duplicating more than necessary.
Let me know what you decide.
@abought, done responding! I changed the option labels to map to the option values to make that clearer. ☔
|
2025-04-01T04:54:49.554129
| 2016-11-06T23:51:37
|
187600355
|
{
"authors": [
"abought",
"samanehsan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13497",
"repo": "CenterForOpenScience/exp-addons",
"url": "https://github.com/CenterForOpenScience/exp-addons/pull/197"
}
|
gharchive/pull-request
|
Fix card text overflowing card
Refs: https://trello.com/c/e5LbqRT8/4-text-on-cards-in-sub-categories-runs-off-card
Purpose
On small screen widths, card text runs off of cards.
Summary of changes
Make buckets wider
Make cards wider
Make card text break in between long words
Looks like purely cosmetic changes here; no review needed.
This will look weird on small screens, but ultimately there's not much we can do if the column is narrower than the text.
|
2025-04-01T04:54:49.561231
| 2015-03-23T20:00:26
|
63818765
|
{
"authors": [
"chennan47",
"lbanner",
"lyndsysimon",
"sloria"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13498",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/pull/2289"
}
|
gharchive/pull-request
|
[feature] Multiple emails
Status: ready for review
@lyndsysimon Let's focus this PR on the limited multiple emails requirement i.e. ability to add emails to a user account that are non-existent, or unregistered/unconfirmed -- and then knock this out -- OK
1st pass finished. :point_up:
Ready for second pass.
2nd pass finished. :v:
Bah. For some reason, Travis isn't building this every push.
I've got my repo to build as well, and this passes: https://travis-ci.org/lyndsysimon/osf.io/builds/56771321
Response to 2nd pass complete. Your move, @sloria...
:game_die:
Pass finished. :tanabata_tree:
Third pass review complete.
:pager:ing @sloria.
Is there any way to resend a confirmation email?
Minor "bug": When adding an unconfirmed that is already in the list of unconfirmed emails, the growl message says "Email added" when it should probably say "Email already added" or something similar.
Bug: Whenever I input an unconfirmed email with an uppercase character, I get an error message "Email validation failed", but the email still gets added.
Doesn't have to be added before this is merged, but we might want to add a little helper text (e.g. with a question-mark button) to explain what it means to make an email "primary".
Pass finished. :deciduous_tree:
Some input from Trello re: the "Email added" message: https://trello.com/c/X8aJ9lC7/73-more-information-when-email-is-added
Also: https://trello.com/c/Udvre4jB/75-adding-an-email-for-a-confirmed-account-with-no-email-set
ed20e52 address https://trello.com/c/X8aJ9lC7/73-more-information-when-email-is-added
@sloria rdy for review
:+1:
|
2025-04-01T04:54:49.564785
| 2018-08-27T18:25:55
|
354425238
|
{
"authors": [
"erinspace",
"sloria"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13499",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/pull/8651"
}
|
gharchive/pull-request
|
Account for folders without relying on querying BaseFileNode for path [#PLAT-1053]
Purpose
Yet another addendum to #8643
Changes
Use OsfStorageFileNode (cuz BaseFileNode overrides path)
If path turns out to be a folder and a version is still there, don't error just don't do anything with the folder
QA Notes
Documentation
Side Effects
Ticket
https://openscience.atlassian.net/browse/PLAT-1053
Failing build is due to missing merge migration, which has been fixed on feature/storage-i18n.
|
2025-04-01T04:54:49.568029
| 2019-08-15T20:45:50
|
481331146
|
{
"authors": [
"UdayVarkhedkar",
"pattisdr"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13500",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/pull/9131"
}
|
gharchive/pull-request
|
[ENG-691] Enables gitlab addon to access all gitlab projects with user membership
Purpose
Enable the GitLab addon to access all GitLab projects that the user is a member of. Previously, the addon was limited to GitLab projects that the user is the owner of.
Changes
Modifies the function that retrieves projects to use the general purpose project endpoint with a membership requirement as opposed to using the user/projects endpoint.
QA Notes
Testing the GitLab add on and ensuring that users are able to access GitLab projects they own as well as GitLab projects that they do not own but are members of would be beneficial.
I tested this locally using two separate GitLab accounts.
Documentation
N/A
Side Effects
N/A
Ticket
https://openscience.atlassian.net/browse/ENG-691
LGTM 🚢
|
2025-04-01T04:54:49.573706
| 2017-08-04T20:04:52
|
248104216
|
{
"authors": [
"Johnetordoff",
"coveralls"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13501",
"repo": "CenterForOpenScience/waterbutler",
"url": "https://github.com/CenterForOpenScience/waterbutler/pull/244"
}
|
gharchive/pull-request
|
[SVCS-332][SVCS-292] Make Googledrive give folder children’s metadata on moves and copies
Purpose
On inter/intra moves/copies googledrive doesn't return the folder's children, this fixes that.
Changes
Changes item serializer to recursively call the google drive api include children to metadata.
Side Effects
None that I know of.
Ticket
https://openscience.atlassian.net/browse/SVCS-332
Old PR
https://github.com/CenterForOpenScience/waterbutler/pull/221
Coverage remained the same at 76.68% when pulling ea5986c48aa6f86a6ffe7661db40964f05c8a2a3 on Johnetordoff:Google-drive-return-children into b787a92c180697cb8fb2f5b6c402cf7b931b339d on CenterForOpenScience:develop.
Coverage remained the same at 78.562% when pulling 5f5b5b0a90105c9c6918b1922ca6259943cb76a5 on Johnetordoff:Google-drive-return-children into 761c430396730b72e600f8def98c06e4a5d0316b on CenterForOpenScience:develop.
Coverage decreased (-0.03%) to 78.536% when pulling 547db2b48a29e6455b93ccc9f386a776127eb91b on Johnetordoff:Google-drive-return-children into 761c430396730b72e600f8def98c06e4a5d0316b on CenterForOpenScience:develop.
|
2025-04-01T04:54:49.579793
| 2015-12-08T05:16:05
|
120933000
|
{
"authors": [
"Cerberus98"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13502",
"repo": "Cerberus98/lightshowpi",
"url": "https://github.com/Cerberus98/lightshowpi/issues/7"
}
|
gharchive/issue
|
Always On Mode
Rather than running from cron, we should have a way of running all the time. This would provide for alternate light configurations that don't rely on music to update lights. For example, it would be useful to only play songs at the top of the hour, but leave the lights on at the conclusion of the playlist (or configured subset of the playlist) and the only way to do that is to leave the process running and ensure the GPIO pins stay high
Somewhat wrong here: if you don't explicitly set the GPIO pins to low, they will stay in whatever state the song left them in. In other words, we don't need to run anything. A quick hack to get this up and running is to start a new show in cron, and ensure there are pre and post-cleanup scripts to set all the pins high so the lights stay on the rest of the time.
|
2025-04-01T04:54:49.583059
| 2023-01-10T09:19:27
|
1527037310
|
{
"authors": [
"jiacai2050"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13503",
"repo": "CeresDB/ceresdb",
"url": "https://github.com/CeresDB/ceresdb/issues/554"
}
|
gharchive/issue
|
Optimize local WAL size
Describe This Problem
In my local env, I find WAL(rocksdb based) is very large compared with SST, even after manually flush.
1.5G data/wal
76M data/manifest
371M data/store
1.9G data
Proposal
Do a compaction after delete_range to try remove dead entry from SST.
https://github.com/CeresDB/ceresdb/blob/9ab659a99417b41f3a4007e59bbe9b8f3ff65b8b/wal/src/rocks_impl/manager.rs#L120
Additional Context
https://github.com/facebook/rocksdb/wiki/RocksDB-Tuning-Guide#trigger-compaction-on-deletes
@ShiKaiWi I don't think this issue get fixed, wal encode is a general way to reduce wal size, but the issue here is to optimize for rocksdb-based, they have no direct relation.
|
2025-04-01T04:54:49.642806
| 2023-06-10T18:07:30
|
1751062738
|
{
"authors": [
"javagl",
"jo-chemla"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13504",
"repo": "CesiumGS/3d-tiles-tools",
"url": "https://github.com/CesiumGS/3d-tiles-tools/issues/43"
}
|
gharchive/issue
|
[Feature Request] TilesetCombiner fails for older tilesets with content adressed by url property instead of uri
When using the combine method on tilesets produced by ContextCapture, then an error Content does not have a URI is raised. This is probably because the tile contents in the tileset root json file (plus subsequent tiles) are referred to via the url property rather than the uri property - which references either b3dm or json tile contents.
I know I can use the 3d-tiles-tools upgrade utility on local tilesets I do manage - although it will take a lot of time to upgrade loads of tilesets.
But for tilesets I do not manage, which are stored online on different servers and that I'd like to combine, then the easiest way would be for the combine utility to allow for tilesets which use either the url or the uri property to reference tile content.
We have been pondering this question internally.
The pragmatic view is: All inputs have to be valid. And for the case of the uri vs. url, the recommendation would be: Run upgrade on the input, and then use the upgraded tilesets for any further operations.
But... I can see that this is... invonvenient, particularly when the reason for something not working properly is literally that of a single letter (or "a pixel", as in "the difference between i and l...).
One reason of why we are (currently) not just ignoring this point is that the generalization of this question could be: Which forms of "invalid" inputs should be "ignored"? Or: Which parts of the upgrade should be done autmatically in other operations? At some point, we'd have to spread workarounds for legacy data, handling for special cases, and pseudo-upgrade functionality throughout the code. And it would really be better to summarize this in the upgrade command.
(One example: When you combine two tilesets that use url - should the combined tileset use uri or url then? It should probably use uri, but maybe you want to retain the original url...?)
That being said: The case of the content url was a special one for me as well: The original (pre-refactoring) state of the 3d-tiles-tools had been using the (somewhat unspecified) Tileset sample data extensively for tests. This one uses url. I created a utility function to handle this more transparently. But at some point, I just created a TilesetWithUris from that, to not have to cope with this issue for some tests 😕
So I agree that we should consider to make the tools more resilient for the particular, very special case of uri-vs-url...
Hi again, and again thanks for all this context, very useful.
Understood for the fact that the upgrade utility would be the corner piece before any other workflow operation, which would therefore only have to deal with valid tilesets, simplifying the code a lot. Then indeed, the question of whether to perform such pseudo-upgrade within workflow operations could be important - and even generalized as a cli option for every workflow.
Good question regarding whether the output combined tileset should have a url or uri property. If the tile versioning is a property of each tile json, then since viewers can handle tilesets of different versions, I would say they can be mixed, but I don't know the intricacies of such a decision.
And it's true this uri-vs-url seem like a special corner-case. Being able to combine tilesets without the requirement for upgrading them would make it easier to source external tilesets (if served from a CORS enabled server) without the need to download/serve them and mix multiple sources for content.
Edit final questions related to tiling:
are utilities defined in Cesium3DTilesWriter meant to build a tool to tile arbitrary input meshes to 3D-tiles? Probably something that is used internally by the Cesium Ion pipeline but not meant to be open-sourced as a complete tool yet? All commercial photogrammetry suites have been offering 3D-tiles exporting capabilities for the past few years (RealityCapture, ContextCapture, Agisoft Metashape, etc), but no open-source suite does so at the moment (Colmap/Meshroom etc), probably because of that fact that there is no oss tiler for meshes existing yet, only for pointclouds.
final question: when tiling a pointcloud into a Potree dataset, I can specify the bounding box of the input pointcloud data which will be used as the box of the root tile. Therefore if I reuse the same bbox for tiling multiple pointclouds (where coords are expressed in the same Coordinate Reference System), and PotreeConvert them sequentially with the overwrite option, then it is equivalent to appending pointclouds to the same tiled Potree container. Is there any standard representation as to how to do this for 3d-tiles? This way, we could probably combine tilesets to a master, earth-covering tileset, where tilesets would be appended at a given level in the hierarchy without affecting any other portions of the tileset, and this would be really useful - a way for us to combine a few hundreds indepdent tilesets to a single, earth-wide tileset ala google3dtiles.
Understood for the fact that the upgrade utility would be the corner piece before any other workflow operation,
Ideally, doing an upgrade should never be necessary. It should only be a last-resort option for the case that somebody has a legacy tileset, and wants to try and salvage it. One difficulty is that it is tremendously hard to say which "legacy" elements can be updated (and how), and what exactly the result will be. One example is that of tilesets that contain glTF 1.0 data. It is sometimes simply not possible to update this to glTF 2.0, and whether it is possible depends on many low-level technical factors, and it would be hard or impossible to establish a reliable contract for the behavior of the upgrade in this case.
The url-vs-uri, however, is relatively simple. It only twiddles in the JSON, and can be upgraded with a few lines of code. (So it doesn't really need the broader infrastructure that is offerd by the general upgrade command). We'll still have to see where/how to integrate that step in the most most sensible way. The two "extremes" would be:
Hand this in the most fine-grained manner, on the fly. Instead of writing const uri = content.uri, we'd have to say const uri = Contents.getUri(content), where the latter is the utility function that returns the url from legacy content if necessary
Handle it as a coarse-grained, blanket step. Whenever a tileset.json is read, there could be one pass to PseudoUpgrade.doThatUrlToUriUpgradeIfNecessary(tilesetJson), to have one place where this upgrade happens if necessary, and all the remaining code can use the content.uri directly, without having to worry about the case that it might be a url
Both have pros and cons. We'll have to sort that out..
Regarding the last questions...:
are utilities defined in Cesium3DTilesWriter meant to build a tool to tile arbitrary input meshes to 3D-tiles? Probably something that is used internally by the Cesium Ion pipeline but not meant to be open-sourced as a complete tool yet?
The Cesium3DTilesWriter in particular is just intended to have a mechanism for reading/writing the tileset JSON data. This code is auto-generated from the schema. Users could use this, for example, when they implement a tool that converts meshes to 3D Tiles. Considering the complexity of such a tool, the part that is responsible for writing the JSON would only be a minor, minor part, but ... at least, users wouldn't have to write that part from scratch.
final question: when tiling a pointcloud into a Potree dataset, ...
Sorry, I'm lacking a lot of context here. It sounds like this might be related to external tilesets in general, and maybe something that involves additive refinement. But I don't know Potree well enough to say more here. You might consider bringing this up in the forum at https://community.cesium.com/c/3d-tiles/16 , maybe someone with knowledge about Potree and a better understanding of your goals could chime in there.
Again thanks for all these details, very useful to understand the state of this toolset. And indeed probably not the right place to share these thoughts - the way I did it here is because from my point of view this could be what merge/combine are meant to do - combining tilesets, merging two or more hierarchical structures keeping only the lowest error/highest detail data.
You're right I'll iterate over these thoughts and ask along the way to the community forum instead if I need some feedback. Thanks again for all your help!
As I mentioned, there are some thoughts about possible extensions of the tools in the future. These thoughts differ in how likely it is for them to become actual points on a roadmap.
One example is that we have 'combine' (which creates one large tileset from one that had external tilesets), but we don't have an opposite of that - i.e. there is no "split" function. This could, in some way, be a low-hanging fruit: Just traverse the tileset, and whenever you reach depth x, start writing out a new one.
Now we could go and implement that. But...
nobody needed it until now (i.e. there was no feature request for that)
there is some overlap to concepts of implicit tilesets (the x would be the subtreeLevels, so to speak)
maybe most importantly: This could be generalized ... abritrarily
And before starting something like that, one should at least have a rough idea about how it could be generalized. For example: People might not want to split their tileset into "slices" with x levels each, but maybe based on some geometricError-threshold`, or maybe based on the file/data size of the results...
But if you have ideas (or even specific demand) for a certain functionality, just let us know... (or... open a PR, of course...)
|
2025-04-01T04:54:49.651507
| 2022-04-07T02:56:04
|
1195454987
|
{
"authors": [
"kring",
"nithinp7",
"nzanepro"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13505",
"repo": "CesiumGS/cesium-unreal",
"url": "https://github.com/CesiumGS/cesium-unreal/issues/815"
}
|
gharchive/issue
|
UE5 + Vehicle = exaggerated motion blur
If you create a vehicle template in UE5 and drive the car and notice the wheel motion blur looks fine in the template map
If you then add the Cesium plugin (CesiumForUnreal-500-v1.12.1-ue5 - windows) and create a blank scene, add Cesium World Terrain + Cesium SunSky and then place the car on a city street in Denver area (default area) the motion blur on the tires go crazy. Everything still has default project settings from the vehicle template project.
If this is some setting I'm missing somewhere, please let me know.
I went back and the wheel motion blur is terrible regardless of Cesium being involved in UE4.27
I don't recall this in older versions of Cesium.
I tried various settings of anti-aliasing to no effect in the Cesium map.
same behavior with first person template. its fine until walking on cesium generated terrain. Maybe this is the physics bodies that are being created?
@nzanepro Is there a drop in frame rate when adding the Cesium tilesets? I am not sure how motion blur is computed, but I wonder if it varies based on the delta time between frames.
The CitySample has a better example of car tire motion blur trickery, which works well with Cesium tileset.
The first person template still exhibits this difference between walking on a Geometric "Plane" or "Cube" object and the ground provided by the Cesium World Terrain.
Is the Cesium Physics geometry able to be presented as a StaticMesh to the Physics system?
Cesium tiles are simply static meshes, though they're created and destroyed in a view-dependent manner. I think Nithin's question about frame rate might be highly relevant. Were you able to confirm a frame rate drop?
ok, I figured out something that is reproduceable for you.
install cesium samples project into ue5.01
navigate to 03_CesiumDenver Level
delete Aerometrex Denver photogrammetry from outliner
play
you will start on the "StartPlatform"
use the "a" button to run to the left
notice the motion blur while on the "StartPlatform" vs the motion blur when you drop to the cesium tiles and continue running to the left.
my fps doesn't seem to drop in the same scene, same settings, but different geometry?
video capture attached including fps displayed.
machine stats:
Intel(R) Core(TM) i9-9900KF CPU @ 3.60GHz 32GB memory
NVIDIA GeForce RTX 2080ti 11GB gpu memory
https://user-images.githubusercontent.com/5840082/167763438-ab1f6a9c-845f-42e1-b439-b01773c64b99.mp4
same steps repeated in ue4.27 for reference. motion blur seems to be fine.
https://user-images.githubusercontent.com/5840082/167765248-58f3e64b-5a8c-476f-8c41-490d01ecb6bc.mp4
|
2025-04-01T04:54:49.656228
| 2024-05-07T15:51:33
|
2283702281
|
{
"authors": [
"j9liu",
"kring"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13506",
"repo": "CesiumGS/cesium-unreal",
"url": "https://github.com/CesiumGS/cesium-unreal/pull/1416"
}
|
gharchive/pull-request
|
Add check to disable LOD transitions in material layer
Follow up to #1414. Fixes #1104.
This PR does the following:
Renames UseLodTransitions to EnableLodTransitions, to keep consistent with the rest of the naming in Cesium3DTileset.
Adds a EnableLodTransitions parameter to the ML_DitherFade material layer, to control whether or not the dithering is applied.
Clarifies that EnableLodTransitions is only compatible with temporal AA modes in the comments.
This material parameter corresponds exactly with the Cesium3DTileset's parameter. When false, dithering won't be applied at all. This helps for projects that use non-temporal AA modes. As long as they don't enable LOD transitions (which would be user error), they don't need to worry about these dithering artifacts on their tilesets.
Before
After
Hm, the artifacts don't appear when Forward Rendering is disabled, even for non-temporal AA modes. So I'll adjust the documentation
@j9liu I see pretty significant artifacts in this branch when "Enable LOD Transitions" is enabled. It kind of looks like the old tile is being removed completely before the new one comes in, so I see big holes in the surface. This is with Google Photorealistic 3D Tiles:
I see the same thing with CWT+Bing, though.
|
2025-04-01T04:54:49.746095
| 2016-08-25T15:53:39
|
173241589
|
{
"authors": [
"CassandraCat",
"Huyston",
"badg0003",
"bagnell",
"denverpierce",
"dukeofcool199",
"emackey",
"expatiating",
"ggetz",
"hpinkos",
"lilleyse",
"mramato",
"pjcozzi",
"shunter"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13507",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/issues/4235"
}
|
gharchive/issue
|
Poor billboard quality
Reported on the forum: https://groups.google.com/d/msg/cesium-dev/OLbOuzNRM7U/MHgo_4N6CwAJ
The icon to the left is a billboard, the icon to the right is the original image
I believe this happens because billboards are not aligned to the nearest pixel. When a billboard has a non-integer screen location, it becomes blurred by the fullscreen AA process.
There's a suggested fix on Stack Overflow that involves making a one-line change to BillboardCollectionVS.glsl#L245, to change it from:
gl_Position = czm_viewportOrthographic * vec4(positionWC.xy, -positionWC.z, 1.0);
to:
gl_Position = czm_viewportOrthographic * vec4(floor(positionWC.xy + 0.5), -positionWC.z, 1.0);
This has the effect of greatly reducing blur on both Billboards and Labels.
What do you guys think of this? If you like it, I can make a pull request out of it (assuming we can take code from SO. I presume most projects already have quite a bit of uncredited SO code around).
The reasoning seems clear: Billboards are always raster images, so non-integer screen locations will always be worse quality than pixel-aligned locations.
Ah cool, can you post a screenshot of the difference?
The Stack Overflow answer has a good screenshot:
http://stackoverflow.com/questions/33784256/cesium-label-blurred/33786151#33786151
If this feels like deja-vu, you're right. Here's where clamp-to-pixel functionality was added: #188 Here's where it was removed again: #565.
Good memory @shunter. In experimenting with this code, it looks like the clamp-to-pixel is not what's actually improving this guy's screenshots. He's also turning off scene.fxaa at the same time, and that appears to be what actually cleans up the label.
Yeah, fxaa is probably the real culprit here. Clamping to pixel actually does not do what we would hope it would do (Especially with text). It also causes all kinds of problem when during motion (either because the item is moving or the camera is moving). Not clamping really is the "correct" option.
I asked Dan about possibly not running fxaa on labels and billboards and he things it might be possible but might require too much overhead or other craziness.
I just noticed he's also provided us with an svg, that's probably compounding the problem because the svg is properly be anti-aliased on rasterization by the browser (and depending on the browser it could be doing a bad job).
@hpinkos can you reach out and get the original icon? I didn't notice it in that thread.
I asked @bagnell offline how our fxaa is actually implemented. He suggested this was based on an older approximation where high-contrast edges are intentionally blurred, which is fast enough for realtime, but is a very different implementation from true anti-aliasing. There was talk of this possibly being upgraded to a different algorithm when WebGL 2 becomes more widely available.
So it sounds like the blurring is a design feature of the current fxaa system, making it almost unavoidable without a rewrite or disabling of fxaa.
I don't think it is out of the question to render to one framebuffer that would get FXAA, and another for labels/billboards/points that would not, and then composite them. This would also work well with the need to render labels on top of polygons since the low level-renderer now would have explicit-ish knowledge of labels/billboards/points.
This will be a non-trivial amount of time though so we can't do it for a few months.
Here is a png and svg version of the icon from the user. He says the png looks better than the svg
icon-library.svg.txt
Here is a png and svg version of the icon from the user. He says the png looks better than the svg
This is exactly what I assumed a large part of the problem was. Browsers do not all do an equally good job of rasterizing/resizing SVGs, so the fact that it's an SVG is one of the reasons it looked so blurry.
True, but we still have big problem with fxaa intentionally blurring our billboards and labels. I think @pjcozzi's comment is the path to persue, when time permits.
I think everyone agrees @emackey I just wanted to point out that SVGs can compound the issue.
Related to #2752 and #3279 . If you're considering separate frame buffers, being able to selectively apply it to imagery tiles as well would be helpful. An AA pass on tiles, especially those with labels, reduces quality.
Also reported here: https://groups.google.com/forum/?hl=en#!topic/cesium-dev/NBfMvj80vGI
I investigated this at the bug bash. It would be quite a bit of work to selectively apply anti-aliasing. Here is the post process that would happen after each frustum instead of on the entire scene:
@bagnell can you post a link to the branch-in-progress and before/after screenshots?
The branch is selective-fxaa.
Before:
After:
Another idea is to render pick ids with MRT (or a separate pass to an offscreen framebuffer/texture), and then only blur pixels where the surrounding pick ids are different so only the ends of objects are blurred. Billboards, the globe texture, etc. would not be blurred except at the silhouettes.
Would these fixes apply to #3279 as well?
@denverpierce they would be part of the solution.
I believe that this is still an issue - has there been any additional thought on this topic in the last ~5 years?
Or is there a recommended approach to solve with billboards and text overlays?
Quality issue when moving reported in https://community.cesium.com/t/billboard-with-custom-canvas-image-gets-darker-brighter-on-scrolling-rotating/25221/2.
I'm still having this issue.
Here is an example: sandcastle
I'm trying to create many billboards with a rectangle texture. I tried supplying an image with the "pixelated" style to prevent the image from blurring. Note that the image on the top is not blurred, but the rendered rectangle in Cesium is blurred.
Also tried disabling fxaa and antialias, without success.
Any further investigation into this issue? I can see that even in the sandcastle examples the billboard quality is low as well.
Thanks for the interest. There hasn't been any activity on this item recently, but it is definitely on our radar.
for anyone else that has found this thread, my co-worker came across this other thread explaining how billboards were made intentionally blurry to help with running on mobile devices. and that viewing billboards at a browser zoom level of anything other than 100% will cause this issue to become more noticeable.
The workaround from the thread states to adjust the value of viewer.resolutionScale property
I tested this fix in this sandcastle with some success.
@dukeofcool199 Thanks big time for this. I've been searching and made too many attempts to fix this than I want to admit lol, but your tip did the trick! :)
for anyone else that has found this thread, my co-worker came across this other thread explaining how billboards were made intentionally blurry to help with running on mobile devices. and that viewing billboards at a browser zoom level of anything other than 100% will cause this issue to become more noticeable. The workaround from the thread states to adjust the value of viewer.resolutionScale property I tested this fix in this sandcastle with some success.
UPDATE:
change this setting for the viewer viewer.useBrowserRecommendedResolution to be false
nice bro.
|
2025-04-01T04:54:49.750178
| 2019-05-28T07:23:01
|
449103486
|
{
"authors": [
"OmarShehata",
"hst-m",
"jdarpinian",
"nlucis",
"verybigzhouhai"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13508",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/issues/7871"
}
|
gharchive/issue
|
enableLighting = true causes black globe on Android
On my Pixel 2 XL in Chrome 74, setting enableLighting to true causes the globe to render as completely black. I don't see any errors in the console. It seems to be an issue with the combination of showGroundAtmosphere and enableLighting, as disabling either of them brings the globe back.
Can be reproduced in this sandcastle: https://cesiumjs.org/Cesium/Build/Apps/Sandcastle/?src=Ground Atmosphere.html
Thanks for reporting this @jdarpinian. I can reproduce this on my Pixel 3 in Chrome as well. I can also see that zooming in enough once you get to the fade out distance will trigger the globe to render again. But I don't see a separate code path that triggers when this fade reaches 1 or 0 in this shader:
https://github.com/AnalyticalGraphicsInc/cesium/blob/master/Source/Shaders/GlobeFS.glsl#L286
@bagnell any idea here?
@OmarShehata @jdarpinian I guess this is due to the structure, because this will not be reproduced on the pc platform, ios platform, Huawei mobile phone. the platform using Qualcomm processor seems to have this problem, I am removing all the structs, when I finish After working, I can confirm whether to solve this problem. #7651
This issue still occurs on android chrome, is it possible to fix?
Still presenting as an issue on both iPad OS and the latest Android version on a OnePlus 9 Pro.
|
2025-04-01T04:54:49.767370
| 2020-10-12T22:29:50
|
719710019
|
{
"authors": [
"OmarShehata",
"cesium-concierge",
"lilleyse",
"shelterit"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13509",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/pull/9197"
}
|
gharchive/pull-request
|
Update Model.js
Fix for issue https://github.com/CesiumGS/cesium/issues/9169
Added a simple test for those cases when commands are dynamic in a model
Thank you so much for the pull request @shelterit! I noticed this is your first pull request and I wanted to say welcome to the Cesium community!
The Pull Request Guidelines is a handy reference for making sure your PR gets accepted quickly, so make sure to skim that.
:x: Missing CONTRIBUTORS.md entry.
Please add yourself to the contributors file!
:x: Missing CLA.
Please send in a Contributor License Agreement (CLA) and comment back here to let us know to check this!
:grey_question: CHANGES.md was not updated.
If this change updates the public API in any way, please add a bullet point to CHANGES.md.
:grey_question: Unit tests were not updated.
Make sure you've updated tests to reflect your changes, added tests for any new code, and ran the code coverage tool.
Reviewers, don't forget to make sure that:
[ ] Cesium Viewer works.
[ ] Works in 2D/CV.
[ ] Works (or fails gracefully) in IE11.
Thanks for opening a PR Alex! I know you mentioned testing for unintended consequences, but I'm hesitant to merge when we don't have a way to verify this fix. It may not cause any adverse effects but it may push the error further down the line since the underlying cause is not fixed. Why is an undefined command being pushed here in the first place? Do you know where it originates?
Hey,
That's a good question, however I don't have the internal knowledge of
Cesium to answer how that could happen. :) I'll see if I can talk to the
customer if they can share that piece of data with us, although I'm not
sure (they're a bit particular, hush, hush). I can probably get the source
files, and find out what they used to tile it, maybe there's a race
condition in there somewhere, and hopefully at the least try to make a
different model to replicate the issue.
If I go through the original file, what should I be looking for? It's all
linked to whether a property is present in the right place, and possibly
the tiler / somewhere didn't inject the property at some place? Again, I'm
not smart enough to know what goes into the opacity command being sent and
what that stack looks like.
We've currently got this fix in our build stream, but it would be great to
get it out, of course, to keep things "pure". Do you have a test suite of
data for every build that you could test the 'fix' against?
Cheers,
Alex
On Mon, Nov 2, 2020 at 4:11 AM Omar Shehata<EMAIL_ADDRESS>wrote:
Thanks for opening a PR Alex! I know you mentioned testing for unintended
consequences, but I'm hesitant to merge when we don't have a way to verify
this fix. It may not cause any adverse effects but it may push the error
further down the line since the underlying cause is not fixed. Why is an
undefined command being pushed here in the first place? Do you know where
it originates?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/CesiumGS/cesium/pull/9197#issuecomment-720120660, or
unsubscribe
https://github.com/notifications/unsubscribe-auth/AAFHK5XFH55QDRXJR4EN4O3SNWJDRANCNFSM4SNQGIYA
.
--
Information Alchemist / UX consultant / GUI developer for hire
http://thinkplot.org | http://www.linkedin.com/in/shelterit
Thanks again for your contribution @shelterit!
No one has commented on this pull request in 30 days. Maintainers, can you review, merge or close to keep things tidy?
I'm going to re-bump this in 30 days. If you'd like me to stop, just comment with @cesium-concierge stop. If you want me to start again, just delete the comment.
Thanks again for your contribution @shelterit!
No one has commented on this pull request in 30 days. Maintainers, can you review, merge or close to keep things tidy?
I'm going to re-bump this in 30 days. If you'd like me to stop, just comment with @cesium-concierge stop. If you want me to start again, just delete the comment.
Thanks again for your contribution @shelterit!
No one has commented on this pull request in 30 days. Maintainers, can you review, merge or close to keep things tidy?
I'm going to re-bump this in 30 days. If you'd like me to stop, just comment with @cesium-concierge stop. If you want me to start again, just delete the comment.
Thanks again for your contribution @shelterit!
No one has commented on this pull request in 30 days. Maintainers, can you review, merge or close to keep things tidy?
I'm going to re-bump this in 30 days. If you'd like me to stop, just comment with @cesium-concierge stop. If you want me to start again, just delete the comment.
Thanks again for your contribution @shelterit!
No one has commented on this pull request in 30 days. Maintainers, can you review, merge or close to keep things tidy?
I'm going to re-bump this in 30 days. If you'd like me to stop, just comment with @cesium-concierge stop. If you want me to start again, just delete the comment.
A similar fix was made in https://github.com/CesiumGS/cesium/pull/9271 so I'm going to close this one. Thanks @shelterit.
|
2025-04-01T04:54:49.771323
| 2021-06-10T04:04:47
|
916868623
|
{
"authors": [
"cesium-concierge",
"ptrgags",
"sanjeetsuhag"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13510",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/pull/9607"
}
|
gharchive/pull-request
|
Fixes S2 bounding volumes
This PR fixes some two main issues in the current S2 implementation:
Incorrect conversion from {level}/{x}/{y}/{z} to Hilbert/S2 index.
The parent bounding volume being updated when a child bounding volume is created.
Thanks for the pull request @sanjeetsuhag!
:heavy_check_mark: Signed CLA found.
:grey_question: CHANGES.md was not updated.
If this change updates the public API in any way, please add a bullet point to CHANGES.md.
Reviewers, don't forget to make sure that:
[ ] Cesium Viewer works.
[ ] Works in 2D/CV.
[ ] Works (or fails gracefully) in IE11.
@sanjeetsuhag I'll merge this once CI passes
merged! Thanks @sanjeetsuhag!
|
2025-04-01T04:54:49.773322
| 2024-01-28T02:00:21
|
2103915445
|
{
"authors": [
"CetoBasilius"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13511",
"repo": "CetoBasilius/energy-defender",
"url": "https://github.com/CetoBasilius/energy-defender/pull/20"
}
|
gharchive/pull-request
|
Implemented energy mechanism and UI
Closes #5 and closes #6
Added energy mechanism on GameManager
Added Panel component and energy panel
Implemented energy spend mechanism, locking expensive towers
Added and implemented UIManager
Removed old references to ScrollableCamera to UIManager
Renamed TowerDrag to TowerIcon
Moved things around so folders make more sense now in line with the GDD
No code reviewers are available so will go ahead and merge this
|
2025-04-01T04:54:49.810713
| 2022-09-02T10:08:33
|
1359983566
|
{
"authors": [
"LesnyRumcajs"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13512",
"repo": "ChainSafe/forest",
"url": "https://github.com/ChainSafe/forest/issues/1871"
}
|
gharchive/issue
|
Rubify health check script
Issue summary
Currently, the health check script is written in bash. For readability and extendability and having a uniform technology in the sync check, it would be great to rewrite it in Ruby. The current script doesn't do any magic, just grabs the metrics and compares a few values there, so it should be straightforward to implement.
It might be an excellent entry task to get exposed to Ruby.
Task summary
[ ] rewrite health check script in Ruby
[ ] verify that it works the same way as the former bash script (perhaps some basic unit tests?)
[ ] use it in sync check CI
Acceptance Criteria
[ ] bash script is no longer used in sync check
[ ] new version has at least the same functionality as the former one
Other information and links
https://github.com/ChainSafe/forest/pull/1867
Bash is good, and the sync check is rock solid.
|
2025-04-01T04:54:49.838816
| 2023-01-12T09:13:55
|
1530358955
|
{
"authors": [
"LesnyRumcajs",
"hanabi1224",
"lemmih",
"tyshko5"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13513",
"repo": "ChainSafe/forest",
"url": "https://github.com/ChainSafe/forest/issues/2420"
}
|
gharchive/issue
|
Lint: pedantic clippy warnings
Issue summary
To further increase code quality, it would be interesting to look into pedantic clippy group.
For each lint:
add your GH handle to the lint so that two people are not working on the same lint,
figure out if it's helpful and makes sense in our case,
add it to the lint target in Makefile https://github.com/ChainSafe/forest/blob/d8eae69e556038af86e32a388e7e118b3bd3307d/Makefile#L73-L74
fix the warnings if any
tick the corresponding checkbox,
After they are all resolved, enable the entire pedantic group with perhaps some exceptions.
:warning:
This is a low-priority issue; feel free to tackle a lint or two now and then if there's nothing more important to do.
Lint list (may differ in time & clippy version):
[ ] bool_to_int_with_if
[ ] borrow_as_ptr
[ ] case_sensitive_file_extension_comparisons
[ ] cast_lossless
[ ] cast_possible_truncation
[ ] cast_possible_wrap
[ ] cast_precision_loss
[ ] cast_ptr_alignment
[ ] cast_sign_loss
[ ] checked_conversions
[ ] cloned_instead_of_copied
[ ] copy_iterator
[ ] default_trait_access
[ ] doc_link_with_quotes
[ ] doc_markdown
[ ] empty_enum
[ ] enum_glob_use
[ ] expl_impl_clone_on_copy
[ ] explicit_deref_methods
[ ] explicit_into_iter_loop
[ ] explicit_iter_loop
[ ] filter_map_next
[ ] flat_map_option
[ ] float_cmp
[ ] fn_params_excessive_bools
[ ] from_iter_instead_of_collect
[ ] if_not_else - @jdjaustin
[ ] implicit_clone
[ ] implicit_hasher
[ ] inconsistent_struct_constructor
[ ] index_refutable_slice
[ ] inefficient_to_string
[ ] inline_always
[ ] invalid_upcast_comparisons
[ ] items_after_statements
[ ] iter_not_returning_iterator
[ ] large_digit_groups
[ ] large_stack_arrays
[ ] large_types_passed_by_value
[ ] linkedlist
[ ] macro_use_imports
[ ] manual_assert
[ ] manual_instant_elapsed
[ ] manual_let_else
[ ] manual_ok_or
[ ] manual_string_new
[ ] many_single_char_names
[ ] map_unwrap_or
[ ] match_bool
[ ] match_on_vec_items
[ ] match_same_arms
[ ] match_wild_err_arm
[ ] match_wildcard_for_single_variants
[ ] maybe_infinite_iter
[ ] mismatching_type_param_order
[ ] missing_errors_doc
[ ] missing_panics_doc
[ ] module_name_repetitions
[ ] must_use_candidate
[ ] mut_mut
[ ] naive_bytecount
[ ] needless_bitwise_bool
[ ] needless_continue
[ ] needless_for_each
[ ] needless_pass_by_value
[ ] no_effect_underscore_binding
[ ] option_option
[ ] ptr_as_ptr
[ ] range_minus_one
[ ] range_plus_one
[ ] redundant_closure_for_method_calls
[x] redundant_else
[ ] ref_binding_to_reference
[ ] ref_option_ref
[ ] return_self_not_must_use
[ ] same_functions_in_if_condition
[ ] semicolon_if_nothing_returned
[ ] similar_names
[ ] single_match_else
[ ] stable_sort_primitive
[ ] string_add_assign
[ ] struct_excessive_bools
[ ] too_many_lines
[ ] transmute_ptr_to_ptr
[ ] trivially_copy_pass_by_ref
[ ] unicode_not_nfc
[ ] unnecessary_join
[ ] unnecessary_wraps
[ ] unnested_or_patterns
[ ] unreadable_literal
[ ] unsafe_derive_deserialize
[x] unused_async
[ ] unused_self
[ ] used_underscore_binding
[ ] verbose_bit_mask
[ ] wildcard_imports
[ ] zero_sized_map_values
Other information and links
https://github.com/ChainSafe/forest/pull/2428
To re-iterate two important points:
Don't work on this if you have other work to do.
Carefully consider whether the lint is beneficial. Include your arguments for why the lint should be enabled in your PR. Expect a lot of push-back so make sure your arguments are strong.
How about unwrap_used ? So that we enforce using .expect with a reason.
There has a allow-unwrap-in-tests setting in clippy.toml so that we don't need to fix all tests
How about unwrap_used ? So that we enforce using .expect with a reason.
There has a allow-unwrap-in-tests setting in clippy.toml so that we don't need to fix all tests
@jdjaustin Maybe you could have a go at this one. Coming up with sensible error messages might be tough, though.
Stale, we have more priorities than pedantic clippy warnings now. :)
|
2025-04-01T04:54:49.853474
| 2023-02-13T21:43:51
|
1583113541
|
{
"authors": [
"LesnyRumcajs",
"jdjaustin",
"lemmih"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13514",
"repo": "ChainSafe/forest",
"url": "https://github.com/ChainSafe/forest/pull/2540"
}
|
gharchive/pull-request
|
Changes to walk_snapshot to match Forest and Lotus snapshots
Summary of changes
Changes introduced in this pull request:
Changes to walk_snapshot to match Lotus logic (i.e., don't include identity CIDs in snapshot export).
Reference issue to close (if applicable)
Closes #1884
Other information and links
Change checklist
[x] I have performed a self-review of my own code,
[x] I have made corresponding changes to the documentation,
[x] I have added tests that prove my fix is effective or that my feature works (if possible),
[x] I have made sure the CHANGELOG is up-to-date. All user-facing changes should be reflected in this document.
Calibnet snapshots are matching up now but I'm so far unable to verify mainnet. @LesnyRumcajs
@jdjaustin, what's the issue with mainnet snapshots? Is it because of the hardware limitations? If so, have you tried installing Forest with an alternative, turbo fast, pioneer & bleeding edge backend with make install-with-jemalloc
@jdjaustin, what's the issue with mainnet snapshots? Is it because of the hardware limitations? If so, have you tried installing Forest with an alternative, turbo fast, pioneer & bleeding edge backend with make install-with-jemalloc
Yes, I tried make install-with-jemalloc and I'm still getting No space left on device errors. I can clear out some disk space and try again.
Ah, so your issue is with disk space, not RAM. In this case, you need around 250G of free disk space.
clean up forest directory
download and import the snapshot - this would require around 250G at least - 110G snapshot + 130G DB, more or less.
Delete the snapshot to get back some disk space.
Export the snapshot - again, you will need around 250G plus some more because of the additional chain progression.
@jdjaustin If you still don't have enough disk space after deleting the old files, ask Hubert to set up a droplet for you.
Good news. Verified that the Lotus and Forest mainnet snapshots also match!
@jdjaustin does the shasum of both match?
@jdjaustin does the shasum of both match?
Yepp!
Rock solid! Now it would be fantastic to keep it this way. One approach would be to modify the test a bit in our workflow https://github.com/ChainSafe/forest/blob/main/.github/workflows/rust.yml#L150-L172
Download the Lotus snapshot.
Import it into Forest
Export the snapshot at height specified in Lotus snapshot.
Compare the checksums.
We can do it in a separate issue if it's too complex; I may be missing some gotchas.
Linking the "source" PR for future reference if needed. https://github.com/filecoin-project/lotus/pull/8691
@jdjaustin Perhaps we should match the logic in Lotus for the snapshot, just in case. I am not entirely sure we need it but better safe than sorry. So this:
// We only include raw and dagcbor, for now.
// Raw for "code" CIDs.
switch prefix.Codec {
case cid.Raw, cid.DagCBOR:
default:
continue
}
Should also get included.
@jdjaustin Perhaps we should match the logic in Lotus for the snapshot, just in case. I am not entirely sure we need it but better safe than sorry. So this:
// We only include raw and dagcbor, for now.
// Raw for "code" CIDs.
switch prefix.Codec {
case cid.Raw, cid.DagCBOR:
default:
continue
}
Should also get included.
This block of code doesn't seem to be affecting snapshot exports currently, so my biggest question/concern with adding this is: how will I test the Rust code is working as intended?
I believe just checking if the snapshots are still the same will be okay.
I believe just checking if the snapshots are still the same will be okay.
Got it, I thought it was just going to be dead code, but I was just confused by the Go code; once I started writing the Rust code, I understood what was going on. And the snapshots still match:
|
2025-04-01T04:54:49.858134
| 2024-01-10T16:12:09
|
2074703947
|
{
"authors": [
"nflaig",
"wemeetagain"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13515",
"repo": "ChainSafe/lodestar",
"url": "https://github.com/ChainSafe/lodestar/pull/6279"
}
|
gharchive/pull-request
|
chore: update url of goerli bootnodes file and genesis ssz
Motivation
Files have been removed from eth-clients/eth2-networks, see https://github.com/eth-clients/eth2-networks/pull/92
Running Lodestar with --network goerli flag causes the following error on startup
Error fetching latest bootnodes: HTTPError: Response code 404 (Not Found)
at Request.<anonymous> (/home/nico/projects/ethereum/lodestar/node_modules/got/dist/source/as-promise/index.js:118:42)
at processTicksAndRejections (node:internal/process/task_queues:95:5)
Description
Update url of goerli bootnodes file and genesis ssz
:tada: This PR is included in v1.14.0 :tada:
|
2025-04-01T04:54:49.861361
| 2024-03-19T07:16:02
|
2194212576
|
{
"authors": [
"tuyennhv",
"wemeetagain"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13516",
"repo": "ChainSafe/lodestar",
"url": "https://github.com/ChainSafe/lodestar/pull/6563"
}
|
gharchive/pull-request
|
fix: improve state serialization
Motivation
State serialization happens once per epoch in n-historical state so its performance is important
Description
New validator type to optimize its value_serializeToBytes performance, this helps make state.validators.serialize() 3.4x faster
before
✔ serialize 20000 validators manually 773.4012 ops/s 1.292990 ms/op - 45968 runs 60.0 s
✔ serialize 20000 validators from state 206.3978 ops/s 4.845013 ms/op - 12271 runs 60.0 s
after
✔ serialize 20000 validators manually 768.2498 ops/s 1.301660 ms/op - 45667 runs 60.0 s
✔ serialize 20000 validators from state 694.0964 ops/s 1.440722 ms/op - 41260 runs 60.0 s
use the new serializeToBytes() api of ssz v0.15.1
part of #5968
:tada: This PR is included in v1.18.0 :tada:
|
2025-04-01T04:54:49.873908
| 2023-10-25T08:28:12
|
1960840729
|
{
"authors": [
"entron",
"willydouhard"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13517",
"repo": "Chainlit/chainlit",
"url": "https://github.com/Chainlit/chainlit/issues/503"
}
|
gharchive/issue
|
Table does not parse correctly if there is a title above it.
I am using the latest version and the following code to test:
@cl.on_chat_start
async def main():
msg = """Table 1
| heading | b | c | d |
| - | :- | -: | :-: |
| cell 1 | cell 2 | 3 | 4 |
"""
await cl.Message(content=msg).send()
The output is not parsed as a table:
Table 1
| heading | b | c | d |
| - | :- | -: | :-: |
| cell 1 | cell 2 | 3 | 4 |
Another maybe related problem is if there is markdown list inside a cell. This list will not be parsed either.
I am able to reproduce with your example. However this works
import chainlit as cl
@cl.on_chat_start
async def main():
msg = """
Hello
| heading | b | c | d |
| - | :- | -: | :-: |
| cell 1 | cell 2 | 3 | 4 | """
await cl.Message(content=msg).send()
Looks like indentation is important.
Interesting. The problem is the table and text above and below it are generated by LLM and I can not easily control indentation. Is it possible we make the table parsing more robust?
That is a good question. We use remark-gfm to parse markdown tables. Would have to dig into that.
Do you know what does ChatGPT and Gradio use? At least ChatGPT can produce beatiful tables robustly.
|
2025-04-01T04:54:49.876257
| 2017-11-09T16:23:29
|
272625296
|
{
"authors": [
"AncientFungus",
"NinjaDanz3r"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13518",
"repo": "Chainsawkitten/LargeGameProjectEngine",
"url": "https://github.com/Chainsawkitten/LargeGameProjectEngine/pull/791"
}
|
gharchive/pull-request
|
Continuation puzzle scene 1
Pull request by Jens Lomander and Andreas Flöjt
You can now throw a rock in the first puzzle. Better than showing nothing doodz
SonarQube analysis reported 1 issue
Note: The following issues were found on lines that were not modified in the pull request. Because these issues can't be reported as line comments, they are summarized here:
AncientFungus:LGEPR: Cppcheck cannot find all the include files (use --check-config for details)
|
2025-04-01T04:54:49.881274
| 2017-12-20T19:03:06
|
283664045
|
{
"authors": [
"kingdavidmartins"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13519",
"repo": "Chalarangelo/30-seconds-of-code",
"url": "https://github.com/Chalarangelo/30-seconds-of-code/pull/281"
}
|
gharchive/pull-request
|
Updating PULL_REQUEST_TEMPLATE.md -> Lodash Backlog
Description
MAKING MAINTAINABLE #100
Makes tracking the methods that have been UPDATED or ADDED to the lodash backlog much more easier. Than having the collaborators constantly reading all the snippets and trying to find which has been added or not.
This way all we have to do is just look on lodash backlog timeline and whichever methods.md PR have been merged we can update
What does your PR belong to?
[ ] Website
[ ] Snippets
[x] General / Things regarding the repository (like CI Integration)
Types of changes
[x] Bug fix (non-breaking change which fixes an issue)
[ ] Enhancement (non-breaking improvement of a snippet)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Checklist:
[ ] My code follows the code style of this project.
[x] My change requires a change to the documentation.
[x] I have updated the documentation accordingly.
[x] I have checked that the changes are working properly
[x] I have checked that there isn't any PR doing the same
[x] I have read the CONTRIBUTING document.
Yes Will surely do
|
2025-04-01T04:54:49.906129
| 2024-08-07T21:57:31
|
2454422743
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13520",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/12557"
}
|
gharchive/pull-request
|
General Grievance: Watch content-baer.de
General Grievance requests the watch of the watch_keyword content-baer\.de. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
content-baer\.de has been seen in 0 true positives, 0 false positives, and 0 NAAs.
Approved by Jeff Schaller in Charcoal HQ
|
2025-04-01T04:54:49.907948
| 2018-10-08T11:57:21
|
367755827
|
{
"authors": [
"coveralls",
"normalhuman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13521",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/2460"
}
|
gharchive/pull-request
|
health was renamed as medicalsciences
The site formerly known as Health is now Medical Sciences, but bodyfetcher has the old URL.
Coverage remained the same at 63.784% when pulling f045ad431bd04f4033724d95e27808e37d8e47c4 on normalhuman:patch-7 into cf5d65e5954c7eb7d0d00315c8a0402d56801f43 on Charcoal-SE:master.
|
2025-04-01T04:54:49.911778
| 2020-09-19T09:44:47
|
704850473
|
{
"authors": [
"SmokeDetector",
"double-beep",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13522",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/4874"
}
|
gharchive/pull-request
|
Kulfy: Watch darkhackerworld.com
Kulfy requests the watch of the watch_keyword darkhackerworld\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
darkhackerworld\.com has been seen in 0 true positives, 0 false positives, and 0 NAAs.
From this deleted answer.
Approved by Mast in Charcoal HQ
|
2025-04-01T04:54:49.914566
| 2021-06-28T13:33:17
|
931578831
|
{
"authors": [
"SmokeDetector"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13523",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/6566"
}
|
gharchive/pull-request
|
Ollie: Watch 01329 609 260
Ollie requests the watch of the watch_number 01329 609 260. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
Approved by Spevacus in Charcoal HQ
|
2025-04-01T04:54:49.918029
| 2022-07-28T02:23:56
|
1320307745
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13524",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/7209"
}
|
gharchive/pull-request
|
cocomac: Watch w88xin.com
cocomac requests the watch of the watch_keyword w88xin\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
w88xin\.com has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by Spevacus in Charcoal HQ
|
2025-04-01T04:54:49.921145
| 2023-08-30T07:19:00
|
1873074729
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13525",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/8056"
}
|
gharchive/pull-request
|
Nick: Watch kushals.com
Nick requests the watch of the watch_keyword kushals\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
kushals\.com has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by Mast in Charcoal HQ
|
2025-04-01T04:54:49.924530
| 2023-12-27T16:10:46
|
2057544579
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13526",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/9729"
}
|
gharchive/pull-request
|
Jesse: Watch apkbossnews.com
Jesse requests the watch of the watch_keyword apkbossnews\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
apkbossnews\.com has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by Spevacus in Charcoal HQ
|
2025-04-01T04:54:49.925747
| 2018-02-21T16:38:57
|
299042001
|
{
"authors": [
"izwick-schachter",
"j-f1"
],
"license": "cc0-1.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13527",
"repo": "Charcoal-SE/metasmoke",
"url": "https://github.com/Charcoal-SE/metasmoke/issues/328"
}
|
gharchive/issue
|
Refactor navbar
Maybe integrate this to make our navbar less magic-y?
cc @izwick-schachter
I started trying to do this and it really seems like more trouble that it's worth.
|
2025-04-01T04:54:49.947847
| 2022-05-09T06:14:25
|
1229226686
|
{
"authors": [
"makyen",
"tripleee"
],
"license": "cc0-1.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13528",
"repo": "Charcoal-SE/metasmoke",
"url": "https://github.com/Charcoal-SE/metasmoke/issues/921"
}
|
gharchive/issue
|
Blazer: LEFT OUTER JOIN query never finishes
I created a query which simply ends up displaying the browser's "Aw, snap" sad face eventually. I have had experiences in the past where Blazer queries were slow and timed out, but this seems like a different beast, as the timeout is not caught and reported by Ruby or other parts of the web framework. I'm wondering if I managed to DoS metasmoke somehow.
The problematic query is https://metasmoke.erwaysoftware.com/data/sql/queries/297-phone-number-in-title-is-not-blacklisted
A variation which avoids the (allegedly more efficient) LEFT OUTER join https://metasmoke.erwaysoftware.com/data/sql/queries/298-phone-number-in-title-is-not-blacklisted finishes quickly and returns some 3,000 rows.
You've given us nothing but a link to the query. Queries are automatically run when you go to the page. You say you're concerned about DoS'ing MS as a result of this query. So, you've left us little alternative for investigation other than clicking on the link you've provided and contributing to the DoS ourselves, if that's what's happening.
Sorry about that. Here is the query text:
-- https://stackoverflow.com/a/14710831/874188
SELECT p.id, p.title, p.body FROM p_posts AS p
LEFT OUTER JOIN p_posts_reasons AS r
ON r.post_id = p.id
AND r.reason_id in (47, 164)
-- 47 = phone number in title
-- 164 = bad phone number in title
WHERE r.post_id IS NULL
Here is a query to fetch the information about an individual query by its Blazer ID:
https://metasmoke.erwaysoftware.com/data/sql/queries/299-blazer-query-by-id?query_id=297
I did actually click on the link you provided. Your query was cached about an hour ago. It returned 357,465 rows. I suspect that you didn't crash MS, but that you did crash the browser tab you were using when it tried to display all 357,465 rows of results.
When experimenting/working on a query, I've found it's a good idea to include a LIMIT to get a short look at if the results resemble something you want. When removing the LIMIT when your query is still exceeding the limit you've set, it's a good idea to run it with a COUNT instead of the unlimited query to get a feel for the number of results which you're returning. I've edited the query to have LIMIT 100.
Here's an image of the prior results:
Thanks for investigating! I'm guessing this is a browser bug then actually. (I'm using Brave.)
(For what it's worth, I have deleted the flawed query now; it didn't do what I wanted anyway.)
|
2025-04-01T04:54:49.952055
| 2017-11-13T22:53:36
|
273607314
|
{
"authors": [
"Hyldahl",
"Nixon-Joseph"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13529",
"repo": "Charcoals/PivotalTracker.NET",
"url": "https://github.com/Charcoals/PivotalTracker.NET/issues/23"
}
|
gharchive/issue
|
Issue with dates and json
I am getting this exception thrown on every date
[DateField] = '(new System.Collections.Generic.Mscorlib_CollectionDebugView<PivotalTrackerDotNet.Domain.Story>(stories).Items[0]).[DateField]' threw an exception of type 'System.FormatException'
Using the exact story getting code listed on the home page of this project. I'm guessing you have the api set up to get the milli date, instead of the standard one. Can you add some overload to get the standard date format? XML is fine reading this format, but I am not using XML.
I've forked your project and added 2 methods to the storyService that specifically get certain date types; you'll probably want to change up how it's done, but it works.
https://github.com/Nixon-Joseph/PivotalTracker.NET
Thanks, I'll take a look.
From the screenshot it looks like either you a wrapping the library with your own code and maybe modified the library your self.
The Story object doesn't have AcceptedOn, CreatedOn projects and there is no Note object.
Therefore I have no idea of what you code is doing to cause this issue.
Also from the screenshot it does look like AcceptedAt CreatedAt properties have been deserialized correctly.
Hmm, maybe that's changed since I posted that. Because I didn't change the models at all.
|
2025-04-01T04:54:49.953556
| 2020-10-01T23:44:00
|
713253094
|
{
"authors": [
"CharlesAverill",
"Danowicz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13530",
"repo": "CharlesAverill/satyrn",
"url": "https://github.com/CharlesAverill/satyrn/issues/46"
}
|
gharchive/issue
|
High contrast mode
Make a high contrast mode in addition to light mode #37
CSS contrast will do the work, let me help.
Absolutely! Could you add a sub-dropdown in the View dropdown for Themes, and submit a PR when you're ready? Thank you!
|
2025-04-01T04:54:49.972675
| 2016-11-03T02:31:13
|
186967062
|
{
"authors": [
"dzcpy",
"huan",
"seanxlliu",
"zixia"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13531",
"repo": "Chatie/wechaty",
"url": "https://github.com/Chatie/wechaty/issues/69"
}
|
gharchive/issue
|
[New Puppet] PuppetWechat4u - enable by WECHATY_PUPPET=wechat4u
Today, wechaty has only one puppet support: PuppetWeb, which is powered by Selenium WebDriver with Chrome Browser.
PuppetWeb has the following pros and cons:
Advantages
No need to deal with wechaty server API changes. Webdriver puppet never touches the HTTP protocol of wechaty web server API; It uses webdriver to hook angularjs in the browser. So it will never mind the server API change.
Behavior not like a bot. It's very like a user use the browser.
Disadvantages
Memory pig. The browser will eat about 100-500MB memory. After serval hours, the situation will be worse, might over 1GB with an active wechat account which has more contacts, more big rooms.
Complicated configuration. If you had set a headless browser under a Linux server, you would understand what I mean. (It became a bit better after Dockerized: https://github.com/wechaty/wechaty/issues/66)
About Wechat4U
Wechat4U is an excellent wechat bot framework that supports both Node.js & Browser, with rich features and an active community of experienced contributors.
So I believe it's a bright decision to made a new Puppet like PuppetWechat4U, to let wechaty users have the second way to use wechaty: not only by Browser Hook but also by Server API Call.
The Plan
This practice will be a good chance to check the wechaty abstract/base classes(the codes under /src directory), to start decoupling with the WebDriver and become more robust, more reusable for the future possible PuppetAndroid, PuppetAndroidPad, PuppetIos, PuppetIosPad.
TBD
Reference
Wechat4U v0.5 Update https://github.com/nodeWechat/wechat4u/commit/5d0275e650572a57ea057c8cb1baff803a253b53
Challenge start: finish a new puppet in one night!
Now is June 1 was 10pm.
Start from scratch from 00:00, now is 2:16 am.
Succeed at 1:11 pm!
ding-dong-bot worked with PuppetWechat4u, code had pushed.
$ WECHATY_PUPPET=wechat4u WECHATY_LOG=info DEBUG=* npm run demo
><EMAIL_ADDRESS>demo /home/zixia/chatie/wechaty
> ts-node examples/ding-dong-bot.ts
| __ __ _ _
| \ \ / /__ ___| |__ __ _| |_ _ _
| \ \ /\ / / _ \/ __| '_ \ / _` | __| | | |
| \ V V / __/ (__| | | | (_| | |_| |_| |
| \_/\_/ \___|\___|_| |_|\__,_|\__|\__, |
| |___/
=============== Powered by Wechaty ===============
-------- https://github.com/chatie/wechaty --------
I'm a bot, my superpower is talk in Wechat.
If you send me a 'ding', I will reply you a 'dong'!
__________________________________________________
Hope you like it, and you are very welcome to
upgrade me to more superpowers!
Please wait... I'm trying to login in...
13:13:05 INFO Wechaty v#git[1c24bf1] starting...
13:13:05 INFO Wechaty initPuppet() using puppet: wechat4u
wechat 重启中... +0ms
follow-redirects options { maxRedirects: 21,
protocol: 'https:',
hostname: 'wx.qq.com',
port: null,
path:
'/cgi-bin/mmwebwx-bin/webwxinit?pass_ticket=HVzv4jGSpfn%252F%252BM6SHE3ZR%252BiAZZ5wma7JJQJXaOJKxKm11oNhSM2kT6%252BYKwd56Lfk&skey=@crypt_c117402d_ce0dcad71413fb0db87ccbe3f445b552&r=1091971814',
method: 'POST',
headers:
{ Accept: 'application/json, text/plain, */*',
'user-agent':
'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.109 Safari/537.36',
connection: 'close',
cookie:
'pgv_pvi=11436618067571104000; pgv_si=s331744187855914500; wxuin=1211516682; wxsid=X5DHlBFELj1Pziyl; wxloadtime=1527915233; mm_lang=zh-CN; webwx_data_ticket=gSd1Qet8FHrlIzQl8uQCfF87; webwxuvid=747895d9dac5a25dd3a78175a5e931d820622c2c2e100257730e529062f992fe7b8677a8b2980f54e6ad283848618766; webwx_auth_ticket=CIsBEMqOnqAMGoABbV0bKSiVjFh4uH1OaLDKlkBrdj0sUYPfjR26FICA+K6SUNLBZE/Xrk3r6M6mpYIFnD6Vwuq7L/tAAovTAZxAwNMGZfEXhaGDFKBX1DL+ck8eDHmtwVax5ZEv7MdtJGy8i76oetciEhh2n++z+ps+/wRvYBZ8iobG/CagXHlSLck=',
'Content-Type': 'application/json;charset=utf-8',
'Content-Length': 147 },
agent: false,
auth: undefined } +1ms
wechat 心跳 +2s
follow-redirects options { maxRedirects: 21,
protocol: 'https:',
hostname: 'wx.qq.com',
port: null,
path:
'/cgi-bin/mmwebwx-bin/webwxstatusnotify?pass_ticket=HVzv4jGSpfn%252F%252BM6SHE3ZR%252BiAZZ5wma7JJQJXaOJKxKm11oNhSM2kT6%252BYKwd56Lfk&lang=zh_CN',
method: 'POST',
headers:
{ Accept: 'application/json, text/plain, */*',
'user-agent':
'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.109 Safari/537.36',
connection: 'close',
cookie:
'pgv_pvi=11436618067571104000; pgv_si=s331744187855914500; wxuin=1211516682; wxsid=X5DHlBFELj1Pziyl; wxloadtime=1527915233; mm_lang=zh-CN; webwx_data_ticket=gSd1Qet8FHrlIzQl8uQCfF87; webwxuvid=747895d9dac5a25dd3a78175a5e931d820622c2c2e100257730e529062f992fe7b8677a8b2980f54e6ad283848618766; webwx_auth_ticket=CIsBEMqOnqAMGoABbV0bKSiVjFh4uH1OaLDKlkBrdj0sUYPfjR26FICA+K6SUNLBZE/Xrk3r6M6mpYIFnD6Vwuq7L/tAAovTAZxAwNMGZfEXhaGDFKBX1DL+ck8eDHmtwVax5ZEv7MdtJGy8i76oetciEhh2n++z+ps+/wRvYBZ8iobG/CagXHlSLck=',
'Content-Type': 'application/json;charset=utf-8',
'Content-Length': 354 },
agent: false,
auth: undefined } +6ms
{ MsgId: '3971822301789638769',
FromUserName:
'@@619a6af1d26cc8cb4f27efba3756da634a79d8565fd44dadb22ef5b728c10752',
ToUserName:
'@548e31e461de6823e86c4261fcd5bd5c3b5f99d911dd3b92853893c83be9171b',
MsgType: 43,
Content:
'@5b7fad094cb2b09b91bbf544d02b57b4:\n@859e5247d4ffaed0372f8aa900924b01d11a3f27c2bbb3513c5ef6f37eb76d841f51c97463429b11976df154d8bbe302ee5e939915846ff4b1d40fc94ea747335af430abc2e92a5a96a7fdd957f33334bb4eaac644f1c9a6f7fd4f75b77cccf55dc949b32fc06f6eb218c13f1ad283f4a85414359937d96d80699ad425068ef2448f74c3720ae0414f4b6afaf1d5a0428bb71444fb145a43ac100c9ae05b86006e9bf6454ca13e6f933ee2bfd49876a66570ef155212e889de0cbf299d29f8cdee722495a597fa13d39f3960a481b71fb9f706b3240c8abff5ce3bd3fbbc1863f0de7fba6c04c2e79ed3fc5b4e578280aa4a75f3c96b7f0830545baa8ed9410e1bc7a98b2629e44938099f0c8eebd7b0c9b22dcd1a7857f494bed8f3ca982f3d5098e79fc0adecc0b625a530937e6a03847d3378b5ba27312d062eb5b325cebb580c736b55225e1aaef2b91c779cfdbd094e8cd0cb8ecf0d5157213f001f5e58293b1b37f57fe57f936bae31e65e78789b0e524ea88db429ced06febcb50c5de499eb01d6ada58032316110b5a68390b0d333ab1190ebb4195399bb581a3a8deda3414cc683e4be6caccbfcc7507801a5233d28e8dbcd5adc3e6f2252c402310d2230602815b07ed8692e521b007d18216e99d4096b30fbb43752a2f0bf8a582e06b99ee50dc36648aaa8c864ede4c930dab809f3de34f7ead33c430df3e204099bf2110bd0dc1d548d39a12f43106864cdf4b01ea5b5494a81f3bf83f496465e0c931ec43aa983b2ca3ab3e0c37c6bc4593fb6df958f524e7804ee9a5c377749ce08fb36974be2524c3758294773759ce95e39eb5bf54b5db8ea4c15bfb6b07676371e7dc82ba22c77859efe37753a409c4d7cdad04b8530f2ae4e9af63c4775351ff46bd0d3758dc2054bf6290d7e216eb6e089944724fab71ece7338393b183b0fe1e3d97b5190b36b4bc0daba3c2ba61edac7ae60f7c319845d1cfa91f42e115cf443adaa0d80b45a685c9565ebdeb0d4554fa4b8e9fd3d62b935b82a2f407a8a2aea5c41b60491b18ac07fd44ce1f91d8a2b1385797c27579c4450a2d0bdf38602dd0e2a8376ebb877ba7355a7934ae8dfbafa3ab4883c257e41a4f84bcc1984c0a5b4f94d1e77cc44443ccb70ecaca58dc29fd7f7338344a5e9cb88fa1c63b87630c5c64e3cfcaf27723c05166',
Status: 3,
ImgStatus: 1,
CreateTime:<PHONE_NUMBER>,
VoiceLength: 0,
PlayLength: 15,
FileName: '',
FileSize: '',
MediaId: '',
Url: '',
AppMsgType: 0,
StatusNotifyCode: 0,
StatusNotifyUserName: '',
RecommendInfo:
{ UserName: '',
NickName: '',
QQNum: 0,
Province: '',
City: '',
Content: '',
Signature: '',
Alias: '',
Scene: 0,
VerifyFlag: 0,
AttrStatus: 0,
Sex: 0,
Ticket: '',
OpCode: 0 },
ForwardFlag: 0,
AppInfo: { AppID: '', Type: 0 },
HasProductId: 0,
Ticket: '',
ImgHeight: 512,
ImgWidth: 288,
SubMsgType: 0,
NewMsgId:<PHONE_NUMBER>789638700,
OriContent: '',
EncryFileName: '',
isSendBySelf: false,
OriginalContent:
'@5b7fad094cb2b09b91bbf544d02b57b4:<br/>@859e5247d4ffaed0372f8aa900924b01d11a3f27c2bbb3513c5ef6f37eb76d841f51c97463429b11976df154d8bbe302ee5e939915846ff4b1d40fc94ea747335af430abc2e92a5a96a7fdd957f33334bb4eaac644f1c9a6f7fd4f75b77cccf55dc949b32fc06f6eb218c13f1ad283f4a85414359937d96d80699ad425068ef2448f74c3720ae0414f4b6afaf1d5a0428bb71444fb145a43ac100c9ae05b86006e9bf6454ca13e6f933ee2bfd49876a66570ef155212e889de0cbf299d29f8cdee722495a597fa13d39f3960a481b71fb9f706b3240c8abff5ce3bd3fbbc1863f0de7fba6c04c2e79ed3fc5b4e578280aa4a75f3c96b7f0830545baa8ed9410e1bc7a98b2629e44938099f0c8eebd7b0c9b22dcd1a7857f494bed8f3ca982f3d5098e79fc0adecc0b625a530937e6a03847d3378b5ba27312d062eb5b325cebb580c736b55225e1aaef2b91c779cfdbd094e8cd0cb8ecf0d5157213f001f5e58293b1b37f57fe57f936bae31e65e78789b0e524ea88db429ced06febcb50c5de499eb01d6ada58032316110b5a68390b0d333ab1190ebb4195399bb581a3a8deda3414cc683e4be6caccbfcc7507801a5233d28e8dbcd5adc3e6f2252c402310d2230602815b07ed8692e521b007d18216e99d4096b30fbb43752a2f0bf8a582e06b99ee50dc36648aaa8c864ede4c930dab809f3de34f7ead33c430df3e204099bf2110bd0dc1d548d39a12f43106864cdf4b01ea5b5494a81f3bf83f496465e0c931ec43aa983b2ca3ab3e0c37c6bc4593fb6df958f524e7804ee9a5c377749ce08fb36974be2524c3758294773759ce95e39eb5bf54b5db8ea4c15bfb6b07676371e7dc82ba22c77859efe37753a409c4d7cdad04b8530f2ae4e9af63c4775351ff46bd0d3758dc2054bf6290d7e216eb6e089944724fab71ece7338393b183b0fe1e3d97b5190b36b4bc0daba3c2ba61edac7ae60f7c319845d1cfa91f42e115cf443adaa0d80b45a685c9565ebdeb0d4554fa4b8e9fd3d62b935b82a2f407a8a2aea5c41b60491b18ac07fd44ce1f91d8a2b1385797c27579c4450a2d0bdf38602dd0e2a8376ebb877ba7355a7934ae8dfbafa3ab4883c257e41a4f84bcc1984c0a5b4f94d1e77cc44443ccb70ecaca58dc29fd7f7338344a5e9cb88fa1c63b87630c5c64e3cfcaf27723c05166' }
follow-redirects options { maxRedirects: 21,
protocol: 'https:',
hostname: 'webpush.wx.qq.com',
port: null,
path:
'/cgi-bin/mmwebwx-bin/synccheck?r=1527916396794&sid=X5DHlBFELj1Pziyl&uin=1211516682&skey=@crypt_c117402d_ce0dcad71413fb0db87ccbe3f445b552&deviceid=e726504311449325&synckey=1_679446148%7C2_679450454%7C3_679448912%7C11_679449713%7C201_1527916387%7C203_1527910120%7C1000_1527916093%7C1001_1527894314%7C2001_1527672305',
method: 'GET',
headers:
{ Accept: 'application/json, text/plain, */*',
'user-agent':
'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.109 Safari/537.36',
connection: 'close',
cookie:
'pgv_pvi=11436618067571104000; pgv_si=s331744187855914500; wxuin=1211516682; wxsid=X5DHlBFELj1Pziyl; wxloadtime=1527915233_expired; mm_lang=zh-CN; webwx_data_ticket=gSd1Qet8FHrlIzQl8uQCfF87; webwxuvid=747895d9dac5a25dd3a78175a5e931d820622c2c2e100257730e529062f992fe7b8677a8b2980f54e6ad283848618766; webwx_auth_ticket=CIsBEMqOnqAMGoABbV0bKSiVjFh4uH1OaLDKlkBrdj0sUYPfjR26FICA+K6SUNLBZE/Xrk3r6M6mpYIFnD6Vwuq7L/tAAovTAZxAwNMGZfEXhaGDFKBX1DL+ck8eDHmtwVax5ZEv7MdtJGy8i76oetciEhh2n++z+ps+/wRvYBZ8iobG/CagXHlSLck=; wxpluginkey=1527916093' },
agent: false,
auth: undefined } +9ms
(node:4453) UnhandledPromiseRejectionWarning: Error: from not found
at PuppetWechat4u.<anonymous> (/home/zixia/chatie/wechaty/src/puppet-wechat4u/puppet-wechat4u.ts:535:17)
at Generator.next (<anonymous>)
at fulfilled (/home/zixia/chatie/wechaty/src/puppet-wechat4u/puppet-wechat4u.ts:4:58)
at process._tickCallback (internal/process/next_tick.js:68:7)
(node:4453) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:4453) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
13:13:18 INFO Bot 李卓桓 login
I have a plan that Wechaty switches to use PuppetWechat4u as default puppet in the future version, but this requires the PuppetWechat4u is stable enough for basic operations, at least the same as PuppetPuppeteer.
If anyone is interested in testing PuppetWechat4u with Wechaty v0.17, please check out the latest source code and run the following command to enable wechat4u puppet:
DEBUG=* WECHATY_LOG=silly WECHATY_PUPPET=wechat4u npm start
Please feel free to file issues if you meet any problem, thanks!
Awesome work!
Thank you @seanxlliu .
Did you use wechat4u before?
Never used yet, so I meant the efficiency of this work is very impressive. I would like to contribute to this project too, testing or coding. Just switch to this puppet due to the default puppet's limitation, room contact is incorrect.
However, I just hit some assert failures when using after switching to this one. Maybe it out of date. I am filing a bug now.
Where can I access the source code? Actually what I would like to know is how to use this module together with webchaty? Is there any example code?
You can run wechaty demo with wechaty-puppet-wechat4u by running the following command in this repository:
$ WECHATY_PUPPET=wechaty-puppet-wechat4u WECHATY_LOG=info DEBUG=* npm run demo
|
2025-04-01T04:54:49.977860
| 2018-10-07T22:29:42
|
367600593
|
{
"authors": [
"ThePridestalker",
"pajlada",
"wigoor",
"zneix"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13532",
"repo": "Chatterino/chatterino2",
"url": "https://github.com/Chatterino/chatterino2/issues/779"
}
|
gharchive/issue
|
Mentions channel messages not showing before focusing the tab first
If you start chatterino but you don't focus the tab /mentions, you get a highlight in that tab but you don't get the messages there until you focus it first.
To test, open chatterino, get pinged by a bot or something and see the /mentions tab highlight.
Click it and it has no message.
You'll need to be more specific about how to set up the test. I tried with 2 separate tabs. one with channel pajlada, and one with only a /mentions
I started chatterino with the pajlada tab active, let someone type my name, then I switched to the /mentions tab and the message was there
I was using 4270c402fb6ca3e81d5f685ad2f36b90414ede24 in the video
Removing this line in my settings.json fixes it:
"smoothScrollingNewMessages": true
still exists in Chatterino Nightly 2.1.5 (commit 367b1970f, 22.09.2019)
quick fix is to disable Smooth scrolling on new messages
Brought up again in #1617 so it's definitely still relevant
Yes, I wanted to post that proof on #1617 but I'll do that here instead https://i.nuuls.com/kepuX.mp4
|
2025-04-01T04:54:49.981352
| 2023-10-05T14:54:06
|
1928494501
|
{
"authors": [
"nleach999",
"pedrompflopes"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13533",
"repo": "Checkmarx/ast-cli",
"url": "https://github.com/Checkmarx/ast-cli/issues/592"
}
|
gharchive/issue
|
[BUG] Contributor count breaks with an NPE when encountering an empty repo.
Describe the bug
When running the contributor count utility, it throws an NPE when it encounters a repository that does not contain code. The enumeration ends.
Expected behavior
It should see the repo is empty and move on without stopping the enumeration.
Actual behavior
While enumerating an org's repos, it encountered a repo with no code in it. It threw an NPE and exited the program.
Steps to reproduce
Use a GitHub org (it probably does the same in other SCMs, I only tried GitHub)
Create a new repository, don't add any files to it.
Run the contributor count tool to enumerate the org.
Environment
CLI 2.0.58
Powershell on an unknown Windows version.
Additional comments
N/A
Logs
Hi @nleach999
Thanks for raising the issue. We are currently fixing it under #599
|
2025-04-01T04:54:50.001313
| 2018-07-17T02:13:48
|
341740272
|
{
"authors": [
"BIGKUANG",
"ChenRocks",
"Colanim"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13534",
"repo": "ChenRocks/fast_abs_rl",
"url": "https://github.com/ChenRocks/fast_abs_rl/issues/4"
}
|
gharchive/issue
|
when my dataset is single sentence abstract.
Dear Chen
I train this model in othe dataset of single sentence abstract. but in the process of implementation, i find that the model(abstractor) build an channel by single sentence to single abstract.
If i try to intelligently select K sentence by rouge in extractor stage, than i flatten this sentence to train in abstractor and full_rl, do you think this is feasible?
Does the model of fast_abs_rl apply to single sentence abstract? I would be very grateful for your reply and advices.
Thanks for your interest in using our code. Please read the full paper for the rationale on why we select K sentences then output abstract.
In short, we extract K sentences and then decode K abstract sentences independently. In the experiment we found this is beneficial for multi-sentence summary task. I'm not sure about your task/dataset so it's difficult for me to give advice.
Feel free to fork the repository and then build upon our model for your own experiments.
closing this due to no further discussion
@BIGKUANG
If i try to intelligently select K sentence by rouge in extractor stage, than i flatten this sentence to train in abstractor and full_rl
Did you try this approach ? How are the results ?
|
2025-04-01T04:54:50.008685
| 2017-11-05T18:50:52
|
271303895
|
{
"authors": [
"ChenglongChen",
"hughperkins"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13535",
"repo": "ChenglongChen/pytorch-madrl",
"url": "https://github.com/ChenglongChen/pytorch-madrl/issues/1"
}
|
gharchive/issue
|
License?
Hi,
What is the license?
Hugh
(eg MIT is fairly standard, github has some functions to make adding a license fairly straightforward, though I confess to not having tested them: https://help.github.com/articles/adding-a-license-to-a-repository/ )
MIT license is added. Thanks for the info.
Awesome, thanks! :)
|
2025-04-01T04:54:50.029032
| 2021-02-14T19:33:11
|
808029096
|
{
"authors": [
"dvd101x",
"seeden"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13536",
"repo": "Chia-Network/chia-blockchain-gui",
"url": "https://github.com/Chia-Network/chia-blockchain-gui/issues/30"
}
|
gharchive/issue
|
Local Harvester Plots : Shifted columns for Status and Action
On the table: Local Harvester Plots on the tab for Plots.
For the rows that are in status: plotting or status: queued. The status is in plotting column and actions is in status column. They are one column to the left.
Please put them one column to the right.
Fixed via https://github.com/Chia-Network/chia-blockchain-gui/pull/40
PR was merged
|
2025-04-01T04:54:50.117806
| 2022-03-25T10:50:56
|
1180636713
|
{
"authors": [
"arvidn",
"dannywillems",
"emlowe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13537",
"repo": "Chia-Network/chiavdf",
"url": "https://github.com/Chia-Network/chiavdf/pull/106"
}
|
gharchive/pull-request
|
Add clang-format
Dump .clang-format using:
clang-format -style=llvm -dump-config > .clang-format
and set SortIncludes to false
Run clang-format version 13.0.1-3+b1 on src/h src/.cpp src/.hpp src/.c
Rebase on main, splitted in two commits: one adding clang-format file and one executing clang-format.
@dannywillems I'm generally in favor of using clang format. I think the main benefit from it comes from requiring the canonical format in CI, which this patch does not introduce. This will become outdated very quickly after landing. There would probably need to be some documentation for committers how to apply the clang format before pushing a PR.
I would also have expected the clang format config to somewhat match the existing format. Most importantly, use 4 spaces for indentation rather than 2. Ideally, the change to the source code diff should be as small as possible.
Closing abandoned PR
|
2025-04-01T04:54:50.124465
| 2022-04-25T16:47:17
|
1214778345
|
{
"authors": [
"LVerhulst4321"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13538",
"repo": "ChicagoWorldcon/planorama",
"url": "https://github.com/ChicagoWorldcon/planorama/pull/111"
}
|
gharchive/pull-request
|
Added fields to Venues and Rooms and created Room_Sets.
Henry, Let me know if I did the migrations properly.
Henry, I made the suggested changes. I also figured out how to add a field to the rooms table to reference the room_sets table. This field would contain the default room set for that room. I think Nchanter wants a session to also have a link to the room_sets table and then a conflict report can be run to find sessions with a room_set that doesn't match the room that the session is assigned to.
|
2025-04-01T04:54:50.125685
| 2018-04-23T01:46:49
|
316635254
|
{
"authors": [
"ChiefOfGxBxL"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13539",
"repo": "ChiefOfGxBxL/WC3MapTranslator",
"url": "https://github.com/ChiefOfGxBxL/WC3MapTranslator/issues/12"
}
|
gharchive/issue
|
Info file needs Wiki documentation
Info file (war3map.w3i) is missing documentation on the Wiki, even though the code is implemented. Add to "Other" section.
✔️ Documented the Info translator
|
2025-04-01T04:54:50.144680
| 2015-09-23T14:32:07
|
107933055
|
{
"authors": [
"ChimeraCoder",
"jeffk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13540",
"repo": "ChimeraCoder/anaconda",
"url": "https://github.com/ChimeraCoder/anaconda/pull/104"
}
|
gharchive/pull-request
|
Update README to highlight .Close() functionality.
Creating TwitterApi structs without calling .Close() on them can result in goroutine leaks in long-running applications. This documentation change highlights that case.
Thanks! Merged into master as eca4dfc90a6bd2039af094bfd3d1d153376b29cf.
|
2025-04-01T04:54:50.149866
| 2021-06-11T03:37:33
|
919796651
|
{
"authors": [
"Samsagax",
"massatt212"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13541",
"repo": "ChimeraOS/chimera",
"url": "https://github.com/ChimeraOS/chimera/issues/160"
}
|
gharchive/issue
|
Hide tools
In the installed location, is it possible to hide none games are add a table to install games, seeing proton tools and steam runtime kinda ruins the Console look.
i want to add some extra options and I'm missing the steamcompmgr only seeing original
I'm using ubuntu 20.04
i want to add some extra options and I'm missing the steamcompmgr only seeing original
I'm using ubuntu 20.04
I did not understand the question. What are you trying to accomplish?
-b Disable game focus hack
-p Disable proton/wine color flash suppression hack
-g Enable debug logging for game focus and
I want to try these out, but I don't have the file to do so.
-b Disable game focus hack
-p Disable proton/wine color flash suppression hack
-g Enable debug logging for game focus and
I want to try these out, but I don't have the file to do so.
Are you running steam composer? Or just steam big picture mode?
im running steamos composite plus
im running steamos composite plus
Then follow the Configuration documentation to acchive your goal.
Anyway this conversation went off-topic with the original issue.
i do not have the file to configure it lol, send me one here so I can
i do not have the file to configure it lol, send me one here so I can
Create/Edit ~/.config/steamos-compositor-plus file and set COMPOSITORCMD variable to your desired value. For example:
COMPOSITORCMD="steamcompmgr -b"
ok ill try that
|
2025-04-01T04:54:50.230064
| 2023-12-09T09:54:04
|
2033771366
|
{
"authors": [
"oilrich25"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13542",
"repo": "ChristopherHX/runner.server",
"url": "https://github.com/ChristopherHX/runner.server/issues/283"
}
|
gharchive/issue
|
About WIndow Container getCacheEntry failed problem
I tried using ChristopherHX/runner.server on the windows platform and specifying container as runner-images:windows-ltsc2019.
windows:
runs-on: windows-2019
container: runner-images:windows-ltsc2019
But it was found that it would cause an error of ##[warning]Failed to restore: getCacheEntry failed: connect ETIMEDOUT <IP_ADDRESS>:50299. how to solve
##[group]Run actions/cache@v3
with:
key: toolchains-win-7d9757126edce6986823b174df112b1b55d031724e9815384bc2c7fb2a221936-v1
path: src/build\n
enableCrossOsArchive: false
fail-on-cache-miss: false
lookup-only: false
env:
CACHE_EPOCH: 1
CCACHE_MAXFILES: 0
CCACHE_MAXSIZE: 200M
SCCACHE_CACHE_SIZE: 200M
##[endgroup]
##[command]C:\Windows\system32\docker.EXE exec f85eb2f3dd2369c82121dad0c7902445c031ccb819e0a4aef7fa0594241d67c9 sh -c "cat /etc/*release | grep ^ID"
##[warning]Failed to restore: getCacheEntry failed: connect ETIMEDOUT <IP_ADDRESS>:50299
Cache not found for input keys: toolchains-win-7d9757126edce6986823b174df112b1b55d031724e9815384bc2c7fb2a221936-v1
Windows containers are affeced by the Windows Firewall (Third party firewalls make it even more complicated).
https://github.com/ChristopherHX/runner.server/blob/0b83f0eaa83214b9ae4a9fd37b30f82b733fe375/.github/workflows/nuget.yml#L181C3-L182C71
In the GitHub Actions CI, I decided to turn the firewall off due to beeing a non interactive system I don't host myself.
You should allow the gitea-actions-runner.exe or Runner.Server.exe to be accessible from the outside (or only from the docker for windows network adapter, but that was too complicated for me)
EDIT The tcp port is random, it should be a program based firewall rule for all ports, e.g. allowing in the interactive prompt from Windows worked for development purposes.
EDIT2 Now I understand why my first firewall rule in Actions didn't work in 2021, gharun doesn't open the port 🤦♂️
Okay, I'll test it later
I directly turned off the firewall of the selfed-host host, which solved the problem perfectly.
|
2025-04-01T04:54:50.244893
| 2020-09-09T12:04:50
|
696756170
|
{
"authors": [
"Ciantic",
"redvelociraptor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13543",
"repo": "Ciantic/keymapceditor-vsc",
"url": "https://github.com/Ciantic/keymapceditor-vsc/issues/3"
}
|
gharchive/issue
|
Fix for typescript errors during build
Leaving this here for anyone else trying to get this to build in 2020. I'm an npm newbie, so it took me a while to figure it out.
The code directory needs to be explicitly renamed to keymapceditor rather than keymapceditor-vsc.
vscode must be explicitly installed into the project directory, e.g. npm install vscode (unusual for me, coming from typical gcc compilation experience).
Lastly, to get it to build without code errors, had to force downgrade typescript from v4.0.2 and node from latest (iirc v14.0.2) to get it to compile:
npm install --save<EMAIL_ADDRESS>npm install --save<EMAIL_ADDRESS>
as the typescript run (tsc -p ./) was throwing these errors:
node_modules/@types/node/index.d.ts:164:11 - error TS2300: Duplicate identifier 'IteratorResult'.
164 interface IteratorResult<T> { }
~~~~~~~~~~~~~~
../../../../usr/local/lib/node_modules/typescript/lib/lib.es2015.iterable.d.ts:41:6
41 type IteratorResult<T, TReturn = any> = IteratorYieldResult<T> | IteratorReturnResult<TReturn>;
~~~~~~~~~~~~~~
'IteratorResult' was also declared here.
Build environment: macOS 10.14.6 and npm v6.14.7, installed from homebrew.
I found the solution here after some digging: Stackoverflow
Great! Have you done some changes I should integrate?
Unfortunately, this type of thing is well beyond my programming skill level. I was hoping to use this help me learn to use QMK for my first QMK-driven keyboard.
After trying to use the extension, I discovered it's expecting the web assets from the original web-based version: stack trace: Error: ENOENT: no such file or directory, open '/Users/redvraptor/.vscode/extensions/ciantic.keymapceditor-vsc-0.7.6/out/keymapceditor/index.html'.
I will see if I can resolve these issues and write up some further steps on how I did so.
@redvelociraptor what would you like the extension to do? Why would you want to develop it. Maybe I can add the feature you want.
|
2025-04-01T04:54:50.291157
| 2022-12-19T12:55:02
|
1502900963
|
{
"authors": [
"KyleTryon",
"Shmarkus"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13544",
"repo": "CircleCI-Public/github-cli-orb",
"url": "https://github.com/CircleCI-Public/github-cli-orb/pull/36"
}
|
gharchive/pull-request
|
[semver:minor] Implement Github CLI "pr merge" command
Implemented gh pr merge with the possibility to define the branch or additional parameters
Thank you for reminding me the CHANGELOG is still present here. Happy to take the change here, but we no longer user the changelog. I'll be sure to remove it in a change soon. We now use GitHub's Releases API.
https://github.com/CircleCI-Public/github-cli-orb/releases
When we publish this change, we will use this PR to populate the release notes. I'll be sure to update all that documentation. Thank you.
The code looks good. Could I ask how you intend to use this @Shmarkus ? Is there an example of config you could provide on how you plan to execute this?
By default CircleCI will trigger on a Push (including a Tag), we can augment this a bit more with https://github.com/CircleCI-Public/trigger-circleci-pipeline-action
In what scenario would you run this job?
I can create a concrete example as well but since I'm not at the computer at the moment, I'll just illustrate my use-case.
At the moment, the Dependabot is configured to check for version updates. It creates a pull request with the new version. (Renovate bot has the same functionality)
When the test pipeline succeeds I want to merge the PR automatically (this is where the pain is at the moment)
Added example, removed changelog from changelog
|
2025-04-01T04:54:50.303465
| 2019-03-03T15:07:36
|
416525753
|
{
"authors": [
"Amansingla97",
"satu0king"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13545",
"repo": "CircuitVerse/CircuitVerse",
"url": "https://github.com/CircuitVerse/CircuitVerse/pull/178"
}
|
gharchive/pull-request
|
Fixing Dropdown Position
Fixes #177
Changes Made
added a bootstrap class that fixes the positioning
Thank you!
|
2025-04-01T04:54:50.306381
| 2019-12-21T17:42:17
|
541351949
|
{
"authors": [
"coveralls",
"nitin10s",
"nrs1729"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13546",
"repo": "CircuitVerse/CircuitVerse",
"url": "https://github.com/CircuitVerse/CircuitVerse/pull/757"
}
|
gharchive/pull-request
|
Added ability to delete profile picture
Describe the changes you have made in this PR:
Added a button on the show profile page to delete the profile picture. Issue #595
Screenshots of the changes (If any):
Coverage decreased (-0.3%) to 82.307% when pulling<PHONE_NUMBER>aaff7c32642c6a55d199e27e78f83e on nrs1729:remove-user-photo-595 into 2c01dad8e0b59646b9625313a0da5801dde1e221 on CircuitVerse:master.
Oh sorry forgot about tests, I'll add them
@satu0king Old UI. Please close
|
2025-04-01T04:54:50.322367
| 2024-02-12T15:19:20
|
2130355705
|
{
"authors": [
"DhairyaMajmudar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13547",
"repo": "CircuitVerse/cv-frontend-vue",
"url": "https://github.com/CircuitVerse/cv-frontend-vue/pull/271"
}
|
gharchive/pull-request
|
Removing Warnings and Errors from TextEditor.vue file
Describe the changes you have made in this PR -
Removed warnings and errors from the TextEditor.vue file
Screenshots of the changes (If any) -
Console before changes
Console after changes
Note: Please check Allow edits from maintainers. if you would like us to assist in the PR.
@Arnabdaz @tachyons I have removed about 18 warnings and errors from the TextEditor.vue file,
Pls. have a look : )
|
2025-04-01T04:54:50.342718
| 2022-11-07T05:28:00
|
1437788817
|
{
"authors": [
"poxman",
"qifawu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13548",
"repo": "CiscoPSIRT/openVulnQuery",
"url": "https://github.com/CiscoPSIRT/openVulnQuery/issues/23"
}
|
gharchive/issue
|
get_by_nxos will return first_fixed with None
Hi,
I tried get_by_nxos and noticed that the advisories always had "first_fixed" with None .
ad_nxos = query.get_by_nxos('nxos', '7.3(3)N1(1)')
In order to troubleshoot this, I used get_request to fetch the raw API output for both ios and nxos:
ad_iosxe = query.get_request('iosxe', params={'version':'3.16.1S'})
ad_nxos = query.get_request('nxos', params={'version':'7.3(3)N1(1)'})
From the results, the raw outputs of each ios advisory has the key called "firstFixed":
{
'advisoryId' = {str} 'cisco-sa-appnav-xe-dos-j5MXTR4'
...
'firstFixed' = {list: 1} ['16.12.6']
...
}
Instead, the raw output of each nxos advisory has the key "platforms", which contains the 'firstFixes':
{
'advisoryId' = {str} 'cisco-sa-nxos-cdp-dos-ce-wWvPucC9'
...
'platforms' = {list: 2} [
{'id': '265090', 'name': 'Cisco Nexus 5000 Series Switches', 'firstFixes': [{'id': '289274', 'name': '7.3(12)N1(1)'}], 'vulnerabilityState': 'vulnerable'},
{'id': '265094', 'name': 'Cisco Nexus 6000 Series Switches', 'firstFixes': [{'id': '289274', 'name': '7.3(12)N1(1)'}], 'vulnerabilityState': 'vulnerable'}
]
...
}
I think the reason for the issue lies in no code logic handling nxos outputs in advisory.py. The temporary workaround is to use get_request and get firstFixes manually.
Hi,
I tried get_by_nxos and noticed that the advisories always had "first_fixed" with None .
ad_nxos = query.get_by_nxos('nxos', '7.3(3)N1(1)')
In order to troubleshoot this, I used get_request to fetch the raw API output for both ios and nxos:
ad_iosxe = query.get_request('iosxe', params={'version':'3.16.1S'})
ad_nxos = query.get_request('nxos', params={'version':'7.3(3)N1(1)'})
From the results, the raw outputs of each ios advisory has the key called "firstFixed":
{
'advisoryId' = {str} 'cisco-sa-appnav-xe-dos-j5MXTR4'
...
'firstFixed' = {list: 1} ['16.12.6']
...
}
Instead, the raw output of each nxos advisory has the key "platforms", which contains the 'firstFixes':
{
'advisoryId' = {str} 'cisco-sa-nxos-cdp-dos-ce-wWvPucC9'
...
'platforms' = {list: 2} [
{'id': '265090', 'name': 'Cisco Nexus 5000 Series Switches', 'firstFixes': [{'id': '289274', 'name': '7.3(12)N1(1)'}], 'vulnerabilityState': 'vulnerable'},
{'id': '265094', 'name': 'Cisco Nexus 6000 Series Switches', 'firstFixes': [{'id': '289274', 'name': '7.3(12)N1(1)'}], 'vulnerabilityState': 'vulnerable'}
]
...
}
I think the reason for the issue lies in no code logic handling nxos outputs in advisory.py. The temporary workaround is to use get_request and get firstFixes manually.
Thanks @qifawu
Cisco IOS and Cisco IOS XE doesn’t have the concept of platform, it is a single product. NX-OS,ACI, ASA, etc have the additional fields of platform and hence a first fix release based on per platform rather than just the operating system.
You can define a filter for the platformAlias: example:
filter = query_client.OptionalParameters(query_client.PLATFORMALIAS,'NEXUS9000')
ad_nxos = client.get_by_nxos('default','7.0(3)I6(2)',filter)
If you wish to obtain the first fix for each:
for item in ad_nxos:
... print(item.advisory_title)
... for platform in item.platforms:
... print(" " + platform.name,end=":")
... for fix in platform.firstFixes:
... print(fix['name'])
...
Cisco NX-OS Software MPLS OAM Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(10)
Cisco FXOS and NX-OS Software Cisco Discovery Protocol Denial of Service and Arbitrary Code Execution Vulnerability
Cisco Nexus 9000 Series Switches:9.3(10)
Cisco NX-OS Software OSPFv3 Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:9.3(9)
Cisco Nexus 9000 Series Switches Bidirectional Forwarding Detection Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:9.3(9)
Cisco FXOS and NX-OS Software Cisco Discovery Protocol Service Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:9.3(9)
Cisco NX-OS Software Cisco Fabric Services Over IP Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:9.3(9)
Cisco NX-OS Software NX-API Command Injection Vulnerability
Cisco Nexus 9000 Series Switches:9.3(9)
Multiple Cisco Operating Systems Unidirectional Link Detection Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(10)
Cisco FXOS and NX-OS Software Unidirectional Link Detection Denial of Service and Arbitrary Code Execution Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(10)
Cisco NX-OS Software IPv6 Netstack Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(9)
Cisco NX-OS Software NX-API Cross-Site Request Forgery Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(9)
Cisco FXOS and NX-OS Software Cisco Fabric Services Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(9)
Cisco NX-OS Software Data Management Engine Remote Code Execution Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(8)
Cisco NX-OS Software Call Home Command Injection Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(2)
Cisco NX-OS Software IPv6 Protocol Independent Multicast Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(8)
Cisco IOS, IOS XE, IOS XR, and NX-OS Software One Platform Kit Remote Code Execution Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(8)
Cisco FXOS, IOS XR, and NX-OS Software Cisco Discovery Protocol Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(8)
Cisco NX-OS Software Cisco Fabric Services over IP Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(4)
Cisco FXOS and NX-OS Software Authenticated Simple Network Management Protocol Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(6)
Cisco NX-OS Software Remote Management Memory Leak Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(4)
Cisco NX-OS Software Network Time Protocol Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(6)
Cisco NX-OS Software Virtualization Manager Command Injection Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(6)
Cisco NX-OS Software NX-API Denial of Service Vulnerability
Cisco Nexus 9000 Series Switches:7.0(3)I7(6)
Output without a filter defined:
Cisco NX-OS Software MPLS OAM Denial of Service Vulnerability
Cisco Nexus 3000 Series Switches:7.0(3)I7(10)
Cisco Nexus 9000 Series Switches:7.0(3)I7(10)
Cisco FXOS and NX-OS Software Cisco Discovery Protocol Denial of Service and Arbitrary Code Execution Vulnerability
Cisco Nexus 3000 Series Switches:9.3(10)
Cisco Nexus 9000 Series Switches:9.3(10)
.
.
.
HTH. If need more qualification on this please pop an email to<EMAIL_ADDRESS>
|
2025-04-01T04:54:50.395224
| 2021-05-31T23:09:16
|
907750595
|
{
"authors": [
"codecov-commenter",
"tuomas777"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13549",
"repo": "City-of-Helsinki/liikunta-helsinki",
"url": "https://github.com/City-of-Helsinki/liikunta-helsinki/pull/18"
}
|
gharchive/pull-request
|
Navigation bar logo functionality enhancements
Just a couple of small navigation bar logo related functionalities. Implemented now because an older developer got mad at clicking the logo was not working.
Codecov Report
Merging #18 (f809154) into main (2a03411) will decrease coverage by 0.12%.
The diff coverage is 66.66%.
@@ Coverage Diff @@
## main #18 +/- ##
==========================================
- Coverage 66.26% 66.13% -0.13%
==========================================
Files 38 38
Lines 249 251 +2
Branches 52 53 +1
==========================================
+ Hits 165 166 +1
- Misses 84 85 +1
Impacted Files
Coverage Δ
src/components/navigation/Navigation.tsx
60.00% <66.66%> (-2.50%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 2a03411...f809154. Read the comment docs.
|
2025-04-01T04:54:50.412137
| 2024-05-29T08:38:07
|
2322720505
|
{
"authors": [
"rikuke",
"terovirtanen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13550",
"repo": "City-of-Helsinki/yjdh",
"url": "https://github.com/City-of-Helsinki/yjdh/pull/3054"
}
|
gharchive/pull-request
|
feat: change failure response to ahjo as 200 ok
Description :sparkles:
Ahjo gets confused if the Helsinki-lisä responds to a failing callback request with 400 Bad request, so the
response status needs to be changed to 200 OK.
Issues :bug:
Testing :alembic:
Screenshots :camera_flash:
Additional notes :spiral_notepad:
APPLICANT branch is deployed to platta: https://helsinkilisa-ui-pr3054.dev.hel.ninja :rocket::rocket::rocket:
API branch is deployed to platta: https://helsinkilisa-pr3054.api.dev.hel.ninja/healthz :rocket::rocket::rocket:
TestCafe result is success for https://helsinkilisa-ui-handler-pr3054.dev.hel.ninja :laughing::tada::tada::tada:
TestCafe result is success for https://helsinkilisa-ui-pr3054.dev.hel.ninja :laughing::tada::tada::tada:
|
2025-04-01T04:54:50.422769
| 2024-10-15T07:17:09
|
2587868025
|
{
"authors": [
"rikuke",
"terovirtanen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13551",
"repo": "City-of-Helsinki/yjdh",
"url": "https://github.com/City-of-Helsinki/yjdh/pull/3443"
}
|
gharchive/pull-request
|
Hl 1475 submitted at
Description :sparkles:
HL-1475
Use the submitted_at custom Application model attribute in the Ahjo payload and report csv:s,and created_at if it's not present.
HL-1362
Add an alternate, more compact csv, similar to the powerbi report, to the handler's report screen for downloading.
HL-1483
Show the alteration.recovery_amount correctly in the csv report.
Issues :bug:
Testing :alembic:
Screenshots :camera_flash:
Additional notes :spiral_notepad:
API branch is deployed to platta: https://helsinkilisa-pr3443.api.dev.hel.ninja/healthz :rocket::rocket::rocket:
API branch is deployed to platta: https://helsinkilisa-pr3443.api.dev.hel.ninja/healthz :rocket::rocket::rocket:
HANDLER branch is deployed to platta: https://helsinkilisa-ui-handler-pr3443.dev.hel.ninja :rocket::rocket::rocket:
APPLICANT branch is deployed to platta: https://helsinkilisa-ui-pr3443.dev.hel.ninja :rocket::rocket::rocket:
TestCafe handler result is failed for https://helsinkilisa-ui-handler-pr3443.dev.hel.ninja :crying_cat_face::anger::boom::boom:
API branch is deployed to platta: https://helsinkilisa-pr3443.api.dev.hel.ninja/healthz :rocket::rocket::rocket:
APPLICANT branch is deployed to platta: https://helsinkilisa-ui-pr3443.dev.hel.ninja :rocket::rocket::rocket:
HANDLER branch is deployed to platta: https://helsinkilisa-ui-handler-pr3443.dev.hel.ninja :rocket::rocket::rocket:
TestCafe result is success for https://helsinkilisa-ui-pr3443.dev.hel.ninja :laughing::tada::tada::tada:
TestCafe result is success for https://helsinkilisa-ui-handler-pr3443.dev.hel.ninja :laughing::tada::tada::tada:
|
2025-04-01T04:54:50.431506
| 2023-03-17T22:20:41
|
1629982697
|
{
"authors": [
"ludtkemorgan",
"sarahlazarich"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13552",
"repo": "CityOfDetroit/bloom",
"url": "https://github.com/CityOfDetroit/bloom/issues/1591"
}
|
gharchive/issue
|
Bring upgrades over from Bloom Core
We need to keep major upgrades in sync with core. The following needs to be upgrade
Node 18
Our current version of node is outdated and we should update it to the latest.
This is a continuation of #3317.
Core commit: https://github.com/bloom-housing/bloom/pull/3343
Make sure to update Travis CI pipeline
Next
React
This has been merged and can be tested in https://detroit-public-dev.netlify.app/.
Issues were found with the CSS that will most likely be fixed with https://github.com/bloom-housing/bloom/issues/3283
Issues found will be addressed by UIC alignment PR and can be QA'd as a part of that.
|
2025-04-01T04:54:50.434920
| 2021-10-20T02:15:41
|
1030903882
|
{
"authors": [
"GabeBenjamin",
"YazeedLoonat",
"anders-schneider"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13553",
"repo": "CityOfDetroit/bloom",
"url": "https://github.com/CityOfDetroit/bloom/issues/705"
}
|
gharchive/issue
|
[Email/SMS] Create queue/function to send sms to users
What feature is this part of?
"New Listing" Email/SMS Notifications #695
The current state: the creation of a new listing will trigger SMS notifications to all opted-in users.
The remaining work to do on this:
[ ] Send notifications when an existing listing's status changes to "active" (I have #907 out to address this)
[ ] For the notifications for newly-created listings, make sure the new listing has status "active" (if not, don't send notifications).
[ ] In https://github.com/CityOfDetroit/bloom/blob/main/backend/core/src/sms/services/sms.service.ts, update the SMS notification body (e.g. to include a link to the new listing, and possibly translate the message text).
[ ] The current implementation retrieves a list of all users from the DB, and filters them by "should we send them a notification?" in https://github.com/CityOfDetroit/bloom/blob/main/backend/core/src/sms/services/sms.service.ts. A better-performance approach would be to do the filtering in the SQL query to the DB; this will require changes to UserService too.
closing!
|
2025-04-01T04:54:50.442885
| 2021-12-21T19:29:06
|
1086165517
|
{
"authors": [
"aisiri-murulidhar",
"plunkettgoogle"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13554",
"repo": "CityOfDetroit/bloom",
"url": "https://github.com/CityOfDetroit/bloom/pull/903"
}
|
gharchive/pull-request
|
Fix/features filtering
Issue
Closes #901
Description
A few fixes:
The comparison needed to be "="
We needed to specify listing_features when mapping filter parameters to db queries
We needed to add listing_features to the set of tables that are joined and selected for full listing views
Also adds a coupla tests to verify stuff works.
Type of change
[x] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Prototype/POC (not to merge)
[ ] This change is a refactor/addresses technical debt
[ ] This change requires a documentation update
[ ] This change requires a SQL Script
How Can This Be Tested/Reviewed?
Reseed the db, deploy locally, and navigate to http://localhost:3000/listings/filtered?page=1&elevator=true. You should see 8 listings.
Checklist:
[x] My code follows the style guidelines of this project
[x] I have performed a self-review of my own code
[x] I have reviewed the changes in a desktop view
[x] I have reviewed the changes in a mobile view
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[x] My changes generate no new warnings
[x] I have added tests that prove my fix is effective or that my feature works
[x] New and existing unit tests pass locally with my changes
[ ] Any dependent changes have been merged and published in downstream modules
[x] I have assigned reviewers
[ ] I have updated the changelog to include a description of my changes
[ ] I have run yarn generate:client if I made backend changes
[ ] I have exported any new pieces in ui-components
looks like some tests are failing
looks like some tests are failing
I missed the necessary change to view.spec.ts, oops.
|
2025-04-01T04:54:50.447780
| 2023-12-06T10:40:38
|
2028250730
|
{
"authors": [
"coveralls",
"ixje"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13555",
"repo": "CityOfZion/neo-mamba",
"url": "https://github.com/CityOfZion/neo-mamba/pull/297"
}
|
gharchive/pull-request
|
api: add tokens_count method to NEP11Contract
NEP-11 defines an optional tokens method. the NEP11Contract wraps this method and manually unwraps the iterator such that it does not have to rely on RPC servers that have iterator sessions enabled (which many don't). However, that means there's a ~2000 item limit on the return. This helper is part 1 to resolve that limit.
Part 2 will allow defining a start index into the iterator. Combining both allows us to query all items using a while or for loop.
coverage: 80.593% (-0.4%) from 80.98%
when pulling 514194b6bdb75df20e60387d0bcdb2040a0f8dd5 on token_count
into 4fdf5aa021e61fb55cb55bc522efdbfa24b6dd5b on master.
|
2025-04-01T04:54:50.474147
| 2023-05-12T04:57:49
|
1706924806
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13556",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/10300"
}
|
gharchive/issue
|
⚠️ Bofrost.be has degraded performance
In 3d144fc, Bofrost.be (https://www.bofrost.be) experienced degraded performance:
HTTP code: 200
Response time: 2425 ms
Resolved: Bofrost.be performance has improved in 27130ba.
|
2025-04-01T04:54:50.477386
| 2023-05-19T23:19:17
|
1717930985
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13557",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/11026"
}
|
gharchive/issue
|
⚠️ Bofrost.fr has degraded performance
In ed799c1, Bofrost.fr (https://www.bofrost.fr) experienced degraded performance:
HTTP code: 200
Response time: 1297 ms
Resolved: Bofrost.fr performance has improved in c8dd3ce.
|
2025-04-01T04:54:50.480519
| 2023-06-06T20:28:52
|
1744549693
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13558",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/12385"
}
|
gharchive/issue
|
⚠️ Bofrost.be has degraded performance
In 6943f51, Bofrost.be (https://www.bofrost.be) experienced degraded performance:
HTTP code: 200
Response time: 1396 ms
Resolved: Bofrost.be performance has improved in 3150f8c.
|
2025-04-01T04:54:50.483383
| 2023-06-08T01:52:01
|
1746941232
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13559",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/12495"
}
|
gharchive/issue
|
⚠️ Bofrost.be has degraded performance
In 2f84536, Bofrost.be (https://www.bofrost.be) experienced degraded performance:
HTTP code: 200
Response time: 3731 ms
Resolved: Bofrost.be performance has improved in 42fd929.
|
2025-04-01T04:54:50.486399
| 2023-07-03T09:49:26
|
1785800240
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13560",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/15137"
}
|
gharchive/issue
|
⚠️ Bofrost.be has degraded performance
In cfd2d2c, Bofrost.be (https://www.bofrost.be) experienced degraded performance:
HTTP code: 200
Response time: 932 ms
Resolved: Bofrost.be performance has improved in 245f371.
|
2025-04-01T04:54:50.489680
| 2023-08-21T23:33:52
|
1860317295
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13561",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/21085"
}
|
gharchive/issue
|
⚠️ Bofrost.fr has degraded performance
In a6246b6, Bofrost.fr (https://www.bofrost.fr) experienced degraded performance:
HTTP code: 200
Response time: 808 ms
Resolved: Bofrost.fr performance has improved in 32cc544 after 306 days, 11 hours, 8 minutes.
|
2025-04-01T04:54:50.492811
| 2023-08-25T20:47:12
|
1867665465
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13562",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/21357"
}
|
gharchive/issue
|
⚠️ Bofrost.at has degraded performance
In 27233ec, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 1959 ms
Resolved: Bofrost.at performance has improved in d2da788 after 310 days, 8 hours, 36 minutes.
|
2025-04-01T04:54:50.495701
| 2022-11-09T07:45:04
|
1441580043
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13563",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/221"
}
|
gharchive/issue
|
⚠️ Bofrost.at has degraded performance
In cea0f03, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 986 ms
Resolved: Bofrost.at performance has improved in 4e8a420.
|
2025-04-01T04:54:50.498597
| 2023-09-18T20:03:11
|
1901653633
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13564",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/23409"
}
|
gharchive/issue
|
⚠️ Bofrost.de has degraded performance
In b561447, Bofrost.de (https://www.bofrost.de) experienced degraded performance:
HTTP code: 200
Response time: 1061 ms
Resolved: Bofrost.de performance has improved in a49e29b after 9 minutes.
|
2025-04-01T04:54:50.502264
| 2023-10-01T04:07:37
|
1920543812
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13565",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/24310"
}
|
gharchive/issue
|
⚠️ Bofrost.de has degraded performance
In 6910c1b, Bofrost.de (https://www.bofrost.de) experienced degraded performance:
HTTP code: 200
Response time: 883 ms
Resolved: Bofrost.de performance has improved in 43e136b after 10 minutes.
|
2025-04-01T04:54:50.505464
| 2023-10-22T03:41:41
|
1955747101
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13566",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/25810"
}
|
gharchive/issue
|
⚠️ Bofrost.at has degraded performance
In 6f3ce82, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 1304 ms
Resolved: Bofrost.at performance has improved in 4b66a1d after 6 minutes.
|
2025-04-01T04:54:50.508812
| 2023-11-04T03:12:58
|
1977140339
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13567",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/26913"
}
|
gharchive/issue
|
⚠️ Bofrost.at has degraded performance
In 310fdc4, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 1837 ms
Resolved: Bofrost.at performance has improved in 84c3895 after 15 minutes.
|
2025-04-01T04:54:50.511730
| 2023-11-07T05:24:03
|
1980572338
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13568",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/27240"
}
|
gharchive/issue
|
⚠️ Bofrost.at has degraded performance
In 67068c4, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 1669 ms
Resolved: Bofrost.at performance has improved in 86156c5 after 6 minutes.
|
2025-04-01T04:54:50.515022
| 2023-11-25T23:55:05
|
2010832531
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13569",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/29673"
}
|
gharchive/issue
|
⚠️ Bofrost.de has degraded performance
In 48068b1, Bofrost.de (https://www.bofrost.de) experienced degraded performance:
HTTP code: 200
Response time: 1744 ms
Resolved: Bofrost.de performance has improved in 5585d4e after 14 minutes.
|
2025-04-01T04:54:50.518130
| 2024-04-05T12:30:24
|
2227920124
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13570",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/40408"
}
|
gharchive/issue
|
⚠️ Bofrost.fr has degraded performance
In a364681, Bofrost.fr (https://www.bofrost.fr) experienced degraded performance:
HTTP code: 200
Response time: 1626 ms
Resolved: Bofrost.fr performance has improved in 2b24081 after 20 minutes.
|
2025-04-01T04:54:50.521073
| 2024-04-16T23:20:29
|
2247023365
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13571",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/41971"
}
|
gharchive/issue
|
⚠️ Bofrost.es has degraded performance
In 57af157, Bofrost.es (https://www.bofrost.es) experienced degraded performance:
HTTP code: 200
Response time: 1104 ms
Resolved: Bofrost.es performance has improved in f81eece after 6 minutes.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.