added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:10:31.362102
| 2023-08-23T04:51:21
|
1862563413
|
{
"authors": [
"MartyG-RealSense",
"Susheelabhat"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14559",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/12129"
}
|
gharchive/issue
|
Recording depth and colour frames
| Camera Model D435i/D455
| Firmware Version
| Operating System & Version Ubuntu 17)
| Kernel Version (Linux Only)
| Platform Raspberry Pi
| SDK Version 2.0
| Language python
| Segment others
Hi, I want to record bag files and I was doing it at 6fps with a lower resolution using Realsense-viewer app.
I want to do it without the application now. Is there a possibility of achieving this through command line?
The reason is I'm using a raspberry PI to which I'm not connecting a monitor/screen , so I have to remotely start the recording and change the parameters.
I'm currently using rs-record -f ./test1.bag -t 60
But how to set fps to 6 and resolution rate here?
Thanks in advance!
Hi @Susheelabhat At the empty line 33 of the rs-record program, you could insert additional 'cfg' configurations line to set the minimum FPS speed for the depth and color streams.
https://github.com/IntelRealSense/librealsense/blob/master/tools/recorder/rs-record.cpp#L33
cfg.enable_stream(RS2_STREAM_COLOR, 848,480, RS2_FORMAT_Z16, 6);
cfg.enable_stream(RS2_STREAM_COLOR,1280,720, RS2_FORMAT_RGB8, 6);
For the D455 camera model, the FPS should be set to 5 instead of 6 as D455 does not support 6 FPS.
If you wish to use a different resolution then also edit the resolution in the cfg lines. For example, to set both streams to 640x480:
cfg.enable_stream(RS2_STREAM_COLOR, 640,480, RS2_FORMAT_Z16, 6);
cfg.enable_stream(RS2_STREAM_COLOR, 640,480, RS2_FORMAT_RGB8, 6);
Thanks for the info @MartyG-RealSense. Wanted to check one more thing,
In the ubuntu system, what is the default location of these librealsense tools/files where I can make changes?
Rather than looking in the 'build' folder, the appropriate location in the source code folder is likely to be librealsense > tools > recorder so that you can edit the contents of the rs-record.cpp file and then build librealsense from that modified source code.
I'll try this @MartyG-RealSense . Another follow up question, with the recorded .bag file, is it possible to extract depth frames outside ubuntu(realsense sdk) (where pyrealsense exists)?
I was using this code to extract frames. But how can I do it without pyrealsense support?
https://github.com/IntelRealSense/librealsense/tree/master/wrappers/python/examples/read_example_bag.py
You can run a tool provided by the RealSense SDK called rs-convert to extract depth frames from a bag file recorded in the SDK. Though it is made from C++ code, it is available pre-built so that you can launch the tool without having to compile the code.
If you included the examples and tools with the flag -DBUILD_GRAPHICAL_EXAMPLES=TRUE when building librealsense with CMake then rs-convert should be in the usr/local/bin folder of Ubuntu.
https://github.com/IntelRealSense/librealsense/tree/master/tools/convert
I tried building librealsense again with the code added in rs-recorder.cpp. But when I now run the command rs-record -f ./test1.bag -t 60 , I'm getting segmentation fault error.
pi@pi2:~/Documents$ rs-record -f /home/pi/Documents/test.bag -t 60
Segmentation fault (core dumped)
Was the added code the cfg lines to set 6 FPS?
Hi @Susheelabhat Do you require further assistance with this case, please? Thanks!
Case closed due to no further comments received.
|
2025-04-01T04:10:31.376924
| 2018-08-31T08:21:26
|
355884870
|
{
"authors": [
"MartyG-RealSense",
"claybudin",
"metanamo",
"nachoreimat",
"tropicdragon"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14560",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/2333"
}
|
gharchive/issue
|
How to use multiple RsDevice in Unity to render pointclouds
Required Info
Camera Model
D415
Camera Firmware
<IP_ADDRESS>
Operating System & Version
Win10
Platform
PC
SDK Version
2.16.0
Language
Unity
Issue Description
I need to use at least 2 RsDevices to render pointcloiuds in unity.
I've tried duplicating the RsDevice object and assigning a different serial number for each one. But still seems the camera initialization does not read the serial number field and takes randomly only one of the connected devices to use the same for both instances. So It is rendering twice the same.
Do I need to build my own custom RsDevice to allow that or is there anything I am doing wrong. It seems there is some comented code on the SDK related to this feature, why is that?
Thanks for the help.
The user frenchtouchXR recently published a method for editing a script in the Unity wrapper to enable multiple devices to be recognized.
https://github.com/IntelRealSense/librealsense/issues/2203#issuecomment-412218681
Yes, this code is already in the Live mode of the SDK, but still the RsDevice acts as a singleton so I can not start more than one device.
That´s the error I get
Exception: RsDevice singleton already instanced
RsDevice.Awake () (at Assets/RealSenseSDK2.0/Scripts/RsDevice.cs:85)
My apologies, I see now that you already commented about your detection issue immediately after frenchtochXR's comment in that linked-to post.
What firmware driver version are your cameras currently using please? Updating to the latest firmware can sometimes help to reduce or resolve problems.
https://downloadcenter.intel.com/download/28076/Latest-Firmware-for-Intel-RealSense-D400-Product-Family?v=t
Thanks for the quick reply.
I am using latest firmware <IP_ADDRESS>
The problem is in the RsDevice class from the SDK as it is checking another instance is running when it awakes.
void Awake()
{
if (Instance != null && Instance != this)
throw new Exception(string.Format("{0} singleton already instanced", this.GetType()));
Instance = this;
}
Any idea how to solve it?
Would it be possible to check whether you have the problem in SDK version 2.15.0 (the previous one)?
Same problem is there at least from sdk version 2.14
RsDevice is a Singleton
Thanks. I asked about trying an earlier version because I had a case over at the Intel Support site where a user had problems using two D415s where only one of them would work properly, but their project worked fine when three cameras were connected.
https://communities.intel.com/message/563718#563718
This is likely an issue that one of the Intel development staff on this forum such as Dorodnic or RealSense-Customer-Engineering can answer, unfortunately.
Thanks so much for your help on this case! :)
I published this method a couple of weeks ago as an answer to an issue.
@frenchtouchXR could you send me the link to this publication? thanks
Hi,
I've turned back to unity test to give it another try. And I'm still facing the same problem when trying to stream multiple D415 to Unity. "RSDevice Singleton already instanced..."
Using the latest SDK/Unitypackage and latest Unity version this should be solved...I was assuming.
I have no clue what I'm doing wrong here. I just add 2 devices with pointcloud to unity. One is working but two cause the error. devide is set to public (static button deactivated in unity)
Well, I guess this issue remains since MAY?
Heya. The problem comes from the rsdevice class is static. You need to make it public only and correct all the errors. Then use the s/n and create new shaders. I published this method a couple of weeks ago as an answer to an issue.
Do you happen to still have a link to this method you published ? I am interested in it, since I am facing multi-cam issues in Unity as well.
It is probably this link:
https://github.com/IntelRealSense/librealsense/issues/2203?language=en_US#issuecomment-412218681
That link doesn't have all of the information about how to do it, and a lot of the links off of that link are broken (it looks like Intel changed its Forum format).
@frenchtouchXR (or anyone else): do you have the code you used to address the issue? I made RsDevice non-static, which introduced a bunch of errors, many in other files that now need an instance to reference. There's also a CustomProcessingBlock reference which apparently can take a static variable but not an instanced one. I could probably puzzle all of this out, but I'm not an expert C# programmer and I don't have a lot of time, so if someone has figured all this out and can send me the source files, I would appreciate it.
The link you provide doesn't address the process described above (make the RsDevice non-static then fix the errors that result). I too would love to get the code with the changes.
|
2025-04-01T04:10:31.381384
| 2017-05-15T04:12:30
|
228602197
|
{
"authors": [
"dorodnic",
"nizihabi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14561",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/497"
}
|
gharchive/issue
|
Can I use librealsense under Windows 7 x64?
Required Info
Camera Model
SR300
Firmware Version
Operating System & Version
Windows 7 x64
Kernel Version (Linux Only)
Build System
VS2013
when I used this to initial sr300 under the Win7, it got "Unable to open device via WinUSB" error, does it support Windows 7?
Hi,
librealsense does not support Windows 7 out of the box. The reason you are getting the error is because SR300 relies on a standard WinUSB driver. On later versions you see it as:
However, Win7 does not recognize this device. It is not impossible to get it to show-up, and then librealsense will most likely work fine. Google enabling WinUSB on Windows 7.
That said, Windows 7 has several severe bugs inside Media Foundation core that Microsoft is not planning to patch, so proceed with caution.
ok , thanks a lot : )
|
2025-04-01T04:10:31.385957
| 2020-10-10T16:17:05
|
718652340
|
{
"authors": [
"MartyG-RealSense",
"a-m-p-m"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14562",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/7539"
}
|
gharchive/issue
|
questions about camera settings
Hello,
Could you please point me to:
A full list of the camera settings for various RealSense cameras that can be set via camera's API (D435i, D435, and D415 cameras are of particular interest); and
Information about the camera settings that affect other camera settings (for example, I see that setting Enable Auto Exposure to 1 first and then setting Exposure to an exposure value leads to Enable Auto Exposure setting on the camera switched to 0; are there any other parameters like these ones that affect each other?).
Thanks.
There is an excellent unofficial documentation website that draws its content from the official documentation and enables SDK functions to be searched by type using menus at the top of the page.
https://unanancyowen.github.io/librealsense2_apireference/classes.html
It should be noted though that Advanced Mode functions are not fully documented. This is because they interact with each other in complex ways, so Intel controls them with machine-learning algorithms instead of providing documentation. RealSense users are free to do trial and error experimentation with Advanced Mode functions to see how changes affect the image though.
Other than auto-exposure deactivating when exposure or gain is manually changed, the other function that comes to mind that is affected by value changes in an option is that if the Laser Power option is set to '0' then the laser turns off.
https://github.com/IntelRealSense/librealsense/wiki/API-How-To#controlling-the-laser
Thank you, @MartyG-RealSense
Hi @a-m-p-m Do you still require assistance with this case, please? Thanks!
Case closed due to no fiurther comments received.
|
2025-04-01T04:10:31.412546
| 2020-11-16T08:15:42
|
743624481
|
{
"authors": [
"MartyG-RealSense",
"weiyi-jie",
"yzkzero"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14563",
"repo": "IntelRealSense/librealsense",
"url": "https://github.com/IntelRealSense/librealsense/issues/7777"
}
|
gharchive/issue
|
On the data disorder of d455 depth image
In my normal use process, the camera d455 has the phenomenon of depth data dislocation. When I restart the device, it has a certain probability to repair and output complete and correct depth data. (this phenomenon is completely random).
My version number: librealsense2.so.2.39.0
I would like to ask you what causes this phenomenon and how to avoid it. Thank you very much!
Hi @weiyi-jie Could you provide more information please about what you mean by 'depth data dislocation'. Does the depth data freeze or disappear? And are you using the camera indoors or outdoors? Thank you very much.
Hello, I use it indoors. The depth image data has not been frozen, and the real-time depth data is still being transmitted. Only when the image is displayed, there is a long transverse crack, the whole image is split and moved downward, or the image acquisition is incomplete, only part of the image can be obtained, and the screen flickers.Hello, I use it indoors. The depth image data has not been frozen, and the real-time depth data is still being transmitted. Only when the image is displayed, there is a long transverse crack, the whole image is split and moved downward, or the image acquisition is incomplete, only part of the image can be obtained, and the screen flickers.
Hello, @MartyG-RealSense I use it indoors. The depth image data has not been frozen, and the real-time depth data is still being transmitted. Only when the image is displayed, there is a long transverse crack, the whole image is split and moved downward, or the image acquisition is incomplete, only part of the image can be obtained, and the screen flickers.
Hi @weiyi-jie Could you provide more information please about what you mean by 'depth data dislocation'. Does the depth data freeze or disappear? And are you using the camera indoors or outdoors? Thank you very much.
Hello, @MartyG-RealSense I use it indoors. The depth image data has not been frozen, and the real-time depth data is still being transmitted. Only when the image is displayed, there is a long transverse crack, the whole image is split and moved downward, or the image acquisition is incomplete, only part of the image can be obtained, and the screen flickers.
Hi @weiyi-jie Could you provide more information please about what you mean by 'depth data dislocation'. Does the depth data freeze or disappear? And are you using the camera indoors or outdoors? Thank you very much.
Hello, @MartyG-RealSense I use it indoors. The depth image data has not been frozen, and the real-time depth data is still being transmitted. Only when the image is displayed, there is a long transverse crack, the whole image is split and moved downward, or the image acquisition is incomplete, only part of the image can be obtained, and the screen flickers.
Hi @weiyi-jie I tried to visualise what you are experiencing. It would be very helpful if you could provide an image in this discussion though if that is possible.
Does the flickering depth image that you are experiencing look anything like the one in the link below please:
https://github.com/IntelRealSense/librealsense/issues/7268
HI @MartyG-RealSense, I'm sorry, it was my negligence that I forgot to upload the problem image. My problem image is shown in the figure below. There will be all these phenomena
Here is the normal image
The camera distance is 3.9m
Are you using the RealSense Viewer program to generate images, please? If you are using the Viewer, could you check please whether you have a filter enabled in the Post-Processing controls that is called Threshold Filter. If this filter is enabled - shown by a blue (On) icon beside the Threshold Filter name - then by default it limits the D455 camera's observable depth range to only 4 meters.
You can either left-click on the blue icon to change its color to red (Off) or use the Maximum Distance slider of this filter to set it to its maximum distance.
Sorry, @ MartyG-RealSense I didn't use realsense viewer. I used demo program in librealsense master / build / examples. I can't find this problem in realsense viewer for the time being (probably due to too few tests).
I will test as you describe.
Please do update me when you have done tests. Thanks!
@ MartyG-RealSense Hello, I set the settings according to what you said. I continued to test without any repair, but I used the realsense viewer without any error. Is it true that the depth data acquisition methods of realsense viewer are different? How was it acquired? Can I get its data acquisition program?
The RealSense Viewer is open-source code that you can modify for your own needs.
https://github.com/IntelRealSense/librealsense/tree/master/tools/realsense-viewer
You may certainly experience differences in the images that you get from a program that you create yourself, compared to the images achieved in the RealSense SDK. This is partly because the RealSense Viewer enables a lot of functions such as post-processing and colorization by default, whereas these functions have to be manually programmed into your application when creating your own scripts. Example programs that do not include certain features like the Viewer does may also generate different-looking images.
I ran some tests with the Viewer, pointing the camera at a set of windows as you seem to be doing in the image above. The windows were visible normally if Auto-Exposure was enabled in the Stereo Module controls of the Viewer, and the windows disappeared if Auto-Exposure was disabled. As the Viewer has auto-exposure enabled by default when it launches, I wonder if the example that you used did not have Auto-Exposure enabled.
@Martyg realsense Sorry to reply you now. I modified it as you said, and then tested it many times without any problems; but when I power down the camera and plug it in again, it will happen again. Now my camera will report the following error message:
24/11 16:35:02,671 WARNING [140128753518336] ( sensor.cpp:632 ) Exception was thrown when inspecting Raw RGB Camera property Controls white balance of color image. Setting any value will disable auto white balance
24/11 16:35:02,935 WARNING [140128921757504] ( types.cpp:78 ) hwmon command 0x7d( 0 0 0 0 ) failed (response -21= No data to return)
I will seek advice from Intel about your latest problem with the above errors appearing. Thanks very much for your patience!
I received guidance from Intel. Could you confirm please in the RealSense Viewer if the depth can stream without any problems but the RGB cannot stream.
Yes, in realsense viewer, depth can be streamed without any problems, and RGB can be streamed normally.
Thanks very much. Are you still experiencing the WARNING messages or did they only happen once when plugging the camera in.
Yes, there will always be a warning message at the terminal, not only at the first startup.
I reviewed your case agan from the beginning for new insights. If there is a probability of the camera functioning correctly with your Python script if the camera is reset, could you try adding a hardware_reset() instruction to your script please? This should reset the camera automatically when the program launches, like when the Hardware Reset option is selected from the 'More' menu in the RealSense Viewer.
https://github.com/IntelRealSense/librealsense/issues/3329#issuecomment-475293475
Hi Weiyi-jie,
Hi MartyG,
I think the problem is very similar with what I met and mentioned in:
https://support.intelrealsense.com/hc/en-us/community/posts/360051428274-Incorrect-depth-image-by-D455
From your screenshot, I also noticed the "band and shifting" phenomenon in this reply:
https://github.com/IntelRealSense/librealsense/issues/7777#issuecomment-729343146
So it looks like a common issue of D455. It happens very frequently and it should be solved as soon as possible.
@yzkzero I do not want to make a comment that pre-empts any official technical information that Intel releases for the D455. I do understand your frustration and ask for your kind patience, please.
Hi MartyG,I'm sorry to reply you now. I just came back from a business trip. I tested this method as you said at the first time. So far, there are no problems before the occurrence. Thank you very much for your help, thank you!
@weiyi-jie That's great to hear - thanks very much for the update!
Hi @weiyi-jie Do you require further assistance with this case, please? Thanks!
HI @MartyG-RealSense Thank you very much for your help. This problem does not appear at present.
Hi @MartyG-RealSense , I'd like to ask how to set the depth data to obtain a relatively complete depth image in the sun.
I use realsense_ The viewer can get relatively complete data by setting preset: default. How to set this mode in the program?
The first image is the depth
/media/xj/Work/222.jpg
preset: default
/media/xj/Work/111.jpg
Is it necessary to set something in the program? This is the result. Thanks!
/media/xj/Work/111.jpg
The first image is the depth
preset: default
preset: default
Is it necessary to set something in the program? This is the result. Thanks!
If you wish to set the SDK's Default preset with scripting, there is a method for loading existing presets supplied in the SDK that is simpler than the code that is typically involved in loading a custom-made preset.
In the example code in the link below, change RS2_RS400_VISUAL_PRESET_HIGH_DENSITY to RS2_RS400_VISUAL_PRESET_DEFAULT to load the Default preset.
https://github.com/IntelRealSense/librealsense/issues/7399#issuecomment-697609404
If you want to program in Python, the link below should be a useful reference.
https://github.com/IntelRealSense/librealsense/issues/2577#issuecomment-432137634
If a preset is not defined then the camera should automatically use the default settings for that particular camera model. Some prefer the certainty of defining a preset setting though to be reassured that the camera is being configured precisely how they need it to be.
This is feasible. Thank you for your help @MartyG-RealSense
Hi @weiyi-jie Do you require further assistance with this case, please? Thanks!
@MartyG-RealSense Thank you for your help. It's not available now. Thank you.
Great news @weiyi-jie - thanks for the update! I will close this case now as you have achieved a solution.
|
2025-04-01T04:10:32.097539
| 2023-10-10T12:09:35
|
1935200417
|
{
"authors": [
"lucasjinreal",
"myownskyW7"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14564",
"repo": "InternLM/InternLM-XComposer",
"url": "https://github.com/InternLM/InternLM-XComposer/issues/20"
}
|
gharchive/issue
|
training data?
Will release training data?
Part of our private training data is released at https://github.com/opendatalab/WanJuan1.0.
@myownskyW7 what about else, how many data overall used in training? and what's the postion of English and Chinese data?
Here are details about our pretraining and fine-tuning data.
1. Pretraining Data:
2. Finetuning Data:
Please refer to our technical report for other details.
|
2025-04-01T04:10:32.099674
| 2021-05-25T09:08:44
|
900528539
|
{
"authors": [
"stefkoslowski",
"tmberthold"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14565",
"repo": "International-Data-Spaces-Association/IDS-Messaging-Services",
"url": "https://github.com/International-Data-Spaces-Association/IDS-Messaging-Services/issues/132"
}
|
gharchive/issue
|
Ignore Message Validation when DSC is in TEST_DEPLOYMENT
Change Errors occuring if DAT is not valid to Warning Log Level if TEST_DEPLOYMENT is activated
Suggestion submitted: When booting up the connector (and thus the messaging-services), log/print a warning that you are in TEST_DEPLOYMENT directly at startup and what the consequences are (DAT validity check is not done, errors on the certificate are ignored, etc). Instead of printing the error at DAPS DAT methods
|
2025-04-01T04:10:32.113982
| 2023-09-29T22:08:52
|
1919960667
|
{
"authors": [
"lehins"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14566",
"repo": "IntersectMBO/cardano-ledger",
"url": "https://github.com/IntersectMBO/cardano-ledger/issues/3773"
}
|
gharchive/issue
|
Create Haskell DSL for CDDL
We need a DSL that we can specify our CDDL spec in and use it to:
Generate CDDL files
Generate random binary data and test our decoders
The new DSL has been implemented as cuddle library and a huddle tool in the https://github.com/input-output-hk/cuddle repo
Here are the tickets that track using the new tooling:
#4126
#4125
This ticket can be considered done
|
2025-04-01T04:10:32.118153
| 2024-09-10T13:51:39
|
2516546155
|
{
"authors": [
"Ryun1",
"bosko-m"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14567",
"repo": "IntersectMBO/govtool",
"url": "https://github.com/IntersectMBO/govtool/issues/2000"
}
|
gharchive/issue
|
🐛 Ada holder voting power does not take into account unclaimed rewards OR gov action deposits paid
Area
Delegation Pillar
Domain
gov.tools
Which wallet were you using?
No response
Context
Ada holder voting power is equal to;
total of lovelace UTxOs controlled by the stake account
total of unclaimed lovelace from the reward account (from staking rewards or treasury withdrawals)
total lovelace locked for governance action deposit
see example stake credential (mine)
e1130796e23f5c56594f50ace2c3240923219fc69ee2cb6e2c8facec68
Steps to reproduce
check voting power for e1130796e23f5c56594f50ace2c3240923219fc69ee2cb6e2c8facec68
Actual behavior
resulting value it: wallet balance - unclaimed rewards
Expected behavior
resulting value it: wallet balance + unclaimed rewards
Im putting a P0 on this @Ryun1 .
|
2025-04-01T04:10:32.119535
| 2024-10-14T22:41:17
|
2587205000
|
{
"authors": [
"jorisdral"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14568",
"repo": "IntersectMBO/lsm-tree",
"url": "https://github.com/IntersectMBO/lsm-tree/pull/433"
}
|
gharchive/pull-request
|
Fix a race condition in RWVar
retry is interruptible, which can leave the RWVar in a state that is unrecoverable.
Kudos to @jasagredo for finding this subtle bug!
This is now ready for review. My takeaway is that I should put more thought into interruptibility when writing exception safe code :stuck_out_tongue_closed_eyes:
|
2025-04-01T04:10:32.121746
| 2024-05-03T13:23:48
|
2277677992
|
{
"authors": [
"abishekrsrikaanth",
"olivervogel"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14569",
"repo": "Intervention/image",
"url": "https://github.com/Intervention/image/issues/1344"
}
|
gharchive/issue
|
Is there a way to strip or clear exif data when encoding a file?
I am using the imagick driver. Is there a way to strip or clear exif data when encoding a file?
I am trying to clear all exif data before saving the file to the Storage Driver.
This is not possible with the current range of features. There is already an issue on this topic: https://github.com/Intervention/image/issues/1110
Will calling the stripImage() using the the core()->native() strip the metadata. I created a modifier to call this before encoding the file as below. Will this work?
public function apply(ImageInterface $image): ImageInterface
{
$imagick = $image->core()->native();
$imagick->stripImage();
}
|
2025-04-01T04:10:32.131084
| 2024-06-17T10:38:13
|
2357006332
|
{
"authors": [
"matthieuuu",
"olivervogel",
"pierrejoye"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14570",
"repo": "Intervention/image",
"url": "https://github.com/Intervention/image/issues/1370"
}
|
gharchive/issue
|
Failed image format detection
Describe the bug
$image = $manager->read('logo.png');
failed to detect a valid PNG file.
Code Example
use Intervention\Image\ImageManager;
use Intervention\Image\Drivers\Gd\Driver;
// create image manager with desired driver
$manager = new ImageManager( new Driver());
$image = $manager->read('test.png');
$image->toPng()->save('foo.png');
Expected behavior
logo.png loaded and saved as foo.png.
The reason is if there is no exif data, it is not a corrupt png, but exif_read_data will fail (horrible API):
$ exif logo.png
Corrupt data
The data provided does not follow the specification.
ExifLoader: The data supplied does not seem to contain EXIF data.
$ php -r "exif_read_data('logo.png');"
PHP Warning: exif_read_data(logo.png): File not supported in Command line code on line 1
PHP Stack trace:
PHP 1. {main}() Command line code:0
PHP 2. exif_read_data($file = 'logo.png') Command line code:1
While the imagesize function does the format validation (signature+header):
php -r 'print_r(getimagesize("test.png"));'
Array
(
[0] => 434
[1] => 83
[2] => 3
[3] => width="434" height="83"
[bits] => 8
[mime] => image/png
)
If fileinfo is available, I would recommed to use it instead:
php finfo.php
image/png; charset=binary%
using
<?php
$finfo = new finfo(FILEINFO_MIME);
echo $finfo->file('test.png');
However getimagesize is good enough for a sanity check (signatures can be tricked, so check of w/h could be done, GD/imagic do it as well).
It was working on older version of this package, so not really sure when this exif_read_data was introduced :)
Images
(any without exif would do it but here is one)
Environment (please complete the following information):
PHP Version: 8.3.8
OS: Linux/Windows/FreeBsd
Intervention Image Version: 3.7
GD
Thanks for the bug report. You are right, the xxx function gives a warning when reading one of the mentioned images. I'm just wondering why you see the warning since it should be suppressed. Do you have other settings active that disable the error control operator?
Hello! Just ran into the same issue :) I use an error handler with set_error_handler() function, and the error was not suppressed since I didn't check error_reporting() return. I fixed it and now everything works fine, the error is suppressed as intended!
Thank you @matthieuuu . I suspected something similar. However, the implementation in the library can still be improved, as certain file formats cannot contain any EXIF data at all, but exif_read_data is still called. I will try to improve this.
Hello,
Sorry for the late reply, been busy :)
Yes, it was the issue.
However I would still recommend not to use this function but to actually read the exif information. I suspect most users don't need these exif but on specific cases.
For the purpose of selecting a codec, getimagesize or fileinfo would do a faster job.
I have the code for a generic imageload (which provides the type as part of the gd instance then) in libgd, I need to do a release and add it to php, that would be even better :-)
As of version 3.7.1, calls to exif_read_data are avoided for PNG formats and other formats that do not support EXIF data. Adjustments to a custom error handler are no longer absolutely necessary.
@olivervogel Danke! :)
|
2025-04-01T04:10:32.279471
| 2020-07-07T16:40:52
|
652466538
|
{
"authors": [
"dannylamb",
"nigelgbanks"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14575",
"repo": "Islandora-Devops/isle-buildkit",
"url": "https://github.com/Islandora-Devops/isle-buildkit/pull/65"
}
|
gharchive/pull-request
|
Add support for postgres where possible, and now support database bot…
…h database for fcrepo.
This is significantly bigger pull as I've refactored as much of the database logic down into the base image so it can be shared by all the services.
This required changing the defaults from 'database' to the specific implementation since 'matomo' does not work with postgresql we can't assume only one database in the stack (if users choose to use postgres) so the defaults have changed to 'mariadb'.
To be tested along with https://github.com/Islandora-Devops/isle-dc/pull/86
Address https://github.com/Islandora-Devops/isle-buildkit/issues/9
@nigelgbanks Can you resolve the conflicts on this? Normally I'd do it myself but Github's not letting me through the UI.
@dannylamb will do
@dannylamb ready to go
Testing now :rocket:
|
2025-04-01T04:10:32.299794
| 2018-09-05T00:07:31
|
357019065
|
{
"authors": [
"codecov-io",
"tapashmajumder"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14576",
"repo": "Iterable/swift-sdk",
"url": "https://github.com/Iterable/swift-sdk/pull/22"
}
|
gharchive/pull-request
|
[ITBL-6124] - logging
add logLevel and IterableLogDelegate to IterableConfig.
Codecov Report
Merging #22 into master will increase coverage by 0.13%.
The diff coverage is 83.67%.
@@ Coverage Diff @@
## master #22 +/- ##
==========================================
+ Coverage 69.09% 69.23% +0.13%
==========================================
Files 24 25 +1
Lines 1851 1872 +21
==========================================
+ Hits 1279 1296 +17
- Misses 572 576 +4
Impacted Files
Coverage Δ
swift-sdk/IterableUtil.swift
95.23% <ø> (+1.36%)
:arrow_up:
swift-sdk/IterableLogging.swift
83.67% <83.67%> (ø)
swift-sdk/IterableAPI.swift
30.95% <0%> (+2.38%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 80410a9...357f62e. Read the comment docs.
|
2025-04-01T04:10:32.307893
| 2024-07-23T08:50:04
|
2424632735
|
{
"authors": [
"Bhattu-Anitha-07"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14577",
"repo": "Its-Aman-Yadav/Community-Site",
"url": "https://github.com/Its-Aman-Yadav/Community-Site/issues/1202"
}
|
gharchive/issue
|
Validating email using firebase
Describe the feature
Hello PA,
This issue is about validating email using firebase.
This provides users to login with valid email, that prevents unauthorized access.
Email verification is done by sending a verification email to registered email.
Add ScreenShots
Record
[X] I agree to follow this project's Code of Conduct
[X] I'm a GSSOC'24 contributor
[X] I want to work on this issue
@Its-Aman-Yadav
Kindly assign this issue to me, also add labels.
Will work on this issue.
Thank you :)
@Its-Aman-Yadav
Can you please assign this issue to me?
@Its-Aman-Yadav
please assign this issue to me.
Thank you.
@Its-Aman-Yadav
Kindly assign this issue to me.
Thank you :)
@Its-Aman-Yadav
Please assign this issue to me, Will work on this issue.
Thank you.
|
2025-04-01T04:10:32.310485
| 2024-05-25T14:41:40
|
2317065602
|
{
"authors": [
"cxde-Sujal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14578",
"repo": "Its-Aman-Yadav/Community-Site",
"url": "https://github.com/Its-Aman-Yadav/Community-Site/issues/420"
}
|
gharchive/issue
|
BUG: On hover in footer links : it must change to original icon colors
Is there an existing issue for this?
[X] I have searched the existing issues
What happened?
On hover ,
all get blue colors but they must change to their original icon colors , like linkedIn is getting blue color , so instagram must get pink and other icons also similarly
Record
[X] I agree to follow this project's Code of Conduct
[X] I'm a GSSOC'24 contributor
[X] I want to work on this issue
@Its-Aman-Yadav please assign it to me for GSSoC
|
2025-04-01T04:10:32.313205
| 2024-05-14T11:32:12
|
2295183117
|
{
"authors": [
"DevNandini02",
"Its-Aman-Yadav"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14579",
"repo": "Its-Aman-Yadav/Community-Site",
"url": "https://github.com/Its-Aman-Yadav/Community-Site/pull/193"
}
|
gharchive/pull-request
|
Updated images to slideshow #128
I have removed the previous images and added darker images to the slideshow. #128
After changes
Open Source Village.webm
Please open the link.
Thank you @DevNandini02 just remove the conflicts before merging
Thank you @DevNandini02 just remove the conflicts before merging
@Its-Aman-Yadav sir i am not able to resolve the conflicts in the same branch. Can i make another branch and then send the pull request again?
yeah sure
@Its-Aman-Yadav sir since there is no conflicts now, so do i have to merge the updates to the main branch or you'll do it?
Thank you @DevNandini02
|
2025-04-01T04:10:32.327017
| 2014-10-05T12:35:23
|
44913096
|
{
"authors": [
"StevenPuttemans",
"askerpro",
"upperwal",
"vpisarev"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14580",
"repo": "Itseez/opencv",
"url": "https://github.com/Itseez/opencv/pull/3306"
}
|
gharchive/pull-request
|
Added New Sample: Stereo Camera Calibration in Real Time
Captures images in real time by detecting chessboard corners and
calibrate the stereo camera after that.
Keyboard Shortcuts:
Initial Mode: Detecting (Chessboard Corners)
'c': Capturing (Start capturing the stereo images)
'p' Calibrate (Once all the images are captured, calibration can start by pressing 'p')
@vpisarev hey! Any update? I have been eagerly waiting for this pr to get merged.
@upperwal it seemed that @vpisarev was on holidays until 1 or 2 days ago. He has been back and is graduatly merging pull requests. Just wait a little bit more! I am sure it will get merged since it is a nice addition.
Also, please merge both commits, else this will not get accepted!
Thanks for replying @StevenPuttemans :) and I will definitely merge both the commits asap. :D
Hello @upperwal, terribly sorry for long response. I would prefer if this sample will be combined with the offline stereo_calib. We are trying to cleanup OpenCV samples, remove duplicates or close variations, so it would be nice if we have a single calibration sample. This may be difficult, then I'd prefer that at least monocular camera calibration and stereo calibration are 2 samples, not 3 or more.
Hey @vpisarev, thanks for replying and don't be sorry, I know you guys are very busy. OpenCV samples contain 3 type of calibration files 3calibration.cpp, calibration.cpp and stereo_calib.cpp. Are you suggesting to merge all these samples into one? As my sample is an extension to stereo_calib.cpp (which calibrate using pre saved stereo images, where as my code works on the live video feed) Will it be fine to merge these two samples and let the user decide which method to use?
@upperwal, yes, I meant merging your sample and stereo_calib.cpp. 3calibration.cpp can probably be removed and we plan to extend calibration.cpp. So, there is no need to merge everything together, just retaining a single stereo calibration sample would be good enough solution for now.
:+1: got that. Will try to merge my sample with stereo_calib.cpp and then I will get back to you. Thanks :)
@vpisarev @StevenPuttemans
What from the following is advisable, to load stereo images in the program:
Maintain a images name file (such as images.yml) from which name of an image is read and that string is used in imread().
Ask the user to give a folder, image prefix (Ex: "image_left_", "image_right_"), image post fix (Ex: "jpg") and no of images and iterate using image prefix+i+image post fix to read the images. Given image are image_left_1.jpg, image_right_1.jpg etc.
I am sorry if this is a lame question, but I came across the first method in almost all samples and according to me the second one is more efficient to do. (user does not need to create a file for the name of the images). Please correct me, if I am wrong.
Thanks
@upperwal it kind of depends I guess. Since you are building the sample, you can choose the method you like. The reason why people use a name file with locations, is because reading files is universal in different OS. Reading from a folder gives different sequences since ordering is different on several OS.
But hey, you choose what fits you the best!
Thanks @StevenPuttemans :+1: I used the second method for the time being. To have a single sample for stereo calibration I have merged stereo_calib with my sample. The sample can now work on already saved stereo images or click the images on the go.
Great :+1:
Could you also use your PR to delete the double samples?
@StevenPuttemans As stereo_calib is updated, there are two calibration samples left. 1. "3calibration.cpp" and 2. "calibration.cpp". So "3calibration.cpp" should be deleted as suggested by @vpisarev as "calibration.cpp" is used for single camera calibration. Correct?
@vpisarev hi. Can you please review the code?
@upperwal I know it is taking some time, but please be patient, I have the idea that some of the devs are on holidays or taking a break :)
:+1: @StevenPuttemans
Thanks @StevenPuttemans. You are a great support :+1: :)
got an error with imread
cv::imwrite could not find a writer for the specified extension
//std::ostringstream leftString, rightString;
//leftString << dir << "/" << prefixLeft << pairIndex << postfix;
//rightString << dir << "/" << prefixRight << pairIndex << postfix;
std::string leftString = dir + "/" + prefixLeft + to_string(pairIndex) + "." + postfix;
std::string rightString = dir + "/" + prefixRight + to_string(pairIndex) + "." + postfix;
imwrite(leftString, leftImage);
imwrite(rightString, rightImage);
works with this
@askerpro :+1: will fix it ASAP. Thanks :)
@upperwal I do not think you should close it down. Just update this branch?
Hey @StevenPuttemans This PR already have 9 commits, I was thinking of creating a new PR after resolving the merge conflict and incorporating the above suggestion. Will that be fine?
@upperwal there is a very nice tool in git called git rebase -i HEAD~9 (change number accordingly) which allows you to concatenate commits and make it a clean PR again. No needc in making a new branch for the fixes!
@StevenPuttemans Very well then. Let me try it and reopen this PR. Thanks Steven. :)
Sorry @StevenPuttemans I tried using rebase but it was all messed up and then I did a force push to my repo. I think, now I need to create a new PR for this.
Ok no problem at least you tried :D
:) Thanks @StevenPuttemans
|
2025-04-01T04:10:32.333352
| 2024-04-02T01:38:35
|
2219399517
|
{
"authors": [
"GreeSpirit",
"balinorLiang"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14581",
"repo": "IvorySQL/Ivory-www",
"url": "https://github.com/IvorySQL/Ivory-www/pull/114"
}
|
gharchive/pull-request
|
内容更新
1、第一个提交是测试,不影响外部内容
2、第二个提交是修改背景图片
@GreeSpirit At first, thanks for your PR. However, I need to mention that, we need to follow IvorySQL Community Contribution Guide. The source branch needs to be checkouted from main/master branch,
Hi @GreeSpirit, At first, thanks for your PR.
However, it's important to note that we should adhere to the IvorySQL Community Contribution Guide since it's an open-source project.
Regarding this PR, there are similar issues as follows:
1.The source branch couldn’t be main/master from your forked project, instead, we can create a new branch from main/master as the source branch
2.The test commit is unnecessary, please remove it. Additionally, in open source community, we prefer the use of English for commit information, as well as PR details and comments.
|
2025-04-01T04:10:32.346634
| 2016-06-12T08:58:52
|
159816529
|
{
"authors": [
"J7mbo",
"ThatGamerBlue"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14582",
"repo": "J7mbo/twitter-api-php",
"url": "https://github.com/J7mbo/twitter-api-php/issues/195"
}
|
gharchive/issue
|
oauth_access_token_secret
Do I have to pass the oauth secret, what if I don't have it? Do I pass it anything I want to?
Sorry I didn't get around to this in time! I'm just cleaning up issues. If you still need help with this, please feel free to post, otherwise closing for now.
still need help pls, ty
You need all four values, as per the twitter documentation. If you don't have it, you need to get it from the twitter site.
ok, thanks
|
2025-04-01T04:10:32.355372
| 2022-10-25T15:50:18
|
1422701244
|
{
"authors": [
"danholdaway",
"travissluka"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14583",
"repo": "JCSDA-internal/oasim",
"url": "https://github.com/JCSDA-internal/oasim/pull/1"
}
|
gharchive/pull-request
|
Feature/oasim
Description
Add initial code to develop. (I take zero credit for doing this!)
Required by:
https://github.com/JCSDA-internal/ufo/pull/2339
https://github.com/JCSDA-internal/ufo-bundle/pull/83
Dependencies
None: can be merged at any time, does not require coordination.
Testing
when built as part of the ufo-bundle, the ctest oasim_test passes (with the below mentioned uncommitted change added)
@danholdaway, There is one bug fix to get the ctest to work. I don't have write permission on this repo so I can't push it. (and I can't request reviewers either :man_shrugging: )
@travissluka I just made you an admin
|
2025-04-01T04:10:32.386482
| 2015-09-17T12:46:22
|
106976040
|
{
"authors": [
"eyalbe4",
"huerlisi",
"pvblicltd"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14584",
"repo": "JFrogDev/bower-art-resolver",
"url": "https://github.com/JFrogDev/bower-art-resolver/issues/6"
}
|
gharchive/issue
|
DEPTH_ZERO_SELF_SIGNED_CERT self signed certificate
Node v0.12.7 - We can't get resolver to complete a bower install. We think it may be related to this issue with the "request" library https://github.com/request/request/issues/418.
Has anyone else experienced the same problems?
This could be solved by using the strictSSL=false together with the fix in #3. This PR fixes the bower configuration not being passed to the actual request.
@pvblicltd,
Can you please approve that version 2.0.2 fixes this issue by using the strictSSL=false as @huerlisi suggested?
@pvblicltd,
May we close this issue?
@pvblicltd and all, I'm closing this issue for now. Do let us know if you still experience this issue.
|
2025-04-01T04:10:32.420371
| 2024-10-21T14:19:02
|
2602708574
|
{
"authors": [
"JLSteenwyk",
"aebaci"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14586",
"repo": "JLSteenwyk/ClipKIT",
"url": "https://github.com/JLSteenwyk/ClipKIT/issues/59"
}
|
gharchive/issue
|
[YYYY-MM-DD]: [BUG TITLE] cannot read phylip or maf files
Bug Summary
I have a large alignment file of CDS (248 Mb in maf format and 238 Mb in phylip format). I need to remove gaps for using it in codeml, but I get an error with both.
If I don't specify the format, the error is the following:
(venv) (base) antonella@obelix:~/proteins_decont/CDS/fixedCDS$ clipkit output_alignmentwnnA.phy
Format type could not be read.
Please check acceptable input file formats: fasta, clustal, maf, mauve, phylip, phylip_sequential, phylip_relaxed, stockholm
Traceback (most recent call last):
File "/home/antonella/venv/bin/clipkit", line 8, in
sys.exit(main())
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 220, in main
execute(**process_args(args))
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 166, in execute
trim_run, stats = run(
TypeError: cannot unpack non-iterable NoneType object
(venv) (base) antonella@obelix:~/proteins_decont/CDS/fixedCDS$ clipkit output_alignmentwnnA.phy -if phylip -co -s nt -l
Traceback (most recent call last):
File "/home/antonella/venv/bin/clipkit", line 8, in
sys.exit(main())
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 220, in main
execute(**process_args(args))
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 166, in execute
trim_run, stats = run(
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 82, in run
alignment, input_file_format = get_alignment_and_format(
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/files.py", line 32, in get_alignment_and_format
alignment = AlignIO.read(open(input_file_name), file_format.value)
File "/home/antonella/venv/lib/python3.9/site-packages/Bio/AlignIO/init.py", line 389, in read
raise ValueError("More than one record found in handle")
ValueError: More than one record found in handle
(venv) (base) antonella@obelix:~/proteins_decont/CDS/fixedCDS$ clipkit CDSsfinaleswnamenA.maf -if maf -co -s nt -l
Traceback (most recent call last):
File "/home/antonella/venv/bin/clipkit", line 8, in
sys.exit(main())
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 220, in main
execute(**process_args(args))
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 166, in execute
trim_run, stats = run(
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/clipkit.py", line 82, in run
alignment, input_file_format = get_alignment_and_format(
File "/home/antonella/venv/lib/python3.9/site-packages/clipkit/files.py", line 32, in get_alignment_and_format
alignment = AlignIO.read(open(input_file_name), file_format.value)
File "/home/antonella/venv/lib/python3.9/site-packages/Bio/AlignIO/init.py", line 389, in read
raise ValueError("More than one record found in handle")
ValueError: More than one record found in handle
Technical Details
Python version: Run python --version.
Python 3.9.19
ClipKIT version: Run clipkit -v
clipkit 2.3.0
Hi @aebaci,
Thank you for bringing this issue to our attention. We apologize for the inconvenience.
To facilitate resolving this issue, would you be willing to share your input files?
Please feel free to send them an email if you'd like https://jlsteenwyk.com/contact.html.
Best,
Jacob
Hi @aebaci,
Thank you for sending the file via email.
I looked at the input file and noticed that multiple alignment files have been concatenated together (specifically, it looks like the input was generated using something like this command: cat single_gene_alignments* > msa.file).
The input of ClipKIT is a single multiple sequence alignment file at a time; e.g., clipkit single_gene_alignment.
If there is any other confusion, please let me know and I would be happy to help you further.
Cheers,
Jacob
|
2025-04-01T04:10:32.432769
| 2023-06-07T01:53:36
|
1744890533
|
{
"authors": [
"JLSteenwyk",
"hanxiaoxu1110"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14587",
"repo": "JLSteenwyk/orthosnap",
"url": "https://github.com/JLSteenwyk/orthosnap/issues/4"
}
|
gharchive/issue
|
Single-copy orthologous genes identified is 0
Hi Jacob
I want to use OrthoSNAP identify single-copy orthologous genes nested within larger multi-copy gene families. But I have a question: Single-copy orthologous genes identified is 0. following is my orthogroup tree. I think this orthgroup should have multiple subsets of single-copy orthologous genes. I'm not sure if I had a problem with my data preparation. The fasta file is in FASTA format
test.faa.zip
(species0|01G00667-RA:0.5279117043,(((species4|01G04915-RA:0.3408345014,((((((((((((((((((species0|02G00068-RA:0.2833922109,species0|03G01048-RA:0.2202978172)87:0.0438419260,species0|06G05308-RA:0.0342599303)89:0.0234458576,((species0|08G00253-RA:0.0300651474,species0|01G06890-RA:0.1020280754)93:0.0297929956,species0|05G03661-RA:0.0292380589)81:0.0003640542)68:0.0148926744,(species0|10G05798-RA:0.0100306324,species0|10G05809-RA:0.0099046340)100:0.0425698431)65:0.0239536424,species0|03G03143-RA:0.0037587110)64:0.0901618950,((species0|06G04570-RA:0.2009566336,(species0|09G00242-RA:0.0276064627,species0|09G00252-RA:0.0000010000)98:0.0444388887)95:0.0272777858,species0|02G00069-RA:0.0184186553)92:0.0247560096)65:0.0684129707,(species0|03G02980-RA:0.0311118160,species0|01G01632-RA:0.0081907706)48:0.0068875751)30:0.0000022508,species0|06G06430-RA:0.0171662576)47:0.0824734516,species0|08G01583-RA:0.0582321366)74:0.0698105372,(((((((((species0|06G04577-RA:0.1029693336,species0|08G01179-RA:0.0058083040)65:0.0037224542,species0|06G06159-RA:0.0062634697)18:0.0000010000,species0|02G02446-RA:0.0031141986)9:0.0000010000,species0|02G04676-RA:0.0000010000)13:0.0000010000,species0|05G06142-RA:0.0031219956)34:0.0000023735,species0|05G00002-RA:0.0000010000)66:0.0176742730,species0|05G02195-RA:0.0000010000)67:0.0270298851,species0|07G00780-RA:0.0111600524)93:0.0810354911,species0|05G06683-RA:0.0094745401)56:0.0033010953)75:0.0139675175,species3|04G02997-RA:0.0479693279)54:0.0070020470,((((species0|UnG00061-RA:0.1019755829,((species0|05G04281-RA:0.0100712598,species0|UnG01151-RA:0.0000010000)63:0.0000010000,species0|08G00680-RA:0.0033416155)97:0.0075728491)76:0.0047777879,species0|07G05284-RA:0.0205874047)34:0.0000022133,((((species0|11G02225-RA:0.0000010000,species0|11G02226-RA:0.0034223709)100:0.0206299805,species0|UnG00947-RA:0.0101420749)53:0.0000024204,species0|05G06682-RA:0.0050443888)54:0.0000010000,species0|03G03240-RA:0.0151959726)73:0.0025157288)62:0.0224812810,species0|03G01047-RA:0.0274408894)31:0.0000444198)61:0.0209780864,(species3|09G02646-RA:0.0263452359,species0|06G04032-RA:0.0190750798)96:0.0027947135)44:0.0000024207,species0|07G03150-RA:0.0339043024)65:0.0126331495,(((((species3|03G03823-RA:0.1104765017,species3|07G00184-RA:0.0282350932)68:0.0021200555,species3|UnG04073-RA:0.0334773575)63:0.0049922998,(((species0|05G00255-RA:0.0717264427,species3|01G02394-RA:0.2180804066)93:0.0232292895,species3|07G03562-RA:0.0328082244)12:0.0000022505,species3|02G01645-RA:0.0456063875)9:0.0000024289)15:0.0000028490,species3|08G02640-RA:0.1005701721)72:0.0080842576,species0|09G00240-RA:0.0608569853)76:0.0151201690)61:0.0085221777,species2|09G00394-RA:0.1317027250)66:0.0083495796,species0|03G03342-RA:0.0000010000)67:0.0112977953,((((((species2|11G04129-RA:0.1107995970,species2|11G04120-RA:0.0000010000)74:0.0000024273,species2|11G04119-RA:0.0000010000)81:0.0000027880,(species2|11G04138-RA:0.0000022138,species2|11G04146-RA:0.0164280934)100:0.0500509920)100:0.0318182024,species2|07G02618-RA:0.0105404848)96:0.0180265075,(((species2|11G03854-RA:0.0286863546,(((species2|06G01852-RA:0.0708707538,species2|11G00501-RA:0.0364722483)87:0.0123542412,species2|11G02708-RA:0.0300289749)56:0.0000027713,species2|05G01547-RA:0.0300265612)70:0.0000024696)90:0.0257159554,species2|04G04476-RA:0.0156910819)52:0.0000026998,species2|05G02497-RA:0.0301598075)72:0.0000024693)98:0.0200837784,((((species2|10G02227-RA:0.3108728301,species2|05G03233-RA:0.0000026898)97:0.0464486734,species2|10G04334-RA:0.0489527240)91:0.0170403849,(species2|02G02902-RA:0.0602162416,(species2|06G02624-RA:0.0000010000,species2|06G02625-RA:0.0000010000)96:0.0018256185)99:0.0294009743)17:0.0000023342,((species2|08G02827-RA:0.0711202157,species2|01G04476-RA:0.0898182083)97:0.0348380375,species2|01G05041-RA:0.0233416284)49:0.0000027600)23:0.0000028939)100:0.1365387571)66:0.0777546471)61:0.1920394558,((species1|04G01317-RA:0.1160132618,species1|02G00465-RA:0.1715145619)99:0.1552966767,species1|11G02518-RA:0.1294596501)94:0.1533573617)61:0.0288696434,species1|05G02443-RA:0.3565432133)78:0.0900322084,(species3|07G04113-RA:0.4880719806,(((((species0|09G00325-RA:0.0689512854,species0|09G00326-RA:0.0829119289)100:0.3091446375,species0|05G03566-RA:0.3080354823)82:0.0870903929,(species0|08G05223-RA:0.0376953437,species0|05G03094-RA:0.0244950842)62:0.0000025151)82:0.0439946415,species0|01G06839-RA:0.0354282611)47:0.0322824766,species3|05G04758-RA:0.0950203399)36:0.0000024398)98:0.2712607015);
test.faa.zip
The fasta file is in FASTA format
Hi Xiao Xu,
After looking at the phylogeny, I don't see any instances where OrthoSNAP should be finding a subgroup of single-copy orthologous genes (or SNAP-OGs). To test the behavior of OrthoSNAP, I ran your files but changed the occupancy threshold to have at least two sequences. As expected, SNAP-OGs were identified. Based on the tree topology, are you expecting specific SNAP-OGs?
Assuming no SNAP-OGs being the correct output, there are some clear cases wherein this could happen. For example, some multi-copy gene families have unclear evolutionary histories that make highly confident SNAP-OG detection difficult. Unclear evolutionary histories may be driven by complex patterns of duplication and loss or a lack of information, making tree inference challenging.
Hope this helps!
All the best,
Jacob
Hi! Jacob
Thank you very much for your answer, which helped me a lot. I still have a few questions I would like to ask: my ultimate goal is to test whether two different multicopy orthgroups exist coevolution using PhyKIT. Therefore, I would like to use orthosnap to identify single-copy orthologous genes first. If one has many subsets of multicopy orthgroups, which subset should I choose for coevolution identification? In addition, like in my data, there are five species, but genes of only two species are identified in the subset. Can this incomplete SNAP-OGs represent its multi-copy orthgroups
Hi Xiaoxu,
Thank you so much for adopting the software we have developed. Perhaps others may also be of use to you. A complete list of software our team has engineered can be viewed here: https://jlsteenwyk.com/software.html.
If you are coevolutionary calculations with five species, which is relatively few, I would recommend using the occupancy argument when running OrthoSNAP; specifically, set this parameter to five. In doing so, OrthoSNAP will only report SNAP-OGs with full taxon occupancy (N=5).
To directly address your question, if multiple SNAP-OGs are identified in a larger multi-copy gene family, use them all instead of just one. This can substantially increase the gene space examined.
Please feel free to comment with any other questions you may have. I will close the ticket since the main issue has been addressed.
Thank you again for choosing to use OrthoSNAP and PhyKIT.
best,
Jacob
|
2025-04-01T04:10:32.443033
| 2017-01-18T12:23:19
|
201560036
|
{
"authors": [
"JMPerez",
"hughrawlinson"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14588",
"repo": "JMPerez/passport-spotify",
"url": "https://github.com/JMPerez/passport-spotify/issues/12"
}
|
gharchive/issue
|
Greenkeeper integration
Hey, it looks like you don't have greenkeeper set up on this repo. Worth doing to make sure you support latest versions of express 😄
@hughrawlinson I hadn't tried Greenkeeper before. I have just integrated it, so I'll keep an eye on PRs with updates for the dependencies.
Fixed through https://github.com/JMPerez/passport-spotify/pull/13
Thanks @JMPerez! :)
|
2025-04-01T04:10:32.462769
| 2023-05-04T15:24:56
|
1696240881
|
{
"authors": [
"daredevil2033",
"manticore-projects"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14589",
"repo": "JSQLParser/JSqlParser",
"url": "https://github.com/JSQLParser/JSqlParser/issues/1784"
}
|
gharchive/issue
|
Possible WebSite improvements
Mode switch has three buttons (half-full circle, sun, crescent) but only two change the mode (crescent doesn't)
sphinx-tabs-tab color doesn't account for dark mode (used in Compile from Source Code and Build Dependencies)
Syntax Railroad diagrams don't account for dark mode
Some of the Syntax Railroad diagrams are obstructed (such as JsonFunction, JsonAggregateFunction, CreateTable and AlterExpression)
Some of the Syntax EBNF diagrams are obstructed (such as Ll and Lu)
Thanks for reporting.
Mode switch has three buttons (half-full circle, sun, crescent) but only two change the mode (crescent doesn't)
This comes from the used Sphinx Theme and I honestly don't know how to fix that.
sphinx-tabs-tab color doesn't account for dark mode (used in Compile from Source Code and Build Dependencies)
This comes from the used Sphinx Theme and I honestly don't know how to fix that.
Syntax Railroad diagrams don't account for dark mode
Yes, this one is a problem. We will need to look into the svg.css if it can be amended for dark mode.
Some of the Syntax Railroad diagrams are obstructed (such as JsonFunction, JsonAggregateFunction, CreateTable and AlterExpression)
Yes, but the only solution was to refactor the Productions in the Grammar.
Especially the ALTER TABLE... and CREATE TABLE... need some serious work.
Some of the Syntax EBNF diagrams are obstructed (such as Ll and Lu)
I will look into this one. Also the Spelling is broken and needs to be fixed.
|
2025-04-01T04:10:32.465245
| 2018-07-11T09:12:22
|
340156439
|
{
"authors": [
"wumpz",
"zhouruifeng"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14590",
"repo": "JSQLParser/JSqlParser",
"url": "https://github.com/JSQLParser/JSqlParser/issues/639"
}
|
gharchive/issue
|
Parse the conditional expression id in () into id
Actual Behavior
I want to parse sql and do some custom operations. However, when parsing id in (), there will be cases that do not meet expectations, and id in() is converted to id.
Expected Behavior
Expect id in () to resolve to id in()
Steps to Reproduce the Problem
1.maven dependency:1.2
2.System.out.println(CCJSqlParserUtil.parseCondExpression("id in ()"));
3.output:id
Specifications
Version:1.2
Platform:Mac,Linux
Subsystem:Java
The expression is parsed only for complete expressions. () Is incomplete therefore only I'd is parsed.
|
2025-04-01T04:10:32.466408
| 2019-10-05T12:31:55
|
502957571
|
{
"authors": [
"azamzamy"
],
"license": "cc0-1.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14591",
"repo": "JStumpp/awesome-android",
"url": "https://github.com/JStumpp/awesome-android/pull/287"
}
|
gharchive/pull-request
|
Update Instabug listing
Update Instabug Sponsorship Listing
• Upload new photo
• Modify content
• Modify URL
@JStumpp Hi Jürgen, This PR includes updates to the Instabug section to match the new branding. Thanks!
|
2025-04-01T04:10:32.480259
| 2024-07-11T19:18:56
|
2403919110
|
{
"authors": [
"JaJuMa",
"rhoerr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14593",
"repo": "JaJuMa-GmbH/awesome-hyva",
"url": "https://github.com/JaJuMa-GmbH/awesome-hyva/issues/3"
}
|
gharchive/issue
|
Add $attributes option typehint
Thanks for the great module!
Could you add the $attributes option to the parameter list of all the icon methods?
The icon classes have for instance:
* @method string adHtml(string $classnames = '', ?int $width = null, ?int $height = null)
But the underlying formatHtml() in Hyva has a fourth $attributes parameter available. This is important for instance when providing a title tag, to change the mouseover text from the icon name.
public function renderHtml(
string $icon,
string $classNames = '',
?int $width = 24,
?int $height = 24,
array $attributes = []
): string {
Example usage:
<?= $awesomeiconssolid->listHtml("w-5 h-5", 25, 25, ['title' => $escaper->escapeHtmlAttr(__('Add to list'))]) ?>
Currently this works (because it's a magic method), but PhpStorm doesn't think the parameter exists.
Thanks a lot @rhoerr for pointing that out!
Fixed with release 2.1.6
|
2025-04-01T04:10:32.606139
| 2019-02-11T23:57:47
|
409035769
|
{
"authors": [
"JacksonMaxfield"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14594",
"repo": "JacksonMaxfield/pywslegislature",
"url": "https://github.com/JacksonMaxfield/pywslegislature/issues/7"
}
|
gharchive/issue
|
Restructure services.py to be class based with a class for each web service
Example being:
class CommitteeService(obj):
def __init__(self):
# attached the header attribute
# a description would be neat
# construct attributes for every function that uses this service
This will make it easier for people to not have to remember or have the WSL Web Services API open in another window.
Completed in [gh-2].
|
2025-04-01T04:10:32.610477
| 2016-01-12T01:17:21
|
126075222
|
{
"authors": [
"JacobEvelyn",
"andypearson"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14595",
"repo": "JacobEvelyn/friends",
"url": "https://github.com/JacobEvelyn/friends/issues/69"
}
|
gharchive/issue
|
All command-line arguments should result in prompts if not supplied
It would be nice to allow this, similar to how activity descriptions can be entered via prompt, but for all commands. So I could just say friends graph.
I'd be interested in digging in to this a little bit as I've been using friends quite a lot recently.
One thing I'm wondering, in the case of things like graph it could default to generating a graph of all your activities, which is then "filtered" by passing in the first argument.
Ah, that's an interesting use case of graph that I hadn't thought of, and it kind of throws a wrench in this feature since friends graph was one of the main motivations (though friends add friend is still another where it would be useful).
I just made #83 (which it sounds like you may be interested in working on?) to address your graph suggestion—it's yours if you want it!
I'm also interested in hearing your thoughts on this issue (assuming #83 happens) and #84, which is somewhat related.
I'm up for making #83 my first PR :)
I haven't had a review of all commands, but I was wondering if it's possible to adopt a common pattern where all "read" commands have a sensible "unfiltered" result, which you can then refine using additional arguments and all "write" commands provide an interactive way of completing the full command.
I think both of those ideas make sense; the latter is really what I was getting at with this issue originally.
Closing this for now, as it seems this Issue has transformed a bit, and other Issues that have spawned from it can be completed independently.
|
2025-04-01T04:10:32.635653
| 2020-10-22T08:54:36
|
727187350
|
{
"authors": [
"AndreasDeCrinis",
"MBtech",
"Slashoper",
"jahstreet",
"maciekdude"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14596",
"repo": "JahstreetOrg/spark-on-kubernetes-helm",
"url": "https://github.com/JahstreetOrg/spark-on-kubernetes-helm/issues/46"
}
|
gharchive/issue
|
File file:/tmp/history-server does not exist when configuring history server
Hi,
we are strugling around with configuring the history server in livy using these env vars:
LIVY_SPARK_EVENT1LOG_ENABLED: {value: "true"}
LIVY_SPARK_EVENT1LOG_DIR: {value: "file:///tmp/history-server"}
LIVY_LIVY_UI_HISTORY0SERVER0URL: {value: "https://historyserver.mycluster.lan"}
after we trigger a job, we see this error message in the driver container:
Exception in thread "main" java.io.FileNotFoundException: File file:/tmp/history-server does not exist
does anybody have a clue what we are doing wrong?
BR
Andreas
Hi, the error you observe signals that there are not such file/directory in your Spark HS container. To make it work by default you should create the NFS PVC with name eg. nfs-pvc in the Spark HS namespace and configure the spark-cluster Helm chart with the following values:
historyserver:
pvc:
# to use a file system path for Spark events dir, set 'enablePVC' to true and mention the
# name of an already created persistent volume claim in existingClaimName.
# The volume will be mounted on /data in the pod
enablePVC: true
existingClaimName: nfs-pvc
eventsDir: "/"
Then you do not need to override LIVY_SPARK_EVENT1LOG_DIR to make it work.
Alternatively you need to provide the configs to Spark HS so that it could access for instance HDFS compatible file system. For additional details please refer https://github.com/helm/charts/tree/master/stable/spark-history-server docs.
Just create it in the underlying image ;)
RUN chmod +x /opt/entrypoint.sh && \
chmod g+w $SPARK_HOME/work-dir && \
mkdir -p /tmp/spark-events
@maciekdude , then how will Spark containers write history logs to it? You need to have the shared directory to make it work.
Executors do not write logs there even on shared FS like hdfs/s3. It's only driver, so if you have problem with the spawning jobs you can always disable evenlogin, get some share storage like s3 or just create this folder ;)
If this is the way you are ok to go with then I have no arguments ;)
I am running into a similar issue. I have created a PV and PVC and added the following corresponding settings for the historyserver charge:
pvc:
enablePVC: true
existingClaimName: events-dir
eventsDir: "/"
nfs:
enableExampleNFS: false
pvcName: events-dir
pvName: events-dir-pv
Which configurations do I need to change for the livy chart?
I have changed these two:
env:
# Configure History Server log directory to write Spark logs to
LIVY_SPARK_EVENT1LOG_ENABLED: {value: "true"}
LIVY_SPARK_EVENT1LOG_DIR: {value: "file:///data"}
Which other configurations do I need to change for the livy chart? Persistence configurations?
Figured it out. I needed to configure the Kubernetes Volumes configuration properties for Spark driver and executors as mentioned here.
Figured it out. I needed to configure the Kubernetes Volumes configuration properties (in the request for the livy batch job) for Spark driver and executors as mentioned here.
how do you reslove it, can i see your submit spark job config?
|
2025-04-01T04:10:32.646844
| 2015-04-30T14:13:15
|
72174594
|
{
"authors": [
"JakeWharton",
"rmielnik"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14597",
"repo": "JakeWharton/butterknife",
"url": "https://github.com/JakeWharton/butterknife/issues/259"
}
|
gharchive/issue
|
Multiple ButterKnife.inject(target, View) causes references are reset
Hi,
I created ListView with some header and footer. My initialization is like:
{
...
final View header = inflater.inflate(R.layout.header, this, false);
addHeaderView(header, null, false);
ButterKnife.inject(this, header);
(*)
final View footer = inflater.inflate(R.layout.footer, this, false);
addFooterView(footer, null, false);
ButterKnife.inject(this, footer);
(**)
...
}
I used on click annotations and everything worked fine:
@Optional
@OnClick(R.id.some_view_from_header)
public void someWorkingFunction() {
}
@Optional
@OnClick(R.id.some_view_from_footer)
public void someOtherWorkingFunction() {
}
Then I needed to get references to those views. Tried with:
@Optional
@InjectView(R.id.some_view_from_header)
protected View viewFromHeader;
@Optional
@InjectView(R.id.some_view_from_footer)
protected View viewFromFooter;
What happens is at point (*) I have a properly set reference to viewFromHeader. At point (**) viewFromHeader is null and viewFromFooter is properly set.
Is this expected behavior?
Calling inject will set a reference to every view, present or not. The fact that the click listeners continue to work is an accident–don't rely on that. The right way to do multiple injections is to use the view holder pattern for secondary injections.
static class HeaderViews {
@InjectView(R.id.header_title) TextView title
}
final HeaderViews header = new HeaderViews();
// ...
ButterKnife.inject(header, headerView);
header.title.setText("HEADER!!!!!!");
|
2025-04-01T04:10:32.648309
| 2016-10-06T11:44:17
|
181392891
|
{
"authors": [
"JakeWharton",
"TeodorKolev"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14598",
"repo": "JakeWharton/butterknife",
"url": "https://github.com/JakeWharton/butterknife/issues/767"
}
|
gharchive/issue
|
Support identifiers
Is there support for next code:
Drawable message = (Drawable) getResources().getDrawable(getResources().getIdentifier(myText, "drawable", MainActivity.PACKAGE_NAME));
No, there's no point. The only reason to use that syntax is to do dynamic loading at runtime and Butter Knife resolves everything at compile time. Stick to what you already have.
|
2025-04-01T04:10:32.649918
| 2023-08-27T15:09:15
|
1868529485
|
{
"authors": [
"JakobLichterfeld"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14599",
"repo": "JakobLichterfeld/ansible-role-latex_texlive",
"url": "https://github.com/JakobLichterfeld/ansible-role-latex_texlive/issues/1"
}
|
gharchive/issue
|
The conditional check 'remove_texlive_doc' failed
When executing, the Task "Install LaTeX - remove texlive doc packages" fails with "conditional check 'remove_texlive_doc' failed".
The variable is not called correct:
https://github.com/JakobLichterfeld/ansible-role-latex_texlive/blob/e461793201074c82df1d1cfca996c377e7c93e2b/tasks/main.yml#L116
Must be latex_texlive.remove_texlive_doc
|
2025-04-01T04:10:32.657053
| 2016-11-24T16:53:12
|
191563562
|
{
"authors": [
"namniak",
"vespakoen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14600",
"repo": "Jam3/voice-activity-detection",
"url": "https://github.com/Jam3/voice-activity-detection/pull/1"
}
|
gharchive/pull-request
|
Support more runtimes (by getting rid of the fat arrow function)
Amazing stuff on your repository, hats off to you guys!
I am playing with this package and am trying to run it in CEF (I noticed you guys are familiar with AE as well), that doesn't support fat arrow functions at the moment though.
This fixes it =)
Thanks for your great work!
Thanks. totally.
I tried to get rid of ES6 stuff before publishing but totally missed this bit. thanks!
|
2025-04-01T04:10:32.664618
| 2018-11-09T20:29:08
|
379310906
|
{
"authors": [
"JamesHenry"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14601",
"repo": "JamesHenry/ng-custom-element",
"url": "https://github.com/JamesHenry/ng-custom-element/pull/4"
}
|
gharchive/pull-request
|
fix(directive): normalize names in prop and event bindings
It seems when jQuery is used, it will usually wrap the original CustomEvent which breaks the
interface of the expected detail payload
:tada: This PR is included in version 2.0.2 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T04:10:32.666329
| 2023-09-15T14:58:35
|
1898598560
|
{
"authors": [
"Marc-Eid"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14602",
"repo": "JamesPartsafas/ConcordiaCurriculumManager",
"url": "https://github.com/JamesPartsafas/ConcordiaCurriculumManager/pull/2"
}
|
gharchive/pull-request
|
Front-End setup completed
Created a React app with Vite and Typescript
configured Eslint and Prettier for static code analysis and formatting
added React-Router
closes #1
for the CSS we should use both. in the index we have the global styling, like font and colors for example. and then eventually we'll need to change the CSS for specific pages which then we'll make a css page for the specific file
I also added the folder structure in wiki. I can mention that in the meeting
|
2025-04-01T04:10:32.668393
| 2016-03-08T16:14:52
|
139319023
|
{
"authors": [
"JamesRandall",
"MrKWatkins"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14603",
"repo": "JamesRandall/Simple-Paging-Grid",
"url": "https://github.com/JamesRandall/Simple-Paging-Grid/pull/32"
}
|
gharchive/pull-request
|
Fix to pass row data into row templates.
I was trying to use Handlebars in row templates but the row data wasn't being passed in; this change fixes it.
Thanks for the contrib - much appreciated!
Would the following slight change to your proposed fix work for you:
var tr = $(that._settings.rowTemplates[rowTemplateIndex]({index: rowTemplateIndex, data: rowData }));
I think this will achieve what you require but also allow anybody who was using the row index to also still achieve what they were doing. It's a breaking change but I can bump the version accordingly.
It's a while since I've been in this code so feeling a bit rusty!
Just tried it, that works fine for me.
|
2025-04-01T04:10:32.672076
| 2015-10-29T13:44:56
|
114052269
|
{
"authors": [
"Applejaxc",
"JustDevin",
"SirWaddles"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14604",
"repo": "JamesWilko/Payday-2-BLT",
"url": "https://github.com/JamesWilko/Payday-2-BLT/issues/36"
}
|
gharchive/issue
|
Game crahes wheni start it up.
Crashes after all the into videos are done,when the Press any key to continue tries to load up the game crashes.
Can you post your crash.txt, BLT log and binary sha1?
I'm having the same issue. Reinstalled last night, game kept throwing access violation errors at me until I added the d39dxwhateverwhatever .dll fix. Now with BLT trying to run, I crash as soon the intro videos are done (can't see my character/main menu).
"Sat Oct 31 10:29:40 2015
Application has crashed: C++ exception
mods/Poser/PoserLua/MenuScene.lua:162: attempt to call method 'refresh' (a nil value)
SCRIPT STACK
Callstack:
payday2_win32_release (???) ???
payday2_win32_release (???) ???
payday2_win32_release (???) ???
payday2_win32_release (???) ???
ntdll (???) RtlCreateHeap
ntdll (???) RtlCreateHeap
ntdll (???) NtQueryVirtualMemory
KERNELBASE (???) VirtualQueryEx
KERNELBASE (???) SetUnhandledExceptionFilter
KERNELBASE (???) SetUnhandledExceptionFilter
payday2_win32_release (???) ???
payday2_win32_release (???) ??? "
Yes I've installed (and reinstalled!) the proper C++
Looks like an issue with Poser?
Poser was the issue! How did you know? Is "Kernel" a handle Poser's creator uses or are you just familiar with what it tries to do?
Oh wait, no.
"Application has crashed: C++ exception
mods/Poser/PoserLua/MenuScene.lua:162: attempt to call method 'refresh' (a nil value)"
Sorry. I've been so annoyed by access violation errors since the start of crimefest I guess I gave up actually reading what my crashlogs said. Thank you, SirWaddles.
|
2025-04-01T04:10:32.729671
| 2023-08-23T22:30:47
|
1864101234
|
{
"authors": [
"JanLunge",
"kdb424",
"rayererer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14605",
"repo": "JanLunge/pog",
"url": "https://github.com/JanLunge/pog/issues/4"
}
|
gharchive/issue
|
Missing license
I noticed this project doesn't have a license to use the code. Is this code going to get a license at some point to be used, or just here for personal storage? Love the efforts, and hope to see this get some sort of open license in the future!
A LICENSE has since been added in commit 880a582. So I suppose this issue can be closed.
Yes MIT license has been added.
Seems I had Notifications for Issues and PRs off my bad fixed this now.
|
2025-04-01T04:10:32.761509
| 2023-10-15T16:33:50
|
1943959851
|
{
"authors": [
"Malkovitz",
"freeload101"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14606",
"repo": "JarodMica/audiosplitter_whisper",
"url": "https://github.com/JarodMica/audiosplitter_whisper/issues/15"
}
|
gharchive/issue
|
Could not download 'pyannote/segmentation-3.0' model.
Hello, I recently tried to spin up this program again, but it seems like there's some problem with speaker diarization.
It worked okay month ago, and I have not changed anything in the config file.
I tried to fix it generating new token and making sure I was gated in all three programs like in YouTube tutorial, but it's still failing.
`Could not download 'pyannote/segmentation-3.0' model.
It might be because the model is private or gated so make
sure to authenticate. Visit https://hf.co/settings/tokens to
create your access token and retry with:
Model.from_pretrained('pyannote/segmentation-3.0',
... use_auth_token=YOUR_AUTH_TOKEN)
If this still does not work, it might be because the model is gated:
visit https://hf.co/pyannote/segmentation-3.0 to accept the user conditions.
Traceback (most recent call last):
File "c:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\Scripts\whisperx-script.py", line 33, in
sys.exit(load_entry_point('whisperx==3.1.1', 'console_scripts', 'whisperx')())
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\whisperx\transcribe.py", line 211, in cli
diarize_model = DiarizationPipeline(use_auth_token=hf_token, device=device)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\whisperx\diarize.py", line 19, in init
self.model = Pipeline.from_pretrained(model_name, use_auth_token=use_auth_token).to(device)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\pyannote\audio\core\pipeline.py", line 136, in from_pretrained
pipeline = Klass(**params)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\pyannote\audio\pipelines\speaker_diarization.py", line 128, in init
model: Model = get_model(segmentation, use_auth_token=use_auth_token)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\pyannote\audio\pipelines\utils\getter.py", line 89, in get_model
model.eval()
AttributeError: 'NoneType' object has no attribute 'eval'
Traceback (most recent call last):
File "c:\Users\arkad\Desktop\ai\audiosplitter_whisper\split_audio.py", line 183, in
main()
File "c:\Users\arkad\Desktop\ai\audiosplitter_whisper\split_audio.py", line 180, in main
process_audio_files(input_folder, settings)
File "c:\Users\arkad\Desktop\ai\audiosplitter_whisper\split_audio.py", line 148, in process_audio_files
diarize_audio_with_srt(audio_file_path, srt_file, speaker_segments_dir)
File "c:\Users\arkad\Desktop\ai\audiosplitter_whisper\split_audio.py", line 77, in diarize_audio_with_srt
subs = pysrt.open(srt_file)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\pysrt\srtfile.py", line 151, in open
source_file, encoding = cls._open_unicode_file(path, claimed_encoding=encoding)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\pysrt\srtfile.py", line 292, in _open_unicode_file
encoding = claimed_encoding or cls._detect_encoding(path)
File "C:\Users\arkad\Desktop\ai\audiosplitter_whisper\venv\lib\site-packages\pysrt\srtfile.py", line 279, in _detect_encoding
file_descriptor = open(path, 'rb')
FileNotFoundError: [Errno 2] No such file or directory: 'C:\Users\arkad\Desktop\ai\Pudzian\output\plik.srt'`
Nevermind, I just found the issue.
In YouTube video we are supposed to accept conditions for:
https://huggingface.co/pyannote/segmentation
but the script is requiring
https://huggingface.co/pyannote/segmentation-3.0
which looks like is never version of segmentation.
I agreed to 3.0 conditions, but it's getting stuck on >>Performing diarization...
C:\whisperX\1\audiosplitter_whisper\venv\Lib\site-packages\whisperx\diarize.py
https://hf.co/pyannote/speaker-diarization-3.1 for me ...
|
2025-04-01T04:10:32.769021
| 2024-08-12T20:49:09
|
2461868080
|
{
"authors": [
"matavach",
"zDyanTB"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14607",
"repo": "Jas-SinghFSU/HyprPanel",
"url": "https://github.com/Jas-SinghFSU/HyprPanel/issues/116"
}
|
gharchive/issue
|
[Bug] Can't open options where notifications are
Describe the bug
Can't expand options where notifcations are showed
e.g: notications shows on the right, widgets that are on the right can't expand, or even hover
To Reproduce
Send an notifcation and try open any option on the same side
Expected behavior
overlay ?
Desktop (please complete the following information):
Distribution: Nix
Window Manager/Desktop Environment: Hyprland
https://github.com/user-attachments/assets/11c9244d-3abd-4b3a-ad22-55bc39b11ad1
I was able to reproduce this easily.
Will take a look.
|
2025-04-01T04:10:32.773653
| 2024-10-29T08:53:50
|
2620563976
|
{
"authors": [
"Daholli",
"Jas-SinghFSU",
"albohlabs"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14608",
"repo": "Jas-SinghFSU/HyprPanel",
"url": "https://github.com/Jas-SinghFSU/HyprPanel/issues/395"
}
|
gharchive/issue
|
[Feature] option to render empty workspaces in the bar
I would like to have an option to enable empty workspaces to be displayed in the bar in front of a non-empty workspace.
Let's say I have an existing Firefox client on workspace 3.
Currently, only this one workspace is rendered in the bar.
What I want is to have the 2 empty workspaces rendered before it.
The reason for this is that I use shortcuts like ctrl + 1 to jump to workspaces. If the empty workspaces are rendered before a non-empty one, it would be easier to find the corresponding shortcut.
Would this be satisfied by upping the number of Total Workspaces in Configuration > Bar?
That number forces N number of workspaces to always show. So you can set it to 2 in this case, so it always shows the first 2.
Setting the workspaces to 9 would be a workaround for me and I could live with that.
But just to illustrate the behaviour a bit more, if I have a hpyrland client open on workspace 5, then workspace 4 is not rendered. So again, I could not see the empty workspace in between.
But as I said, I could live with setting the total number of workspaces to 9 as a workaround. The question for me was, does it make sense for you to implement this feature into the codebase?
Ahhh I think I understand what you mean. You want to show all workspaces prior to your higher workspace at all times?
Yessss, that is exactly the behaviour i want. What do you think about adding it?
Yeah, that should be doable and probably useful. I'll look into adding it.
I can also have a look at it and create a MR?
I can also have a look at it and create a MR?
Most welcomed to ❤️
This can also be done by "assigning" workspaces to a monitor such as
workspace=1, monitor:DP-2
workspace=2, monitor:DP-2
workspace=3, monitor:DP-2
workspace=4, monitor:DP-2
in your hyprland config, this will make it so they always show up regardless if they have a window in them or not
|
2025-04-01T04:10:32.775391
| 2015-04-10T11:26:26
|
67573228
|
{
"authors": [
"leleuj",
"mmoayyed"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14609",
"repo": "Jasig/cas",
"url": "https://github.com/Jasig/cas/pull/921"
}
|
gharchive/pull-request
|
Disable authn flow if no service is found, optionally
Handles https://github.com/Jasig/cas/issues/549
+1
|
2025-04-01T04:10:32.785074
| 2023-06-13T13:07:38
|
1754852854
|
{
"authors": [
"JasonDsouza212",
"Prajwal0225"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14610",
"repo": "JasonDsouza212/free-hit",
"url": "https://github.com/JasonDsouza212/free-hit/pull/610"
}
|
gharchive/pull-request
|
Adding website for developer for practice and learning purpose
Related Issue
Closes #610
Description
Added websites
devChallenges.io
Frontend Practice
JavaScript30
Codewars
Checklist
[x] My code adheres to the established style guidelines of the project.
[ ] I have included comments in areas that may be difficult to understand.
[x] My changes have not introduced any new warnings.
[x] I have conducted a self-review of my code.
The descriptions are too long.Make them shorter, with a maximum of 15 words each.
|
2025-04-01T04:10:32.788166
| 2018-04-26T10:57:10
|
317977413
|
{
"authors": [
"Favna",
"JasonEtco"
],
"license": "isc",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14611",
"repo": "JasonEtco/todo",
"url": "https://github.com/JasonEtco/todo/issues/139"
}
|
gharchive/issue
|
Todobot seemingly stopped working entirely
I've added some todo comments to my code on multiple commits now but the bot never picks them up and creates issues anymore. Examples of these commits are here and here, neither time did the bot pick them up as TODO's.
The only thing that comes to mind is that in both of these cases the commits were queued along with other commits when I pushed them but that really shouldn't matter.
So what gives?
p.s. I know they are the same todos, it was an attempt to get the bot to trigger. I temporarily removed it here
Looks like it hit some rate limiting issues. I don't have the bandwidth to solve that, but you're welcome to clone the repo and host your own instance of the app.
|
2025-04-01T04:10:32.824158
| 2016-07-20T08:35:44
|
166527809
|
{
"authors": [
"JayBizzle",
"alberto-bottarini"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14612",
"repo": "JayBizzle/Crawler-Detect",
"url": "https://github.com/JayBizzle/Crawler-Detect/issues/111"
}
|
gharchive/issue
|
some other false positives
hi,
maybe I found three other false positives. Here's the list:
Mozilla/5.0 (Linux; Android 4.4.2; ForwardRuby Build/KOT49H) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/51.0.2704.81 Mobile Safari/537.36
Mozilla/5.0 (Linux; Android 5.1; Cosmos Build/LMY47I) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/<IP_ADDRESS> Mobile Safari/537.36
Mozilla/5.0 (Linux; U; Android 4.2.1; en-ph; MyPhone Agua Vortex Build/JOP40D) AppleWebKit/534.30 (KHTML, like Gecko) Version/4.0 Mobile Safari/534.30
What do you think?
ty
I think you are correct!
Will push a fix later today when I get chance.
Thanks for reporting ❤️
|
2025-04-01T04:10:32.838409
| 2018-07-18T17:41:19
|
342426288
|
{
"authors": [
"Galileo007",
"fredroo"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14613",
"repo": "Jays2Kings/DS4Windows",
"url": "https://github.com/Jays2Kings/DS4Windows/issues/450"
}
|
gharchive/issue
|
Question about PS4 Wireless USB
This dongle works with the last version of DS4Windows?
Someone know the range of PS4 Wireless USB
That's the official dongle, it should work perfectly.
|
2025-04-01T04:10:32.855086
| 2021-09-19T08:32:00
|
1000248638
|
{
"authors": [
"BR00l",
"kvndrsslr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14614",
"repo": "Jean-Tinland/simple-bar",
"url": "https://github.com/Jean-Tinland/simple-bar/issues/223"
}
|
gharchive/issue
|
Mail plugin
Hello there, you have done an asthonishing work here, I really like the whole setup and the idea of this bar!
I have one request though, Is there a possibility for a mail indicator plugin on the bar?
You can just implement this as a user plugin.
The script for simple global inbox count:
osascript -l JavaScript -e '((x) => {x.checkForNewMail();return x.inbox().unreadCount()})(Application("Mail"))'
My personal script (unread count over all mailboxes except my gmail accounts archive ('All Mail') to avoid duplicates):
osascript -l JavaScript -e '((x) => {x.checkForNewMail();return x.accounts().flatMap(a => a.mailboxes()).filter(a => a.name() !== "All Mail").map(a => a.unreadCount()).reduce((a,b) => a+b, 0)})(Application('Mail'))'
|
2025-04-01T04:10:32.857272
| 2022-11-05T15:51:33
|
1437079150
|
{
"authors": [
"Jean-Tinland",
"kristiansordal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14615",
"repo": "Jean-Tinland/simple-bar",
"url": "https://github.com/Jean-Tinland/simple-bar/issues/325"
}
|
gharchive/issue
|
Pin data to the right
I have a problem where I cannot figure out how to pin the data portion of my bar to the right, as seen in the picture, on the right side, the data is pinned on the left hand side, however I would like to pin it to the left.
How do I go about achieving this?
My configuration is as follows:
:root {
--bar-height: 27px;
--bar-radius: 14px;
--item-radius: 8px;
--bar-inner-margin: 6px;
--item-inner-margin: 3px 12px;
--item-outer-margin: 0 0 0 6px;
--display: 50px;
}
.simple-bar--no-bar-background .spaces {
min-width: 642px;
alignSelf: 'flex-end';
}
.simple-bar--no-bar-background .simple-bar__data {
min-width: 647px;
alignSelf: 'flex-end';
}
I think the css you are looking for is:
.simple-bar--no-bar-background .spaces {
min-width: 642px;
justify-content: flex-start;
}
.simple-bar--no-bar-background .simple-bar__data {
min-width: 647px;
justify-content: flex-end;
}
Thank you so much, that was excactly what I was looking for!
|
2025-04-01T04:10:32.858199
| 2021-11-28T21:25:44
|
1065440582
|
{
"authors": [
"JeanPhilippeKernel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14616",
"repo": "JeanPhilippeKernel/RendererEngine",
"url": "https://github.com/JeanPhilippeKernel/RendererEngine/issues/153"
}
|
gharchive/issue
|
Fix SceneViewport size on first initialization
When we launch the editor for the first time, the SceneViewportUI component doesn't update its size with the size of the region available in the DockspaceUI component
fixed by the default layout recently merged see #185
|
2025-04-01T04:10:32.887623
| 2019-04-10T12:55:15
|
431487673
|
{
"authors": [
"Jeffail",
"asiragusa"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14617",
"repo": "Jeffail/benthos",
"url": "https://github.com/Jeffail/benthos/issues/182"
}
|
gharchive/issue
|
Failed to delete consumed SQS messages
I get the following error on the SQS input:
Failed to delete consumed SQS messages: AWS.SimpleQueueService.BatchEntryIdsNotDistinct: Id 34376646-2fb8-4131-ba1c-d2db2f0060cc repeated.
status code: 400, request id: 0c4b21eb-174f-5ae3-a3b3-74ccbad399b1
when one of the processors or outputs returns an error response (response.NewError). Moreover the Acknowledge function does not take into account such errors and voids the at-least-once guarantee. Instead of deleting the message, the visibility timeout should be set to 0, as the doc states: https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-visibility-timeout.html#terminating-message-visibility-timeout
Hey @asiragusa, thanks for raising this.
Looks like I wrapped the input with a preserver https://github.com/Jeffail/benthos/blob/master/lib/input/sqs.go#L51, which should preserve ALO but isn't idiomatic use of SQS. When an error occurs the input retries the same SQS message until success, but if that loop lasts longer than the visibility timeout we'll end up with the same message consumed multiple times, which potentially breaks the batch deletes once the error loop ends (duplicate IDs in the batch request).
I'll try and get a fix out today.
Fixed: https://github.com/Jeffail/benthos/commit/2512b42bdbd7c8d91842e4f259093be66b3e608e
Released: https://github.com/Jeffail/benthos/releases/tag/v1.10.10
@asiragusa, I've reproduced the issue and tested the fix with a relatively low volume queue, but we don't use this consumer in production so I don't have a heavy load to throw at it. If you still get problems please reopen this ticket and I'll do some heavier testing.
Cool, thanks!
|
2025-04-01T04:10:32.907700
| 2017-08-24T18:10:10
|
252685158
|
{
"authors": [
"JeffreyBenjaminBrown"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14618",
"repo": "JeffreyBenjaminBrown/digraphs-with-text",
"url": "https://github.com/JeffreyBenjaminBrown/digraphs-with-text/issues/1"
}
|
gharchive/issue
|
When parsing, the # operator must abut something
Currently, to parse an empty joint, I can use #(). I would like to be able to use # alone (surrounded by non-word characters).
Solved. I was using Megaparsec.Char.symbol to parse the # operator, and symbol is a lexeme, so it had already consumed any trailing space before looking for the label that might follow the #. (The # operator by itself is not a lexeme; only the combination of having looked for that and then looked for the trailing label should be treated as a lexeme.)
So I switched to Megaparsec.Char.string.
|
2025-04-01T04:10:32.910948
| 2017-06-19T02:38:16
|
236766402
|
{
"authors": [
"joshmanders"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14619",
"repo": "JeffreyWay/laravel-mix",
"url": "https://github.com/JeffreyWay/laravel-mix/issues/888"
}
|
gharchive/issue
|
Unexpected behavior of build between production and development
Laravel Mix Version: #.#.# (npm list --depth=0) 1.0.0-beta (but latest stable gives same error)
Node Version (node -v): 8.1.2
Yarn Version (yarn -v): 0.24.6
OS: macOS Sierra
Description:
Using jQuery (Semantic-UI) with Vue in development mode everything works fine, but anytime I run production build something changes and causes Transition: Element is no longer attached to DOM. Unable to animate. Use silent setting to surpress this warning in production.
I've tested this by including the production build of vue on the page along with jquery and semantic-ui with only the following code being compiled
new Vue({
el: '#app',
mounted() {
$('.dropdown').dropdown();
}
});
hmmm weird, I ditched my whole webpack.mix.js file and started over and now it seems to work.
|
2025-04-01T04:10:32.947074
| 2019-12-27T10:03:01
|
542820888
|
{
"authors": [
"JeremySkinner",
"benmccallum"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14620",
"repo": "JeremySkinner/FluentValidation",
"url": "https://github.com/JeremySkinner/FluentValidation/issues/1285"
}
|
gharchive/issue
|
Async PropertyValidator Documentation
System Details
FluentValidation version: 8.4.0 (via FluentValidation.AspNetCore 8.4.0)
Web Framework version: ASP.NET Core 2.2
Issue Description
We're currently investigating performance/deadlock issues in our codebase and a use of .GetAwaiter().GetResult() inside an implementation of a PropertyValidator's IsValid method in an extremely hot path is concerning me.
I've looked at #999, but wanted to confirm that the suggestion to use AsyncValidatorBase will prevent our deadlocking issues. Specifically, that this line is OK: https://github.com/JeremySkinner/FluentValidation/blob/c24d6929f013b4ba003f0c6819c87672b0e984cf/src/FluentValidation/Validators/AsyncValidatorBase.cs#L49
If it's all g, would be happy to help write up a docs PR on the use of this base class; just let me know.
Cheers,
Ben
Use of GetAwaiter().GetResult() forces an asynchronous method to run synchronously.
This is only ever invoked if you have async validators, but you’re calling them synchronously (with a call to Validate at the top level), which you should never do. If your validator includes async property validators then you should always call ValidateAsync, and never Validate.
If you are using asp.net’s automatic validation (as part of model binding) then be aware that the validation pipeline is entirely synchronous and you should never use any async validators at all, as they will only ever be invoked synchronously (this is a limitation of asp.net, not FluentValidation)
Hope that clears things up.
Ahh, so the correct usage when needing async is to use IsValidAsync; and essentially I could/should just throw in the IsValid sync method to highlight any bad executions (e.g. via Validate not ValidateAsync)?
I did read about the limitation of asp.net. Is that applicable to ASP.NET Core as well as the old-school stack (MVC5)? If so, not sure how we're going to work around that :/ Any suggestions? Write an async middleware?
For 2nd question, just found #1175. Hmm, that's a fun one. Will do some googling for solutions, but shout out if you have any. Appreciate the help. Happy holidays :)
essentially I could/should just throw in the IsValid sync method to highlight any bad executions
Yes, that's correct. I'm actually considering making this the default behavior for 9.0.
Is that applicable to ASP.NET Core as well
Yes. The model binding and validation processes in all versions of ASP.NET Core is synchronous.
If so, not sure how we're going to work around that
You'd have to stop using automatic validation and instead inject the validators into your controllers. If your controller actions are then async, you can then directly invoke the validator's ValidateAsync inside the controller action.
Thanks Jeremy. Kind of puzzling the asp.net core team didn't go async in the validator pipeline. I'll do some reading to find out why and if it's on the horizon.
I think since we are using MediatR, I might introduce a middleware into the pipeline before our handlers that does the validation and uses ValidateAsync. I'm sure I won't be the first to do this so will take a
Edit: Thanks for the further detail on data annotations. Makes sense :)
|
2025-04-01T04:10:32.951948
| 2017-06-28T21:56:19
|
239309605
|
{
"authors": [
"JeremySkinner",
"gregatarcurve"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14621",
"repo": "JeremySkinner/FluentValidation",
"url": "https://github.com/JeremySkinner/FluentValidation/issues/500"
}
|
gharchive/issue
|
ASP.NET Core FluentValidation disables default data annotation validation
Enabling FluentValidation in an ASP.NET Core WebApi application appears to disable/break the default validation provided the framework - such that data annotations placed on models (e.g. StringLength, Required, etc.) have no effect on ModelState.IsValid. If the AddFluentValidation method is removed then the default framework validation works as expected and ModelState.IsValid works as expected.
Is this intentional? This did not occur in older versions of FluentValidation (targeting ASP.NET MVC 4). Of course I could replace these attributes with fluent validations, but I'd prefer to have the flexibility of placing the standard attributes directly on the model properties.
services.AddMvc()
.AddFluentValidation(options => options.RegisterValidatorsFromAssemblyContaining<Startup>())
class MyModel
{
[StringLength(10)] // This only works if you remove the call to AddFluentValidation
string MyProperty {get; set;}
}
It should allow both. Which version of FluentValidation are you running?
I'm using FluentValidation.AspNetCore 7.1.0, though I also noticed the behavior with 7.0.3.
So 7.0 completely disabled data annotations, but this was changed after the release (either 7.0.1 or 7.0.2), so DataAnnotations should be working in 7.1.
Note that you can't mix FV and DataAnnorations in the same model object (this is a a limitation of aspnet core, not of FV), but you should be able to use them both in the same project on different classes...if thats not the behaviour you're seeing then I'll try and look into this further later this week.
That's exactly what the problem/behavior is, thank you. I didn't realize this was a limitation of ASP.NET Core. Thank you for your quick response.
|
2025-04-01T04:10:33.061330
| 2022-04-06T03:50:02
|
1193971062
|
{
"authors": [
"hsz",
"ice1000"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14622",
"repo": "JetBrains/gradle-intellij-plugin",
"url": "https://github.com/JetBrains/gradle-intellij-plugin/issues/939"
}
|
gharchive/issue
|
java-compiler-ant-tasks for 221?
> Task :instrumentCode
[gradle-intellij-plugin :intellij-arend:instrumentCode] Cannot resolve java-compiler-ant-tasks in version: 221.5080.93
[gradle-intellij-plugin :intellij-arend:instrumentCode] Resolved closest lower java-compiler-ant-tasks version: 213.7172.33
Is this an issue?
No. As there is no java-compiler-ant-tasks available in the version that matches precisely the SDK you use, the closest lower one is picked to keep everything working.
Tyvm
|
2025-04-01T04:10:33.068335
| 2024-01-30T13:17:27
|
2107845834
|
{
"authors": [
"Drjacky",
"hsz"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14623",
"repo": "JetBrains/intellij-platform-gradle-plugin",
"url": "https://github.com/JetBrains/intellij-platform-gradle-plugin/issues/1533"
}
|
gharchive/issue
|
PublishPlugin Task Failed
What happened?
Github Action -> publishPlugin task fails:
https://github.com/Drjacky/Avocado/actions/runs/7709858760/job/21011916928
latest try: https://github.com/Drjacky/Avocado/actions/runs/7709858760/job/21017535162
All the secrets(including PUBLISH_TOKEN) are provided!
Related thread: https://github.com/JetBrains/gradle-intellij-plugin/issues/1513
Relevant log output or stack trace
> Task :jarSearchableOptions FROM-CACHE
> Task :buildPlugin
FAILURE: Build failed with an exception.
> Task :signPlugin SKIPPED
* What went wrong:
> Task :publishPlugin FAILED
Execution failed for task ':publishPlugin'.
18 actionable tasks: 14 executed, 4 from cache
> token property must be specified for plugin publishing
Configuration cache entry stored.
Steps to reproduce
Tru to publish the plugin through Github Actions
Gradle IntelliJ Plugin version
1.16.1
Gradle version
8.5
Operating System
None
Link to build, i.e. failing GitHub Action job
https://github.com/Drjacky/Avocado/actions/runs/7709858760/job/21017535162
Duplicate of #1540
|
2025-04-01T04:10:33.070478
| 2022-03-24T11:28:08
|
1179381494
|
{
"authors": [
"llehn",
"vepanimas"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14624",
"repo": "JetBrains/js-graphql-intellij-plugin",
"url": "https://github.com/JetBrains/js-graphql-intellij-plugin/issues/602"
}
|
gharchive/issue
|
Enable environment variables
At the moment there seems to be no way to make the plugin use the value of environment variable, other than explicitly giving it via the "Enter missing variable..." dialog.
This basically turns the advertised functionality of using something like ${env:TOKEN} into something like
${SHOW_PROMPT} (or SHOW_PROMPT_1, SHOW_PROMPT_2, to support different values)
I saw the discussions about supporting env files, which IMHO totally have the wrong point - the concern which env files in which order should be applies is one which is not what this plugin is about.
I hope somebody corrects me and i'm just too stupid for not being able to find a way to use the enviroment variable (the thing that System.getenv() gets the value for)
They are set of course.
@llehn I fixed that, now env variables should be available.
|
2025-04-01T04:10:33.091855
| 2022-03-24T09:00:18
|
1179206074
|
{
"authors": [
"En3Tho",
"auduchinok"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14625",
"repo": "JetBrains/resharper-fsharp",
"url": "https://github.com/JetBrains/resharper-fsharp/pull/358"
}
|
gharchive/pull-request
|
Changed TreeBuilder to align let/use ILocalBindings and let!/use!/and! ILocalBindings
Now in every case they like:
ILetOrUseExpr
ILocalBinding
FSharpTokenType
*TokenElement
Before this change in case of let!/use!/and!:
ILetOrUseExpr
FSharpTokenType
*TokenElement
ILocalBinding
Added tests to make sure they all look alike.
For now I've changed ProcessAndLocalBinding to ProcessAndBangLocalBinding to reduce confusion.
So far I haven't tried to merge binding processing together. Waiting for your decision.
For now I've changed ProcessAndLocalBinding to ProcessAndBangLocalBinding to reduce confusion.
Yup, it helps, thanks! I've added a minor comment, otherwise it seems good to go in!
|
2025-04-01T04:10:33.094874
| 2022-03-11T23:10:03
|
1166965882
|
{
"authors": [
"elbebegato",
"philipto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14626",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/13746"
}
|
gharchive/pull-request
|
Higher University of San Andrés
file.zip
Web Page
https://www.umsa.bo/
Mail Domain
umsa.edu.bo
Universidad Mayor de San Andrés
La Paz - Bolivia
Plaza del Obelisco, Av. Camacho, La Paz
https://goo.gl/maps/XaFwHubV8hH9SRc2A
@elbebegato I am sorry, I have to close the request without merging, because you made a mistake in a request and requested to merge too many files, most of them do not relate to the organization which you would like to add to this repository. Please ensure that you are familiar with this repository rules located at the bottom of this page and submit a new request with the correct information. Thank you for your understanding!
|
2025-04-01T04:10:33.098427
| 2023-05-09T21:26:49
|
1702778485
|
{
"authors": [
"philipto",
"soda4fries"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14627",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/17467"
}
|
gharchive/pull-request
|
Universiti Malaya email: siswa.um.edu.my add
The University website is : https://www.um.edu.my/
The Link to the Computer Science Faculty of UM : https://fsktm.um.edu.my/#
This is official UM guide that show information that this email address is recognized by UM: https://umsitsguide.um.edu.my/applicant/pdf_app/guides_app/UM System Access.pdf
@soda4fries I am sorry, the domain siswa.um.edu.my is removed from our trusted domains list because the school allows to register an email address there even if you are not a student of the school. If you are the student of this school, please apply with official document (OFFICIAL DOCUMENT tab at https://www.jetbrains.com/shop/eform/students).
|
2025-04-01T04:10:33.104360
| 2024-03-25T01:03:42
|
2204667784
|
{
"authors": [
"Ensophea-Toch",
"philipto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14628",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/20436"
}
|
gharchive/pull-request
|
tgi
TUX Global Institute
@Ensophea-Toch lib/domains/kh/edu/tgi: this request has an incorrect file path format, please ensure that you are familiar with this repository rules located at the bottom of this page and change your request accordingly. If you don't know how to change request, close it and open a new one. If you are not able to make the pull request according the instructions, please send us an email to<EMAIL_ADDRESS>Please mention this request URL in the message.
@Ensophea-Toch I have to close this request without merging, because you requested to add a non-existent domain to this repository: Host tgi.edu.kh not found: 3(NXDOMAIN) Please ensure that you are familiar with this repository rules located at the bottom of this page and submit a new request with the correct information. Thank you for your understanding!
Dear JetBrains Team,
tgi domain is tuxglobalinstitute.edu.kh thanks!
On Thu, 28 Mar 2024 at 18:33 Philip Torchinsky @.***>
wrote:
Closed #20436 https://github.com/JetBrains/swot/pull/20436.
—
Reply to this email directly, view it on GitHub
https://github.com/JetBrains/swot/pull/20436#event-12280058095, or
unsubscribe
https://github.com/notifications/unsubscribe-auth/BG5W65NKN4YPRG7NSAFM7ZLY2PWX5AVCNFSM6AAAAABFGEYG4SVHI2DSMVQWIX3LMV45UABCJFZXG5LFIV3GK3TUJZXXI2LGNFRWC5DJN5XDWMJSGI4DAMBVHAYDSNI
.
You are receiving this because you were mentioned.Message ID:
@.***>
|
2025-04-01T04:10:33.105371
| 2024-04-22T14:30:27
|
2256695351
|
{
"authors": [
"hnikp",
"philipto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14629",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/20711"
}
|
gharchive/pull-request
|
Create ESMT.txt
Adding the email domain of European School of Management and Technology (ESMT), Berlin
@hnikp Thank you for your patience while waiting for a review and processing your request. Pull request merged.
|
2025-04-01T04:10:33.109539
| 2024-05-17T16:53:33
|
2303205091
|
{
"authors": [
"rmelchorv"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14630",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/20965"
}
|
gharchive/pull-request
|
Adding UNISTMO domain
"bianni" is the subdomain for Ixtepec campus of UNISTMO domain.
University official website URL: https://www.unistmo.edu.mx/
URL of a page on the official website where a long-term (>1 year) IT related course is offered: https://www.unistmo.edu.mx/lic_informatica.html
URL of a page or some other proof (.pdf or a screenshot) showing that the university recognizes the domain which you are submitting as an official email domai: https://correo.unistmo.edu.mx/mail
"bianni" is the subdomain for Ixtepec campus of UNISTMO domain.
University official website URL: https://www.unistmo.edu.mx/
URL of a page on the official website where a long-term (>1 year) IT related course is offered: https://www.unistmo.edu.mx/lic_informatica.html
URL of a page or some other proof (.pdf or a screenshot) showing that the university recognizes the domain which you are submitting as an official email domai: https://correo.unistmo.edu.mx/mail
|
2025-04-01T04:10:33.112689
| 2024-07-21T06:05:13
|
2421243098
|
{
"authors": [
"Alejandrom199",
"philipto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14631",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/21954"
}
|
gharchive/pull-request
|
Add domain for Universidad de Guayaquil
Domain Name:
ug.edu.ec
University:
Universidad de Guayaquil
Official Website URL:
https://www.ug.edu.ec/
@Alejandrom199 I am sorry, the domain ug.edu.ec is removed from our trusted domains list because it was abused in the past. If you are the student of this school, please enroll to GitHub Student Developer Program. Then apply for JetBrains educational subscription with your GitHub account (GitHub tab at https://www.jetbrains.com/shop/eform/students).
|
2025-04-01T04:10:33.114059
| 2017-02-06T10:31:12
|
205542794
|
{
"authors": [
"DoggyCode",
"philipto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14632",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/2733"
}
|
gharchive/pull-request
|
Added Østfold Fylkeskommune Schools domain. ostfoldfk
Was told by you to create a new one.
@DoggyCode Pull request merged. Thank you!
@philipto Thank you very much, do you know when the list will be updated at jetbrains so I can collect my student lincense? :) Right now it's telling me: "Your email address does not belong to any university that we know. Contribute your university to our list.". I assume this is because it updates on intervals.
|
2025-04-01T04:10:33.114900
| 2019-09-20T05:45:08
|
496157261
|
{
"authors": [
"longhn",
"philipto"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14633",
"repo": "JetBrains/swot",
"url": "https://github.com/JetBrains/swot/pull/6844"
}
|
gharchive/pull-request
|
Hong Bang International University
Please review and add Hong Bang International University for the Learning license
@longhn Pull request merged. Thank you!
|
2025-04-01T04:10:33.134412
| 2019-08-19T18:10:33
|
482453598
|
{
"authors": [
"bemasc"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14634",
"repo": "Jigsaw-Code/Intra",
"url": "https://github.com/Jigsaw-Code/Intra/pull/213"
}
|
gharchive/pull-request
|
Revert "Split APK by ABI"
The Play Store doesn't permit multiple APKs for different architectures to share a version code, so this splitting configuration is basically not usable in practice. We'll just have to stick with the unified APK until we can get bundles working.
@alalamav FYI
|
2025-04-01T04:10:33.143174
| 2017-09-19T16:53:49
|
258891799
|
{
"authors": [
"JimBobSquarePants",
"perosb"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14635",
"repo": "JimBobSquarePants/ImageProcessor",
"url": "https://github.com/JimBobSquarePants/ImageProcessor/pull/630"
}
|
gharchive/pull-request
|
Update libwebp.dll to 0.6
Updates libwebp to latest 0.6.
https://github.com/webmproject/libwebp/releases/tag/v0.6.0
Aw nice! I always find it a chore doing this. 👍
@JimBobSquarePants it seems my versions require vc2017 runtime, I'll see if I can recompile with vc2013 or vc2015..not sure what people use today
|
2025-04-01T04:10:33.150999
| 2021-05-29T16:15:26
|
906506606
|
{
"authors": [
"Germwalker",
"jimmyfm"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14636",
"repo": "JimmyLaurent/torrent-search-api",
"url": "https://github.com/JimmyLaurent/torrent-search-api/issues/133"
}
|
gharchive/issue
|
Dead repo ?
Heya,
I need some updates present in the pull request queue.
The first one is more or less one year old, with any comments.
Is this repo dead ?
Best regards
I fear so :(
|
2025-04-01T04:10:33.156447
| 2024-02-28T16:50:17
|
2159398402
|
{
"authors": [
"JinShil",
"pm64"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14637",
"repo": "JinShil/BlazorWebView",
"url": "https://github.com/JinShil/BlazorWebView/issues/9"
}
|
gharchive/issue
|
Child windows
Is there any guidance or pattern we should follow with respect to opening and managing child windows? Is this supported?
It depends on what kind of child window you wish to use. Do you want to open a dialog window in the browser with the <dialog /> tag? Do you want to create a message box in the browser with alert() or confirm()? Or do you want to display a native window on the platform, like a file selection dialog window?
For the windows in the browser, you can use the <dialog /> tag, alert() function, confirm() function or even create your own HTML/CSS/Javascript window.
Native windows, like a file selection dialog window, should also work fine because this is a Blazor hybrid project that utilizes the local .NET framework on the device that the program is running on, as opposed to Blazer Server that utilizes the .NET framework on a web server, or Blazor WASM that utilizes the .NET framework loaded into the browser respectively.
So, on Windows, you should be able to add a reference to any native Windows .NET library and utilize it. You can also P/Invoke into any native unmanaged libraries should you need to. However, any native libraries will not be portable across platforms, so you'll probably have to add an implementation for each platform.
I think Maui might make that more portable, but I don't have any experience with it.
Hey @JinShil, thanks for your reply. The child windows I had in mind are in a 3rd category -- proper child windows that can move beyond the bounds of the main application window, which themselves are displaying Blazor content.
This works, but it may not be what you were thinking:
@using Microsoft.Extensions.DependencyInjection
<button @onclick="OpenWindow">Test</button>
@code {
void OpenWindow()
{
var window = Gtk.Window.New();
window.SetDefaultSize(800, 600);
// Add the BlazorWebView
var serviceProvider = new ServiceCollection()
.AddBlazorWebViewOptions(new BlazorWebViewOptions()
{
RootComponent = typeof(WebKitGtk.Test.App),
HostPath = "wwwroot/index.html"
})
.BuildServiceProvider();
var webView = new BlazorWebView(serviceProvider);
window.SetChild(webView);
window.Show();
}
}
If you wanted to open a window with javascript's window.open(), I believe you will need to tap into the WebView.OnCreate event and create the window there.
Closing as I'm assuming there is nothing more to discuss. Reopen if there is anything further.
|
2025-04-01T04:10:33.157865
| 2023-10-06T05:41:05
|
1929471516
|
{
"authors": [
"Jingkang50",
"codeprogredire"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14638",
"repo": "Jingkang50/OpenPSG",
"url": "https://github.com/Jingkang50/OpenPSG/issues/109"
}
|
gharchive/issue
|
PSG dataset download
How to download the PSG Dataset? In the link they have provided for downloading, they have JSON files and COCO.zip file. Can anyone explain how to make sense of them.
Please check https://github.com/Jingkang50/OpenPSG/blob/main/tools/Visualize_Dataset.ipynb
|
2025-04-01T04:10:33.160022
| 2021-10-13T18:13:43
|
1025554291
|
{
"authors": [
"JoernT"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14639",
"repo": "Jinntec/Fore",
"url": "https://github.com/Jinntec/Fore/issues/71"
}
|
gharchive/issue
|
handle id resolution within repeats
ids must be resolved in their respective repeat context to make functions like index() work that refer to another element.
there's one use case that even exists beyond functions - not sure how we can address this.
Consider a repeated control that uses a label - this would usually look like this:
<fx-repeat ref="...">
<fx-control ref="...">
<label for="foo">a label</label>
<input id="foo">
....
This would currently fail as the browser will just always take the first control it finds. This is especially a problem with checkboxes where you want the label to be clickable to check/uncheck the box.
probably an overwritten version of fn:id() is needed that uses the new resolveId function to accomplish the use case like this:
<fx-repeat ref="...">
<fx-control ref="...">
<label for="id('foo')">a label</label>
<input id="foo">
|
2025-04-01T04:10:33.174584
| 2021-06-08T16:11:31
|
915217899
|
{
"authors": [
"JoDehli",
"lawrenceakka"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14640",
"repo": "JoDehli/pyloxone-api",
"url": "https://github.com/JoDehli/pyloxone-api/pull/4"
}
|
gharchive/pull-request
|
Message rework
For discussion:
This is a fairly extensive rework of message parsing. Messages are now parsed by specific classes. I have also subclassed the websocket to enable more detailed logging, and to centralise message handling a bit more.
I found a couple of race-conditions in the original code. If a keep alive message or token refresh request was sent to the mini server at the same time as an incoming message was being processed then the incoming message could become garbled. I have fixed these, I hope, but there is still scope for this to happen. You can see this if you set the refresh and keep alive times to, eg, 10 seconds, and cause the mini server to send a number of messages quickly (eg changing a slider up and down). This is not likely to happen often, but it is a possibility.
@lawrenceakka thanks for all the work. I will try it as soon as possible. At the moment I have not so much time but I will try it and merge it. Thanks.
|
2025-04-01T04:10:33.177195
| 2016-11-03T22:23:52
|
187206876
|
{
"authors": [
"Davidrums",
"JoakimSoderberg"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14641",
"repo": "JoakimSoderberg/imageclipper",
"url": "https://github.com/JoakimSoderberg/imageclipper/pull/14"
}
|
gharchive/pull-request
|
add c++ version flags for compiler
I was having trouble compiling on Ubuntu 16.04 with all the packages reinstalled. This flag solved this error
imageclipper/src/imageclipper.cpp:83:3: error: ‘nullptr’ was not declared in this scope nullptr, ^
Thanks
|
2025-04-01T04:10:33.197937
| 2023-09-10T04:35:34
|
1888971185
|
{
"authors": [
"JoelKanyi",
"Tonnie-Dev"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14642",
"repo": "JoelKanyi/KomposeCountryCodePicker",
"url": "https://github.com/JoelKanyi/KomposeCountryCodePicker/issues/7"
}
|
gharchive/issue
|
LocalSoftwareKeyboardController Crash
I am using Material 3 Version 1.2.0-alpha07 and Compose UI Version 1.5.1 and I am getting this log crash while trying out KomposeCountryCodePicker.
java.lang.NoClassDefFoundError: Failed resolution of: Landroidx/compose/ui/platform/LocalSoftwareKeyboardController;
at com.joelkanyi.jcomposecountrycodepicker.component.KomposeCountryCodePickerKt.KomposeCountryCodePicker(KomposeCountryCodePicker.kt:74)
Caused by: java.lang.ClassNotFoundException: Didn't find class "androidx.compose.ui.platform.LocalSoftwareKeyboardController" on path: DexPathList[[dex file
There is a similar issue that I saw on this unrelated bug report.
When I downgraded MD to 1.1.1 the code did not crash but crucial material three colors could not be resolved e.g. MaterialTheme.colorScheme.surfaceContainerLowest.
I also tried using Integration with TextField but the code still crashed.
Any leads on how to avert this crash will greatly be appreciated.
Let me look into the issues, then will get back to you
I have just created a sample, and all seems well: sample usage
Kotlin Version: 1.9.10
Compose BOM Version: 2023.09.00 which translates to Compose Version 1.5.1
Compose Compiler Version: 1.5.3
KomposeCountryCodePicker Version: 1.0.3
Which stable version of material 3 is this offered? - MaterialTheme.colorScheme.surfaceContainerLowest
Which stable version of material 3 is this offered? - MaterialTheme.colorScheme.surfaceContainerLowest
Only in
Thanks for the prompt review and the tips
After upgrading from Kotlin 1.9.0 -> 1.9.10, Compose Compiler 1.5.0 -> 1.5.3 and Downgrading from Material3:1.2.0-alpha07 -> Material3:1.2.0-alpha03, KomposeCountryCodePicker is now working without crashing
MaterialTheme.colorScheme.surfaceContainerLowest is only available from on Material3:1.2.0-alpha03 onwards. Let me play around with the your Library and I will report in case of any difficulties.
Thanks Man 👍🏽
|
2025-04-01T04:10:33.222062
| 2019-07-08T18:41:57
|
465404422
|
{
"authors": [
"JohnSully",
"mikefaraponov"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14643",
"repo": "JohnSully/KeyDB",
"url": "https://github.com/JohnSully/KeyDB/issues/52"
}
|
gharchive/issue
|
Does KeyDB support WAIT command in active-active to wait for data propagation?
x.y.z.d:6379> wait 1 1
(error) ERR WAIT cannot be used with replica instances. Please also note that since Redis 4.0 if a replica is configured to be writable (which is not the default) writes to replicas are just local and are not propagated.
This feature is vital for me
probably it relates to https://github.com/JohnSully/KeyDB/issues/45
This looks like just an oversight - the feature should work. Let me test it and unblock it.
Upon testing this fix I found a scenario where the replica is undercounting the backlog. Outside of this command it may result in extra full syncs however with this command it will wait forever.
Once that issue is resolved I can enable this for Active Replication.
This is now fixed in the unstable branch.
Change:
a5784ef09e71a9a45780a8f3dbab875b1f1fe1a5
I’ve done quite a bit of testing as well as adding new tests to track this.
This will be our v5 release slated for Monday.
This command is enabled in v0.9.6. Docker is updated with this new version.
|
2025-04-01T04:10:33.223833
| 2016-05-12T15:33:57
|
154513364
|
{
"authors": [
"JohnSundell",
"alexaubry"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14644",
"repo": "JohnSundell/Unbox",
"url": "https://github.com/JohnSundell/Unbox/pull/60"
}
|
gharchive/pull-request
|
Changes sources layout to support Swift Package Manager
This commit enables the use of Unbox in a Swift Package Manager-based environment and possibly on Linux (needs testing) with Swift 2.2 (DEVELOPMENT-SNAPSHOT-2016-03-01-a). Extending support to SPM required the 1.5.2 release to be created (Package.swift versioning).
So awesome, thank you very much for this @alexaubry! 🚀
Perfect, thanks again! 👍
|
2025-04-01T04:10:33.252892
| 2024-10-20T19:44:09
|
2600821787
|
{
"authors": [
"Jollywatt",
"NiceWaffel",
"aaron-jack-manning",
"brynhum"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14645",
"repo": "Jollywatt/typst-fletcher",
"url": "https://github.com/Jollywatt/typst-fletcher/issues/55"
}
|
gharchive/issue
|
Deprecation warnings since Typst 0.12 (use of measure and style)
Fletcher 0.5.1 now generates warnings with Typst 0.12, sometimes directly from code in fletcher, and sometimes from code in cetz:0.2.2. Presumably the latter error can be fixed by making it depend on cetz:0.3.1 instead, and the former warnings are related to code in fletcher itself.
warning: `style` is deprecated
┌─ @preview/fletcher:0.5.1/src/diagram.typ:482:5
│
482 │ box(style(styles => {
│ ╭───────^
483 │ │ let options = options
484 │ │
485 │ │ options.em-size = measure(h(1em)).width
· │
543 │ │ render(grid, nodes, edges, options)
544 │ │ }))
│ ╰────^
│
= hint: use a `context` expression instead
warning: calling `measure` with a styles argument is deprecated
┌─ @preview/fletcher:0.5.1/src/node.typ:357:24
│
357 │ let (width, height) = measure(box(
│ ╭───────────────────────────^
358 │ │ node.label,
359 │ │ width: inner-size.at(0),
360 │ │ height: inner-size.at(1),
361 │ │ ), styles)
│ ╰──────────────^
│
= hint: try removing the styles argument
The deprecation warnings in fletcher seem to be fixed with Pull Request #51, not the ones from the outdated CeTZ though.
@Jollywatt when do you plan to release the next package version, so these fixes are generally available?
@NiceWaffel I've looked at upgrading to cetz 0.3.0 but it introduces some breaking changes which require some time. I'm a starting PhD student at present so have very little time unfortunately!
@Jollywatt Also note that Cetz changed licenses in the update latest update, https://github.com/cetz-package/cetz/pull/598. I will have a bit of available time soon and am happy to contribute to this, but I would need clarification on if fletcher is happy to go GPL too then.
The breakages I observed with cetz 0.3.0 were luckily fixed in cetz 0.3.1, so it wasn't too bad to update. Fletcher 0.5.2 is now available and requires typst >=0.12.0.
|
2025-04-01T04:10:33.256232
| 2023-02-06T17:50:02
|
1573017284
|
{
"authors": [
"JonasKruckenberg",
"Ludea"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14646",
"repo": "JonasKruckenberg/tauri-build",
"url": "https://github.com/JonasKruckenberg/tauri-build/issues/215"
}
|
gharchive/issue
|
Uploading file on Windows
Tauri build correctly artifact
Running candle for "main.wxs"
Running light to produce D:\a\Sparus\Sparus\src-tauri\target\release\bundle/msi/sparus_0.1.0_x64_en-US.msi
Finished 1 bundle at:
D:\a\Sparus\Sparus\src-tauri\target\release\bundle/msi/sparus_0.1.0_x64_en-US.msi
When using softprops action-gh-release I get
Pattern 'D:\a\Sparus\Sparus\src-tauri\target\release\bundle\msi\sparus_0.1.0_x64_en-US.msi' does not match any files.
🤔 D:\a\Sparus\Sparus\src-tauri\target\release\bundle\msi\sparus_0.1.0_x64_en-US.msi not include valid file.
I don't know if it come from tauri-build or softprop action
Can't say much without more input (a repro or project to look at) but I just ran the test suite again and it passes without issues (including uploading the msi) so it's most likely an issue with your setup
Log : https://github.com/Ludea/Sparus/actions/runs/4103480208
Workflow : https://github.com/Ludea/Sparus/blob/main/.github/workflows/release.yml
https://github.com/softprops/action-gh-release/issues/280#issuecomment-1322187892 can be relevant
Closing, this issue is not from tauri-build but from softprops action
|
2025-04-01T04:10:33.267090
| 2015-02-10T20:07:26
|
57229084
|
{
"authors": [
"JonyEpsilon",
"puredanger"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14648",
"repo": "JonyEpsilon/gorilla-repl",
"url": "https://github.com/JonyEpsilon/gorilla-repl/issues/192"
}
|
gharchive/issue
|
Folded segments
In building some training material with Gorilla, it would be useful if segments (or possibly groups of segments, but I'm not that greedy) had a "folded" state and if you could toggle it in the UI. That way you can put:
a text segment with instructions
a code segment with some shell of a solution where they can try things
a hidden segment (text or code) with the answer
Section folding would be a great feature, and there was an issue where we discussed this a bit #163 . It eventually closed as "won't fix" because I couldn't figure out how to do it without unreasonable effort! I think to do it properly the worksheet would have to move to a tree-shaped data model, rather than the current list.
The simpler idea of having segments fold would be much easier to implement. To make this persistent I think we'd need to have some segment-level metadata. This is something I've been consistently resisting, as it seems like an increase in complexity, and also the thin end of a wedge - once the metadata exists there's some temptation to add feature-after-feature. But maybe it's time to accept that metadata is a necessary complication.
I guess an alternative implementation would be to have something like IPython's "cell magics", which are magic strings that can appear at the start of a segment. But I never thought this was particularly elegant, so I'm not sure it's a good solution.
|
2025-04-01T04:10:33.280423
| 2019-03-01T06:31:37
|
415961425
|
{
"authors": [
"JordanAdams",
"viko16"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14649",
"repo": "JordanAdams/devdocs-cli",
"url": "https://github.com/JordanAdams/devdocs-cli/pull/2"
}
|
gharchive/pull-request
|
fix postinstall hook
fixed #1
Cheers for this 🎉
@viko16 Would you like to be added to the authors section of the package.json? :)
No, I’m good.
Thank you @JordanAdams . 😄
🚀 Published in v0.1.3
|
2025-04-01T04:10:33.288614
| 2020-10-01T22:08:01
|
713215967
|
{
"authors": [
"DrAg0n141",
"JordanKnott",
"almereyda",
"jpdsceu",
"wuast94"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14650",
"repo": "JordanKnott/taskcafe",
"url": "https://github.com/JordanKnott/taskcafe/issues/78"
}
|
gharchive/issue
|
Taskcafe PostgreSQL Port
Hello,
i need to change the postgresql port to another, what ENV is the right one? I have tested TASKCAFE_DATABASE_HOST_PORT but it is not working.
Thanks for your help.
DrAg0n141
In docker-compose.yml, add in bold.
postgres:
image: postgres:12.3-alpine
restart: always
networks:
- taskcafe-test
**ports:
-'YOURPORT:5432'**
environment:
POSTGRES_USER: taskcafe
POSTGRES_PASSWORD: taskcafe_test
POSTGRES_DB: taskcafe
volumes:
- taskcafe-postgres:/var/lib/postgresql/data
This method will only work if TASKCAFE_DATABASE_HOST_PORT is a valid environment.
I propose to consider this issue as resolved, since there was no reply of the OP.
Currently it is not possible to set the port of the Postgres DB Taskcafe tries to connect to however it will be possible once I finish the feature I am currently working on.
This is now possible to config as of the 0.3.0 release
i searched in wiki and files but didnt find what the variable for docker is to set a diffrent DB port. so wich env variable i must set?
|
2025-04-01T04:10:33.291332
| 2020-06-14T18:42:12
|
638407082
|
{
"authors": [
"JordanMartinez",
"and-pete",
"milesfrain"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14651",
"repo": "JordanMartinez/purescript-cookbook",
"url": "https://github.com/JordanMartinez/purescript-cookbook/issues/34"
}
|
gharchive/issue
|
Autogenerate Readme's list of recipes
See this comment for more context.
In the meantime, the RoutingLog link in the first column of the ReadMe table is broken (as it refers to RoutingLog instead of the current RoutingHashLog). Also the corresponding text in the Description column refers to a misspelled rouing library. :)
D'oh! Thanks for making us aware of that!
Quick fix for above included in https://github.com/JordanMartinez/purescript-cookbook/commit/27162d0ca3235c927415069c2cf234b0f801ae93
Thanks @milesfrain for fixing that!
|
2025-04-01T04:10:33.296936
| 2024-10-19T19:29:35
|
2599520358
|
{
"authors": [
"JordanSantiagoYT",
"PatoFlamejanteTV",
"PleaseEnterUsernameHere"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14652",
"repo": "JordanSantiagoYT/FNF-JS-Engine",
"url": "https://github.com/JordanSantiagoYT/FNF-JS-Engine/issues/671"
}
|
gharchive/issue
|
custom ratings help
Describe your problem here. If you are modding, please try testing your problem in a clean version of the engine instead, this ensures the problem is actually caused by the engine itself. Also, if you're using an older version of JSE, please try the latest action build in the Actions tab or the latest release build. The issue has probably been fixed since that version. Oh yeah, and also be sure to check the pinned 'JS Engine: Known Issues' issue, to double check if your issue hasn't already been found!
i want to add like custom rating text
but whenever i add it, it works fine at 100% acc
but as soon as it drops below 100% it reverts to the default quotes
why do it do dis
Are you modding a build from source or with Lua?
Lua
What is your build target?
Windows x64
Did you edit anything in this build? If so, mention or summarize your changes.
aside from normal modding stuffs, no
Can you share a video/screenshot?
this is it at 100%
this it at literally anything else
idk what to do to fix it
(yes it has custom lines and everythin)
Did you make a Lua script for changing the quotes? If yes, it maybe is glitching/it's (maybe) not precise at all.
it doesnt need an lua script.. at least i dont think so
Did you make a Lua script for changing the quotes? If yes, it maybe is glitching/it's (maybe) not precise at all.
and to answer this properly, im not using an lua script
did you follow all these steps correctly? https://github.com/JordanSantiagoYT/FNF-JS-Engine/wiki/Adding-Rating-Quotes
i skipped like the whole making ratingQuotes file and tried to do it directly in assets/data/ratingQuotes cuz i thought it would mess with the mod
and after a quick restart and fixin the error i made it all work fine
|
2025-04-01T04:10:33.302674
| 2024-06-15T14:03:09
|
2354879068
|
{
"authors": [
"RealCox",
"moxie-coder"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14653",
"repo": "JordanSantiagoYT/FNF-PsychEngine-NoBotplayLag",
"url": "https://github.com/JordanSantiagoYT/FNF-PsychEngine-NoBotplayLag/issues/386"
}
|
gharchive/issue
|
Fix this please the engine is crashing every time
Describe your bug here. Mind you, if you are modding, please try testing your bug in a clean version of the engine instead, this ensures the bug is actually caused by the engine itself. Also, if you're using an older version of JSE, please try the latest action build in the Actions tab or the latest release build. The issue has probably been fixed since that version. Oh yeah, and also be sure to check the pinned 'JS Engine: Known Issues' issue, to double check if your issue hasn't already been found!
flixel/sound/FlxSound.hx (Line 603)
flixel/sound/FlxSound.hx (Line 743)
flixel/sound/FlxSound.hx (Line 571)
flixel/tweens/misc/NumTween.hx (Line 53)
flixel/tweens/FlxTween.hx (Line 1341)
flixel/FlxGame.hx (Line 748)
flixel/FlxGame.hx (Line 680)
flixel/FlxGame.hx (Line 553)
openfl/events/EventDispatcher.hx (Line 426)
openfl/display/DisplayObject.hx (Line 1398)
Uncaught Error: Null Object Reference
Please report this error my ballz: https://github.com/JordanSantiagoYT/FNF-PsychEngine-NoBotplayLag
Crash Handler written by: sqirra-rng
Command Prompt/Terminal/Crash logs (if existing)
No response
Have you identified any steps to reproduce the bug? If so, please describe them below in as much detail as possible. Use images if possible.
No response
Are you modding a build from source or with Lua?
Lua
What is your build target?
Windows
Did you edit anything in this build? If so, mention or summarize your changes.
No
If you use 1.12.0 or earlier, did you have Optimized Chart Loading turned on?
No
Did you check for any similar issues to what you're reporting? Check, then come back here. If there is a similar issue, then do not report the issue, otherwise it will be marked as a duplicate.
Yes
oh yeah that transform issue, only thing I can do to fix it is to shadow the FlxSound class and then add a null safety for the transform
but I do wanna find a better solution for that
closing due to no activity
|
2025-04-01T04:10:33.358670
| 2022-08-11T02:35:55
|
1335399042
|
{
"authors": [
"kylewatkins1202",
"mepc36",
"the-jasoney",
"vvruspat"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14654",
"repo": "JoshMarler/react-juce",
"url": "https://github.com/JoshMarler/react-juce/issues/297"
}
|
gharchive/issue
|
Pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed
When I tried implementing react-juce with my project, it came up with this error:
Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 42 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 47 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 52 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 69 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 74 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 80 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 85 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 91 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 96 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 42 Error C2039 'evaluateInline': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 42 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 47 Error C2039 'evaluate': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 47 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 52 Error C2039 'evaluateBytecode': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 52 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 69 Error C2039 'registerNativeProperty': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 69 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 74 Error C2039 'registerNativeProperty': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 74 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 80 Error C2039 'invoke': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 80 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 85 Error C2232 '->std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>::reset': left operand has 'class' type, use '.' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 85 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 91 Error C2039 'debuggerAttach': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 91 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 96 Error C2039 'debuggerDetach': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 96
Platform: Windows 11 64-bit
Compiler: MSVC
IDE: VS2022
JUCE: 6.1.6
When I tried implementing react-juce with my project, it came up with this error: Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 42 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 47 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 52 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 69 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 74 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 80 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 85 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 91 Error (active) E0393 pointer to incomplete class type "reactjuce::EcmascriptEngine::Pimpl" is not allowed neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 96 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 42 Error C2039 'evaluateInline': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 42 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 47 Error C2039 'evaluate': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 47 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 52 Error C2039 'evaluateBytecode': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 52 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 69 Error C2039 'registerNativeProperty': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 69 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 74 Error C2039 'registerNativeProperty': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 74 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 80 Error C2039 'invoke': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 80 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 85 Error C2232 '->std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>::reset': left operand has 'class' type, use '.' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 85 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 91 Error C2039 'debuggerAttach': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 91 Error C2027 use of undefined type 'reactjuce::EcmascriptEngine::Pimpl' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 96 Error C2039 'debuggerDetach': is not a member of 'std::unique_ptr<reactjuce::EcmascriptEngine::Pimpl,std::default_delete<reactjuce::EcmascriptEngine::Pimpl>>' (compiling source file ..\..\JuceLibraryCode\include_react_juce.cpp) neutron_SharedCode [my_path_to_react_juce]\react_juce\core\EcmascriptEngine.cpp 96 Platform: Windows 11 64-bit Compiler: MSVC IDE: VS2022 JUCE: 6.1.6
solved that by adding REACTJUCE_USE_HERMES=1
REACTJUCE_USE_HERMES=1
Where did you ad this too?
REACTJUCE_USE_HERMES=1
Where did you add this too?
|
2025-04-01T04:10:33.364876
| 2016-04-23T21:59:03
|
150601193
|
{
"authors": [
"Josverl"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14655",
"repo": "Josverl/Connect-O365",
"url": "https://github.com/Josverl/Connect-O365/issues/1"
}
|
gharchive/issue
|
it should not be needed to update the script each time that a new binary module is avaialble
Currently the configuration is stored as a JSON string as part of the script.
The goal is to externalize this data and store it on GitHub for better maintenance.
the assumption is that when installation is needed, the machine that can perform the installation does have access to the internet to download the modules.
#2 data is now published to git
|
2025-04-01T04:10:33.421844
| 2022-02-05T11:25:36
|
1124904940
|
{
"authors": [
"Juansero29"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14656",
"repo": "Juansero29/Loop",
"url": "https://github.com/Juansero29/Loop/issues/8"
}
|
gharchive/issue
|
[INIT] Change app language from app settings at any given moment
Change app language between English and French at any given moment, changing all non-user-input text in the app to the required languages. More languages may be added in the future.
Related Docs:
Internationalization (i18n) ngx-translate
https://www.freecodecamp.org/news/how-to-add-localization-to-your-react-app/
Translating "IonTabBar" in "App.tsx"
https://forum.ionicframework.com/t/page-transition-with-black-glitch/195698/4
https://blogs.infinitesquare.com/posts/web/ajouter-des-traductions-dans-votre-app-react-avec-i18n
https://ionicframework.com/docs/api/select
https://ionicframework.com/docs/api/select-option
|
2025-04-01T04:10:33.439665
| 2020-10-16T23:21:35
|
723582734
|
{
"authors": [
"codecov-io",
"khosravipasha"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14657",
"repo": "Juice-jl/LogicCircuits.jl",
"url": "https://github.com/Juice-jl/LogicCircuits.jl/pull/46"
}
|
gharchive/pull-request
|
initial githubaction
Just messing around with github actions. Might make it easier in future to have gpu tests, if the JuliaGPU gitlab solution does not work out.
Codecov Report
Merging #46 into master will decrease coverage by 0.81%.
The diff coverage is 45.76%.
@@ Coverage Diff @@
## master #46 +/- ##
==========================================
- Coverage 83.96% 83.14% -0.82%
==========================================
Files 29 29
Lines 2126 2178 +52
==========================================
+ Hits 1785 1811 +26
- Misses 341 367 +26
Impacted Files
Coverage Δ
src/Utils/data.jl
58.82% <42.85%> (-8.35%)
:arrow_down:
src/transformations.jl
97.08% <100.00%> (+0.48%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update fef48c0...6f8361d. Read the comment docs.
Travis is failing but should be okay, github actions pass the (same) tests.
|
2025-04-01T04:10:33.445270
| 2021-04-12T07:52:00
|
855669897
|
{
"authors": [
"pbayer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14658",
"repo": "JuliaActors/Actors.jl",
"url": "https://github.com/JuliaActors/Actors.jl/issues/26"
}
|
gharchive/issue
|
World Age Error when changing behavior
If we change behavior of an actor to a newly created λ-function with acquaintance parameter, we get a
world-age error if we try to execute it
julia> using Actors
julia> import Actors: spawn
julia> me = newLink()
Link{Channel{Any}}(Channel{Any}(32), 1, :local)
julia> myactor = spawn((lk,x,y)->send(lk,x+y), me)
Link{Channel{Any}}(Channel{Any}(32), 1, :default)
julia> send(myactor, 123, 456)
(123, 456)
julia> receive(me)
579
julia> become!(myactor, (lk,x,y)->send(lk,x^y), me); # now I change the behavior
julia> send(myactor, 123, 456)
(123, 456)
julia> receive(me)
Actors.Timeout()
julia> info(myactor)
Task (failed) @0x0000000172db0bc0
MethodError: no method matching (::var"#28#29")(::Link{Channel{Any}}, ::Int64, ::Int64)
The applicable method may be too new: running in world age 29647, while current world is 29648.
Closest candidates are:
(::var"#28#29")(::Any, ::Any, ::Any) at REPL[65]:1 (method too new to be called from this world context.)
Stacktrace: ...
In that case Actors doesn't call invokelatest on the behavior function because it is not parameterless.
This use case is not very likely but perfectly legitimate.
Two solution strategies
Get the actor world-age at startup and call invokelatest on every method with a later world age.
Call invokelatest on any anonymous function.
Discussion
This would be my preferred solution but it must use Julia's internals and may break with a later release. It would work with
julia> convert(Int, @cur_world())
29649
with @cur_world defined as
pworld(f) = methods(f).ms[1].primary_world ## will that change in the future?
macro cur_world()
:(pworld(()->1))
end
The 2nd solution will only cover anonymous functions and not functions imported from a script or module after the actor was started.
What should I do? Any other caveats, solutions?
This problem only arises after become! or become. Therefore become could set a flag in the behavior indicating if invokelatest should be called. This could be a keyword argument of the become function.
|
2025-04-01T04:10:33.447099
| 2024-06-17T14:05:05
|
2357441508
|
{
"authors": [
"DanielVandH"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14659",
"repo": "JuliaArrays/InfiniteArrays.jl",
"url": "https://github.com/JuliaArrays/InfiniteArrays.jl/pull/179"
}
|
gharchive/pull-request
|
Allow findfirst(isone) and findfirst(iszero) for InfRanges
Fixes #178
Also changes my previous PR #177 to use the new findfirst(iszero)
This doesn't work for reversed ranges..
findfirst(==(1), 10:-1:-∞) # should be 10?
Should it? Probably it's own issue since this issue already exists on master.
|
2025-04-01T04:10:33.505256
| 2019-11-29T16:45:57
|
530430180
|
{
"authors": [
"ChrisRackauckas",
"NirvikNU"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14660",
"repo": "JuliaDiffEq/SciPyDiffEq.jl",
"url": "https://github.com/JuliaDiffEq/SciPyDiffEq.jl/issues/4"
}
|
gharchive/issue
|
Speed issue
Is there any overhead or are these solvers inherently very slow compared to DifferentialEquations?
See the readme. Run the overhead measurement on your computer to double check. I found this wrapper to be 3x faster than using SciPy from Python with Numba (because of the function Jitting). See what you get!
Okay. Got it!
Is it the same for MATLABDiffEq and deSolveDiffEq?
I just want to ensure that the benchmarking is comparable when run directly
from Julia or if they need to be run from their own languages.
On Fri, Nov 29, 2019 at 12:02 PM Christopher Rackauckas <
<EMAIL_ADDRESS>wrote:
See the readme. Run the overhead measurement on your computer to double
check. I found this wrapper to be 3x faster than using SciPy from Python
with Numba (because of the function Jitting). See what you get!
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/JuliaDiffEq/SciPyDiffEq.jl/issues/4?email_source=notifications&email_token=ANOXCNMIWLQB4W2DK43HXTDQWFKNFA5CNFSM4JTBWJY2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEFPLZLI#issuecomment-559856813,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ANOXCNJM24OHQRLEXGF74GTQWFKNFANCNFSM4JTBWJYQ
.
See the READMEs in each. MATLABDiffEq is about 1-1, deSolveDiffEq introduces an overhead of around 3x.
Got it...another issue is that I see that some solvers like the Sundials,
native Julia solvers come with their own interpolant. So can the
interpolant be considered an intrinsic part of the algorithm? I know for
example in MATLAB the deval function is used to interpolate from the
solution structure (although the MATHWORKS manual page on deval doesnt
mention what type of interpolation is used for each algorithm). How does
SciPyDiffEq, deSolveDiffEq and MATLABDiffEq deal with such interpolations?
Or does it do a linear interpolation for these algorithms? In that case
will comparison across languages/algorithms depend on the type of
interpolation used?
On Mon, Dec 2, 2019 at 3:50 PM Christopher Rackauckas <
<EMAIL_ADDRESS>wrote:
See the READMEs in each. MATLABDiffEq is about 1-1, deSolveDiffEq
introduces an overhead of around 3x.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/JuliaDiffEq/SciPyDiffEq.jl/issues/4?email_source=notifications&email_token=ANOXCNPQPEGJTDSIVGVY5C3QWV7JDA5CNFSM4JTBWJY2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEFVP2EI#issuecomment-560659729,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ANOXCNIG3QQFD6DHZSGPHYTQWV7JDANCNFSM4JTBWJYQ
.
some solvers like the Sundials, native Julia solvers come with their own interpolant
Sundials doesn't expose it post solution, so we utilize their internal one in the integrator but use a Hermite from the solve (since it allows saving du which is all that's needed for Hermite). Since it's 3rd order it works about as well as the Sundials internal one anyways though.
I know for example in MATLAB the deval function is used to interpolate from the solution structure
We don't use that yet, though we could if you have a use for it. Given the speed of MATLAB though, I generally don't put much effort into giving that library all of the features.
How does SciPyDiffEq, deSolveDiffEq and MATLABDiffEq deal with such interpolations?
deSolve doesn't have an interpolation, so it's just linear. SciPy exposes it, so it defaults to using the one returned from SciPy (i.e. it has full integration).
Thanks for clarifying. I dont know what algorithm MATLAB uses for deval,
but I ran some benchmarks to compare the solutions directly from MATLAB and
via Julia and the results are pretty much the same. So I presume its
inconsequential.
On Mon, Dec 2, 2019 at 6:28 PM Christopher Rackauckas <
<EMAIL_ADDRESS>wrote:
some solvers like the Sundials, native Julia solvers come with their own
interpolant
Sundials doesn't expose it post solution, so we utilize their internal one
in the integrator but use a Hermite from the solve (since it allows saving
du which is all that's needed for Hermite). Since it's 3rd order it works
about as well as the Sundials internal one anyways though.
I know for example in MATLAB the deval function is used to interpolate
from the solution structure
We don't use that yet, though we could if you have a use for it. Given the
speed of MATLAB though, I generally don't put much effort into giving that
library all of the features.
How does SciPyDiffEq, deSolveDiffEq and MATLABDiffEq deal with such
interpolations?
deSolve doesn't have an interpolation, so it's just linear. SciPy exposes
it, so it defaults to using the one returned from SciPy (i.e. it has full
integration).
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/JuliaDiffEq/SciPyDiffEq.jl/issues/4?email_source=notifications&email_token=ANOXCNODNJC52WVOJPVNOC3QWWR2RA5CNFSM4JTBWJY2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEFXU3QY#issuecomment-560942531,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ANOXCNJISFA3LMMAY65QWKLQWWR2RANCNFSM4JTBWJYQ
.
Not much else here. Cheers!
|
2025-04-01T04:10:33.555546
| 2015-10-08T15:41:14
|
110478599
|
{
"authors": [
"Sisyphuss",
"stevengj"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14662",
"repo": "JuliaLang/IJulia.jl",
"url": "https://github.com/JuliaLang/IJulia.jl/issues/371"
}
|
gharchive/issue
|
Feature request: a magic to toggle on/off contents shown in Out[] (returned value, standard output, warning)
Update: the returned value showed in Out[ ] by executing a code cell
What return value and warning are you talking about? You mean the return value from executing a code cell?
Yes, the one showed in Out[ ]:.
You can just end the cell with a semicolon to suppress the Out[*] return value.
I changed the title of this issue. This feature can help quick development/presentation switch.
Maybe this is more an IPython issue...
If you want a global setting to suppress all output, I suppose that would should be filed with Jupyter. I still can't figure out what you're trying to accomplish, though.
@stevengj I know what you mean. Thanks.
|
2025-04-01T04:10:33.562805
| 2015-12-10T15:03:36
|
121505871
|
{
"authors": [
"nalimilan",
"sschmitz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14663",
"repo": "JuliaLang/WinRPM.jl",
"url": "https://github.com/JuliaLang/WinRPM.jl/issues/59"
}
|
gharchive/issue
|
Make sure downloaded DLLs are executable
To make use of dynamic linking under Windows, the user must have execution permissions on the DLL file. For default installations, these are not usually restricted -- all files are "executable" from a permissions viewpoint, but only files with the proper extension are actually executed.
Since I'm using the same partition as both my Windows and Linux home folder, I set it up in a way so that newly created files do not have the executable permission set (otherwise, file managers on linux give me a "do you want to open or execute this file?" prompt every time i try to open something). This also affects the files that WinRPM downloads. For packages that use WinRPM as a BinDeps backend, this causes the issatisfied check to fail, since the call to dlopen_e meant to check for the library's existence fails (prompting me to dig through the code of all these packages for two days when I'm just installing julia for the first time ;-)).
I recognize that this is only a problem for specific odd setups as mine, but the Windows analogue of a simple chmod +x would suffice to be on the safe side here.
Would you try to prepare a pull request to do that, with a comment in the code explaining the specific case where this is needed?
I can try. As I said, I have never worked with julia before, but then again, it should be rather straightforward.
... or so I thought. As laid out in JuliaLang/julia#7574, the standard library permissions functions (including stat and chmod) don't make too much sense on Windows right now, and I couldn't find any Julia package to deal with ACLs. It would probably need some C hacking to do this, but, that's probably not worth it. So a proper solution would have to wait for some such package to be written.
Since this issue is unlikely to affect a significant number of people, it's fine with me if you want to close it.
|
2025-04-01T04:10:33.804388
| 2023-06-21T11:39:44
|
1767396720
|
{
"authors": [
"TSGut"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14664",
"repo": "JuliaLinearAlgebra/InfiniteLinearAlgebra.jl",
"url": "https://github.com/JuliaLinearAlgebra/InfiniteLinearAlgebra.jl/pull/140"
}
|
gharchive/pull-request
|
SymTridiagonal support for cholesky
Fix https://github.com/JuliaLinearAlgebra/InfiniteLinearAlgebra.jl/issues/124
Makes cholesky accept SymTridiagonal.
@dlfivefifty Ready for merge or review.
|
2025-04-01T04:10:33.808754
| 2024-05-12T01:05:25
|
2291119354
|
{
"authors": [
"cortner"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14665",
"repo": "JuliaMolSim/EmpiricalPotentials.jl",
"url": "https://github.com/JuliaMolSim/EmpiricalPotentials.jl/issues/13"
}
|
gharchive/issue
|
Suggest renaming get_id
I think this is too generic a name, and we should choose something like get_particle_id.
moved to AtomsCalculatorsUtilities
|
2025-04-01T04:10:33.835964
| 2019-12-18T07:52:30
|
539515636
|
{
"authors": [
"kellertuer",
"mateuszbaran",
"sethaxen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:14666",
"repo": "JuliaNLSolvers/Manifolds.jl",
"url": "https://github.com/JuliaNLSolvers/Manifolds.jl/issues/65"
}
|
gharchive/issue
|
[RFC] Graph manifold
I would like to use the holidays to think about modelling another manifold that extends the PowerManifold to some extend: Manifold-valued data on a graph. The idea stems from this paper and at that time was implemented in Matlab but I was never happy with the implementation. I started some ideas in Julia, but I would like to do that efficiently. It's actually a nice interplay of data on vertices and edges on a graph.
The setting
Let G be a graph with V its set of vertices and E the set of edges (each edge having a start point end an end point from V) and M be a manifold. Then we have four types of data
manifold-valued data on the vertices, i.e. f: V-> M
manifold-valued data on the on the edges, i.e. g: E -> M
tangent vectors on edges, i.e. W: E->TM
tangent vectors on vertices, i.e. Z: V-> TM (which might not that much be used)
If we call the GraphManifold GM, the idea is, that for example log(GM, f, g) for two manifold-valued data sets on the vertices yields tangent vectors on the edges, namely for an edge (v,w) in E we get the term log_{f_v} g_w. We can then mimic (see literature above) the Graph Laplace by producing Z: V -> TM with Z(v) = sum of all edges that start in v of these logs (note they are all log_f(v) - terms. And for example then perform a exp(GM,f,Z)
RFC
I am not 100% sure how such an approach works for manifold-valued data on the edges, but I am curious to ponder about that
My main question for comments is: What is a good and efficient way to represent this? the data f can be done similar to the Power manifold, but the GM also has to store its graph (adjacency matrix for example). Here I would like to ask for ideas/comments what to use – LightGraphs.jl for example?
I wonder how this relates to the noncommutative discrete differential geometry as described here: https://phorgyphynance.wordpress.com/ or here: https://ncatlab.org/ericforgy/files/mexico.pdf ?
Regarding efficient representation, what kinds of relatively complex operations would be the most common? Adding/deleting edges? Computing graph logs and exps? Something else?
Summing over all adjacent edges to assign a vector to a vertex is maybe also something. The exp(GM,f,Z) can be seen as a classical one on the power manifold |V|.
And yes the graph can model a discrete differential geometry, but at first glance your two links look like they act on the integer lattice?
Summing over all adjacent edges to assign a vector to a vertex is maybe also something. The exp(GM,f,Z) can be seen as a classical one on the power manifold |V|.
I guess it might be the best to start with something relatively simple, like a digraph from LightGraphs.jl with custom weights, represented for example by dict from node/edge to an array representing a point or a tangent vector. Alternatively node meta-information can be represented by a vector indexed by node numbers and edge meta-information can be represented by (potentially sparse) matrices. There is MetaGraphs.jl but the way they represent meta-information looks really performance-unfriendly (unless someone is going to do lots of adding or deleting of such meta-information).
And yes the graph can model a discrete differential geometry, but at first glance your two links look like they act on the integer lattice?
As I understand, regular lattices are used just for demonstration of the simplest cases. You can look at section 2.1 of the pdf I've linked. @EricForgy might be able to provide some input here.
I think I would prefer LightGraphs.jl, storing the points in an array (of size of the number of points or edges) anyways. What I am not yet so sure about is how to distinguish data on edges(vertices when it comes to dispatch for exp/log/retractions etc.
I think I would prefer LightGraphs.jl, storing the points in an array (of size of the number of points or edges) anyways.
An array would faster for reading and writing edge/vertex data but slower during graph modification. Anyway, it's also a good choice.
What I am not yet so sure about is how to distinguish data on edges(vertices when it comes to dispatch for exp/log/retractions etc.
What do you mean by distinguishing data here?
I actually have a "graph manifold" type of application, though maybe it doesn't fit in to what you're proposing here.
I'm doing inference on ensembles of protein structures, which we can represent with a graph. A single protein structure can be decomposed into a set of connected rigid bodies, fixing the topology. We interpret the nodes as points on SE(3) (rigid body orientations, i.e. shifts), and we interpret the edges likewise as points on SE(3) (perturbations). We do this so we can construct a probability distribution on the protein by assigning distributions to the edges. We can relate that to observables such as atomic distance distributions by convolving the distributions along the path between two nodes. Since those distributions have parameters, in reality our manifold is a statistical manifold with Dirac mean transformations at the nodes and Brownian/IsotropicDiffusion distributions at the edges. But any draw from that distribution would live on the graph manifold.
I am not sure whether my exp/log ideas above apply (those mixing edges/vertices), but still even with the default exp/log (vertex-vertex or edge-edge as argument, which is pasically power manifold stuff) I think your application fits perfectly :)
What do you mean by distinguishing data here?
I am not 100% sure whether that's necessary. The interesting part ist, that there are different interplays between vertex-data and edge-data. Let's see that in detail and take
x, y as manifold-valued data on V
v tangent to x so tangent-valued data on V)
a, b manifold-valued data on E
w a tangent-valued data on E corresponding to a
Then
exp(GM,x,v) should be a y as manifold-valued vertex data (element wise exp on the vertices)
exp(GM,a,w) the same on the edges, i.e. a b on the edges
log(GM,x,y) yields edge-valued data w, where for any edge (e,f), i.e. e,f \in V are vertices we store w( (e,f) ) = log(M, x_e, y_f) or is this inconsistent? Because even log(GM,x,x) would yield a nonzero w and that seems inconsistent. Otherwise, log(GM,x,y) could also be seen element (vertex-) wise. Then even more the first case need another function/name/types.
Similar issues arise for the log on the edges. So despite distinguishing data I maybe also mean different functions/names.
I think exp and log should primarily work on the whole graph manifold (vertices + edges). Operations on edges or vertices only could probably also be exp and log, differentiated by a keyword argument.
If you want to get tangent vectors on edges from points on vertices, that's not really a log. Maybe we should call it node_log? There could also be a node_exp that works the other way.
Because even log(GM,x,x) would yield a nonzero w and that seems inconsistent.
That's I think because this type of log doesn't make a vector space consistent with the corresponding exp.
By the way, do you consider the graph itself as an object shared between points and tangent vectors, not connected to the manifold object?
Oh, I actually never had data on vertices and edges. Just vertices or edges.
I consider the graph being a property of the manifold, i.e. it is a field of the GraphManifold.
Oh, I actually never had data on vertices and edges. Just vertices or edges.
This is an important point 🙂 . Then we should probably have a type parameter that tells whether a graph manifold is a node-manifold or edge-manifold.
Hm, but (in my mind) the GraphManifold is both and it depends on which data (node data or edge data) as parameters or returns you have? At least in my Matlab and the first Skecth for Julia I had such different data types.
So to make it more precise – a log(GMe,x,x) would return an edge-valued data since GMe is an edge manifold (and it would be the logs to neighbors) and log(GMv,x,x) would be node-valued data (and the zero tangent) then?
The one point I am not sure about it is, whether it would maybe also be an argument of exp/log (if a GraphManifold is both edge and vertex-manifold) and the same for all others where it's not uniquely determined.
For one thing, a manifold is supposed to have a single dimension for all its points, and the dimensions of the edge-manifold and node-manifold are usually different. What would manifold_dimension return if it didn't know whether it represents nodes or edges?
So to make it more precise – a log(GMe,x,x) would return an edge-valued data since GMe is an edge manifold (and it would be the logs to neighbors) and log(GMv,x,x) would be node-valued data (and the zero tangent) then?
Well, something similar to this. The vertex manifold should have both logs -- the node-valued one and the edge-valued one (although I'm not really convinced that calling the second one just log is a good idea).
Oh, the manifold_dimension argument is a good one. You convinced me, we should do two manifolds, then, either
GraphEdgeManifold or EdgeGraphManifold (the second is ordered by size kind of)
GraphVertexManifold or VertexGraphManifold
or one could to that as a parameter as GraphManifold{Vertex,M} and GraphManifold{Edge,M}, which would make it easier to extend to a VertexEdge (i.e. really data on both) or even something HyperGraph-kind (subsets of E) data.
What do you think about the naming?
And yes, maybe it's better to name for log(GMe,x,x) would be incident_log, which would also work for manifold-valued edge data incident_log(GMv,a,b), where it could compute the mean for a vertex (of its incident edges) and return the sum of the logs from there to the edge-values.
I like having a single type GraphManifold with a type parameter that determines whether it's a vertex or edge manifold. This way the graph-related operations can be more easily shared between both.
I think incident_log is a good name for the second type of log.
would also work for manifold-valued edge data incident_log(GMv,a,b), where it could compute the mean for a vertex (of its incident edges) and return the sum of the logs from there to the edge-values.
What would that operation be useful for?
Cool! So we have the same preference on the GraphManifold.
I don't know what it's for, I just thought about what that operation might be and that sounds reasonable. Whether it's also useful – maybe. For someone.
By the way, I recently realized that it's likely possible to extend one of the neater tricks from my PhD thesis to graph manifolds. If you ever needed to find a path in a graph manifold that sort-of looks like a template path, there is a very cool way to find it. The "sort-of looks like" part refers to a Fisher-Rao type metric on a manifold of (discretized) curves.
Sounds quite cool, though I don't know yet where to use that ;)
What is a template path?
That's a known path-like graph manifold over the same basic manifold. It also works for simple DAGs instead of paths (then the algorithm would look for the closest pair of path in both graph manifolds).
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.