added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:37:30.983421
| 2022-05-05T18:42:54
|
1227046125
|
{
"authors": [
"Zelzahn",
"lisadejonghe"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2551",
"repo": "SELab-2/OSOC-5",
"url": "https://github.com/SELab-2/OSOC-5/issues/353"
}
|
gharchive/issue
|
unexpected side effect in "currentsteps" computed property
In MailsOverview.vue is the following code:
currentsteps() {
if (!this.mailStore.mails.has(this.student.id)) return []
const data = (this.mailStore.mails.get(this.student.id) ?? []).filter(mail => {
return this.statuses.includes(parseInt(mail.info))
})
if (!this.currentStep) {
this.currentStep = this.statuses.find(s => !data.map(d=>parseInt(d.info)).includes(s)) // <-- gives side effect error
}
return data
},
The indicated line gives an Unexpected side effect in "currentsteps" computed property. error for me.
That's something @Wouter01 added.
|
2025-04-01T06:37:30.986439
| 2023-04-18T10:33:48
|
1672830846
|
{
"authors": [
"GeertThijs",
"costezki",
"csnyulas"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2552",
"repo": "SEMICeu/style-guide",
"url": "https://github.com/SEMICeu/style-guide/issues/61"
}
|
gharchive/issue
|
Names of enumerations
In § Case sensitivity and charsety. Why should the name of an enumeration start with a lowercase letter? In the UML metamodel an enumeration is a specialisation of the classifier Datatype. Datatype names start with a capital, so why not enumerations?
The logic of this recommendation is to have consistency between the naming style of the enumerations declared in a CV or AP, and other externally maintained, and often reused, vocabularies and authority tables (e.g. at-voc:access-right, at-voc:language).
Does this make sense @GeertThijs ?
We are aware that many CVs and APs currently don't follow this recommendation. This is a style guide, with recommendations. It is up to each group and project to decide if they want to follow or not any given the recommendation in the guide.
We have updated the recommendation, relaxing it on the packages, data types and enumerations.
|
2025-04-01T06:37:30.994133
| 2021-08-09T18:41:18
|
964267260
|
{
"authors": [
"JuliePeeling",
"ernest19"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2553",
"repo": "SERVIR-WA/GALUP",
"url": "https://github.com/SERVIR-WA/GALUP/issues/172"
}
|
gharchive/issue
|
Ernest Opoku-Kwarteng
exercise 1.csv
exercise 2.csv
exercise 3 - rankreciprocal.pdf
exercise 3 - ranksummethod.pdf
Well done, Ernest. You clearly grasped the material outlined in Module 4 and successfully applied it to these exercises. Thank you for your participation! Please be sure to fill out the post-Module survey before you begin the next Module.
|
2025-04-01T06:37:30.995695
| 2020-03-11T10:49:41
|
579172315
|
{
"authors": [
"andyAndyA"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2554",
"repo": "SERaaS/SERaaS-Web-Platform",
"url": "https://github.com/SERaaS/SERaaS-Web-Platform/issues/11"
}
|
gharchive/issue
|
API endpoint usage option from Web Platform
I would like to: Provide an API endpoint usage option from Web Platform
So that I can: Allow users to make a test API call from the Web Platform and visualise it
Done in endpoint-usage-option branch.
|
2025-04-01T06:37:31.105484
| 2017-12-30T12:37:06
|
285200675
|
{
"authors": [
"Ceylo",
"Foaly",
"JonnyPtn",
"LaurentGomila",
"MarioLiebisch",
"binary1248",
"dabbertorres",
"eXpl0it3r",
"eliasdaler",
"mantognini"
],
"license": "Zlib",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2555",
"repo": "SFML/SFML",
"url": "https://github.com/SFML/SFML/pull/1335"
}
|
gharchive/pull-request
|
Initial support for SFMLConfig.cmake, for review
This PR is for review and feedbacks, not supposed to be merged yet!
TODO list:
[x] Have config file generated on all platforms
[x] Support debug/release config
[ ] Have no absolute path in generated config files on Linux / Windows / macOS / iOS? / Android?
[ ] Export internal dependencies in case of static build (ie. remove the need for SFML_DEPENDENCIES)
[ ] Support for components ?
[ ] Exhaustive testing!
=============
I've started doing the changes to support config file generation. At the moment this is working at least for me on macOS with frameworks. I expect it to work on other platforms too because it reuses the definitions of current SFML targets. This implies a LOT of changes though so I would like your feedback before going on.
Especially I have replaced all the listing of dependencies with calls to target_link_libraries() and target_include_directories() for 3 reasons:
this avoids redoing what CMake already does (target_link_libraries() and target_include_directories() already handle lists)
this makes it easier to group the full setup (include+link) of each dependency
this will allow specifying which dependencies must be public in case of static linking (what SFML_DEPENDENCIES currently does in FindSFML.cmake), same eventually if compile definitions or include dirs must be exported (currently only the root include dir of SFML is exported)
In terms of usage right now (example from sfeMovie) of course without any FindSFML.cmake:
find_package (SFML 2.3 COMPONENTS graphics window system audio REQUIRED)
target_link_libraries(sfeMovie PRIVATE sfml-graphics sfml-window sfml-system sfml-audio)
I didn't expose yet the fact that sfml-graphics depends on sfml-window and sfml-system. Could be done and would allow writing only
target_link_libraries(sfeMovie PRIVATE sfml-graphics sfml-audio)
Dunno yet if we want that though.
If you want to look at what the generated config looks like for now:
http://yalir.org/files/SFML/
It gets installed in /usr/local/lib/cmake/SFML by default.
References
https://cmake.org/cmake/help/latest/manual/cmake-packages.7.html#creating-packages
https://cmake.org/cmake/help/latest/command/find_package.html
https://cmake.org/cmake/help/latest/command/export.html
https://cmake.org/cmake/help/latest/module/CMakePackageConfigHelpers.html
https://github.com/SFML/SFML/compare/master...SrTobi:support-config-file-packages
Related issues
https://github.com/SFML/SFML/issues/758
https://github.com/SFML/SFML/issues/937
Amazing work, thanks :+1:
I know very little about modern CMake usage, but what I see looks ok. We should now test it under various scenarios (static/dynamic link, on all OSes).
this will allow specifying which dependencies must be public in case of static linking
What exactly does that mean? Will CMake now automatically export the private dependencies in case of static linking, or does it require additional work? In any case, we no longer need SFML_DEPENDENCIES, right?
I didn't expose yet the fact that sfml-graphics depends on sfml-window and sfml-system. Could be done
Should be done, since the dependency does exist.
Didn't dig too deep into the changes, but with the branch as-is, sfml-main won't build, since its include directories aren't set (and therefore won't find SFML/Config.hpp).
You're basically missing this:
# set the include directory
target_include_directories(sfml-main PUBLIC
$<BUILD_INTERFACE:${PROJECT_SOURCE_DIR}/include>
$<INSTALL_INTERFACE:include>)
from sfml-main's CMakeLists.txt.
Also are you sure about the location of the SFMLConfig.cmake and the other files? I thought they'd have to be placed directly in lib or am I confusing it with PkgConfig?
I know very little about modern CMake usage, but what I see looks ok. We should now test it under various scenarios (static/dynamic link, on all OSes).
The most important part is correctly defining which target properties have PRIVATE, PUBLIC or INTERFACE visibility. Are you familiar with it?
Basically PRIVATE properties are used only when building the target.
PUBLIC properties are used when building the target and by targets that depend on your target.
INTERFACE properties are used only by targets that depend on your target.
And in these properties you can define include directories, libraries to link or preprocessor definitions.
What exactly does that mean? Will CMake now automatically export the private dependencies in case of static linking, or does it require additional work? In any case, we no longer need SFML_DEPENDENCIES, right?
It will, at least that's what I want. But it's not done yet with current changes. And yes SFML_DEPENDENCIES won't be needed anymore if I get it to work. Will update the PR when it's ready.
Should be done, since the dependency does exist.
Right, will update when it's done
Didn't dig too deep into the changes, but with the branch as-is, sfml-main won't build, since its include directories aren't set (and therefore won't find SFML/Config.hpp).
Indeed. Didn't do it yet as it's not needed on macOS, because I wanted feedbacks before going further. Will fix it.
Also are you sure about the location of the SFMLConfig.cmake and the other files? I thought they'd have to be placed directly in lib or am I confusing it with PkgConfig?
There are several possible paths, see https://cmake.org/cmake/help/latest/command/find_package.html
"Each entry is meant for installation trees following Windows (W), UNIX (U), or Apple (A) conventions:"
<prefix>/ (W)
<prefix>/(cmake|CMake)/ (W)
<prefix>/<name>*/ (W)
<prefix>/<name>*/(cmake|CMake)/ (W)
<prefix>/(lib/<arch>|lib|share)/cmake/<name>*/ (U)
<prefix>/(lib/<arch>|lib|share)/<name>*/ (U)
<prefix>/(lib/<arch>|lib|share)/<name>*/(cmake|CMake)/ (U)
<prefix>/<name>*/(lib/<arch>|lib|share)/cmake/<name>*/ (W/U)
<prefix>/<name>*/(lib/<arch>|lib|share)/<name>*/ (W/U)
<prefix>/<name>*/(lib/<arch>|lib|share)/<name>*/(cmake|CMake)/ (W/U)
<prefix>/<name>.framework/Resources/ (A)
<prefix>/<name>.framework/Resources/CMake/ (A)
<prefix>/<name>.framework/Versions/*/Resources/ (A)
<prefix>/<name>.framework/Versions/*/Resources/CMake/ (A)
<prefix>/<name>.app/Contents/Resources/ (A)
<prefix>/<name>.app/Contents/Resources/CMake/ (A)
Still need to figure out where it'll be located in the end. It's probably not portable on this part right now.
I initiated the CI builds, checkout the build logs (Config.hpp can't be found).
Which is exactly what I've stated above. ;)
Build is fixed for sfml-main, everything builds fine locally on Windows (VS2017 32bits).
The dependencies between SFML modules are now public so examples like VOIP only need to depend on sfml-network and sfml-audio, no need to specify sfml-system. The same applies to external projects that will use find_package(SFML) and depend on a SFML module.
I ran the install step with the Debug/Release configuration and did a find_package(SFML) without any FindSFML.cmake and it did work on Windows, with the correct lib depending on selected configuration (Debug/Release).
There doesn't look to be any absolute path in the generated config file on Windows, which is nice for making releases. On the contrary for now the config file generated on macOS has absolute paths (/Library/Frameworks). To be fixed.
Still need to support static builds, most likely the hardest part because of the need to expose internal dependencies.
One important note
I changed a bit how the sfml-main library is created. Instead of custom add_library() and properties, it's created like any other SFML library : with suffix for debug / static, and with pdb next to the generated .lib file. The only customization I kept is that it always remains a static lib, whatever the value of BUILD_SHARED_LIBS.
Nice. I'd say absolute paths aren't that bad for now. Wouldn't they be the same on any machine anyway? No experience with Frameworks, but from my understand they're supposed to sit in that specific directory?
SFMLBuildMaster: Build this please.
New revision is building fine for me on Windows. Do we even have to keep (and therefore maintain) FindSFML.cmake? Or should it be deprecated/removed altogether? From my understanding we wouldn't need both. Or is there some minimum version requirement higher than using a Find….cmake file?
SFMLBuildMaster: Build this please.
Nice. I'd say absolute paths aren't that bad for now. Wouldn't they be the same on any machine anyway? No experience with Frameworks, but from my understand they're supposed to sit in that specific directory?
It depends on what you want to do. Currently FindSFML.cmake allows finding SFML in standard paths or in custom paths if you set SFML_ROOT. I don't see any reason to restrict the benefits of CMake config files to standard installations only.
Consider for example where you want to make sure that the package you're creating for your program is self-contained. In that case you don't want to have SFML installed in standard paths. You can also want to keep standard paths clean, or don't have admin rights, etc.
New revision is building fine for me on Windows. Do we even have to keep (and therefore maintain) FindSFML.cmake ? Or should it be deprecated/removed altogether? From my understanding we wouldn't need both. Or is there some minimum version requirement higher than using a Find….cmake file?
I'm always for keeping a simple and unique way to do things. But I don't think I'm the one to choose this for SFML :)
As for the minimum requirement.. is SFML's CI using CMake 2.8? If it does then it's ok, because I didn't add any conditional check for some CMake version.
Also I realize that currently existing FindSFML.cmake won't find the new sfml-main lib because of changed suffix. Is it acceptable to introduce such break in SFML 2.5?
Also I realize that currently existing FindSFML.cmake won't find the new sfml-main lib because of changed suffix. Is it acceptable to introduce such break in SFML 2.5?
Do you mean the -s affix? Why not just overwrite it after defining the target? Or won't that work?
Do you mean the -s affix? Why not just overwrite it after defining the target? Or won't that work?
Yes the -s suffix, and yes overwriting does work. Committing, thanks.
@MarioLiebisch Do you know where sfml-graphics links against OpenGL ES libraries on Android? I can't find it on SFML/master and I broke link step of this module on Android: https://ci.sfml-dev.org/builders/android-armeabi-v7a-api13/builds/267/steps/compile/logs/stdio
I saw that you worked on Android support so asking you :)
It used to be here:
https://github.com/SFML/SFML/blob/master/src/SFML/Graphics/CMakeLists.txt#L126
but obviously got lost in the rewrite.
It's still there: https://github.com/Ceylo/SFML/blob/feature/CMakeTargetExport/src/SFML/Graphics/CMakeLists.txt#L124
And I don't think you pointed out the correct line, because in Config.cmake I see that Android uses OpenGL ES.
It seems like it was simply not set. But it previously worked because another signature of target_link_libraries() was used: https://cmake.org/cmake/help/latest/command/target_link_libraries.html?highlight=target_link_libraries#libraries-for-both-a-target-and-its-dependents
Which imported linkage for OpenGL ES from sfml-window target definition : https://github.com/SFML/SFML/blob/master/src/SFML/Window/CMakeLists.txt#L270
And it's no more the case with the explicit PRIVATE linkage (it was implicitly PUBLIC). Hopefully easy to fix :)
But that's outside the SFML_OPENGL_ES branching starting in line 113.
SFMLBuildMaster: Build this please.
Remaining discussed points are done!
When building macOS frameworks, CMAKE_INSTALL_PREFIX should be set by user to /Library/Frameworks for SFML to be directly usable. I've added SFML_DEPENDENCIES_INSTALL_PREFIX and SFML_MISC_INSTALL_PREFIX to support what was discussed. SFML_DEPENDENCIES_INSTALL_PREFIX defaults to /Library/Frameworks on macOS and to CMAKE_INSTALL_PREFIX on other platforms. SFML_MISC_INSTALL_PREFIX defaults to CMAKE_INSTALL_PREFIX except on macOS where it defaults to /usr/local. This allows changing CMAKE_INSTALL_PREFIX to install frameworks in /Library/Frameworks without modifying where "misc" contents goes. So basically by default nothing changes except the install prefix for macOS frameworks.
When creating the SFML package for distribution, you should run the "install" target at least 4 times with the same install prefix. Twice for debug/release x twice for static/dynamic. This will generate the following files:
SFMLConfig.cmake
SFMLConfigDependencies.cmake
SFMLConfigVersion.cmake
SFML${type}Targets.cmake
SFML${type}Targets-${config}.cmake
Where ${type} is "Static" or "Shared", and ${config} is one of CMake build configurations (Release, Debug, MinSizeRel, RelWithDebInfo).
The files that are installed by more that one installation configuration are identical between the different configurations that can write it.
For frameworks it's similar except that only release dynamic config needs to be installed to get a complete SFMLConfig.
Either SFMLStaticTargets.cmake or SFMLSharedTargets.cmake is included by SFMLConfig.cmake depending on whether the user has set SFML_STATIC_LIBRARIES. SFML${config}Targets.cmake includes its debug/release child files depending on which ones exist. Note that if a user does a find_package(SFML) with an installation that only contain debug libs for example, these debug libs are used even when he builds his project in release. Can be an issue on Windows but that's how CMake generates these files.
I did test with CMake 3.0.2 and it did work. I checked the installed files, there are slight differences but nothing that prevents using this version. Requirement can't be lower as there are many uses of interface libraries. Will commit the corresponding changes soon.
For dependencies I created interface libraries. This is done by sfml_find_package(). For example sfml_find_package(FLAC INCLUDE "FLAC_INCLUDE_DIR" LINK "FLAC_LIBRARY") creates a target named "FLAC", and when linking against that target, the target that needs it automatically gets the correct compile and link flags. And it looks just like if we'd just built the FLAC library in the same project. The important point is that the created target name in current example is "FLAC". That means that if a user also creates a FLAC target and uses SFMLConfig, there'll be a conflict. I tried to add a namespace to have something like sf::FLAC as target name but namespaces are not allowed for interface libraries. I also considered sfFLAC but it looks weird, requires some changes and in the end I think there's very low probability to have a conflict. So I just left "FLAC".
The variables previously defined by FindSFML.cmake are not defined by SFMLConfig.cmake except SFML_FOUND and SFML_DOC_DIR.
So I guess now is the time for final review & testing! Would be nice to test as a user, by deleting the FindSFML.cmake you may have in your project and generate the SFMLConfig and use it. Especially to check that no use case is missing.
SFMLBuildMaster: Build this please.
Awesome! Thanks for the detailed summary.
I won't be able able to test this for the next 10 days or so, but maybe someone else with a mac can give some feedback. :)
Another thing, in order to include this with the next SFML RELEASE, we also need to update the tutorials. Could you have a look at them? (But maybe it would be wise just to wait for some additional approval of this PR?)
Cmake wise, we also have #1344. Would you like to have a look? You seems more proficient than I to solve it. No pressure though :)
Thanks :)
As for the tutorials indeed I'd prefer feedbacks on usage first. Will check the current tutorials though to get an idea of what needs to be changed.
And for the other PR I'll check, but I prefer to finish this one first and focus. Also a lot of CMake code has changed so the changes for #1344 would most likely conflict.
Concerning last failed build, I've fixed the error on Windows. On macOS though it requires changes to the cmake command line used for CI, because of the introduced install prefixes for dependencies and misc files. I don't know how that can be done because I suppose that the CI config does not depend on the branch, right? I also don't know where the CI steps are defined in order to help on that matter. Basically CMAKE_INSTALL_FRAMEWORK_PREFIX is not used anymore, and both SFML_DEPENDENCIES_INSTALL_PREFIX and SFML_MISC_INSTALL_PREFIX should be set to what you want.
@mantognini For the tutorials you were thinking of FindSFML.cmake or something else? I didn't find any mention about it except in forums and changelogs.
@LaurentGomila @MarioLiebisch @eXpl0it3r @mantognini So what do we do? This PR has been ready for testing for 2 weeks now.
I made a topic on SFML's forum one week ago: https://en.sfml-dev.org/forums/index.php?topic=23676.0
Since that time some of you answered on the forum and fixes were done, but I'm not aware of anyone actually trying to use this for his project. So no feedback on real uses at the moment.
The more other PRs are merged in the meantime, the harder it is to make sure I fix all conflicts correctly, because codebases diverge. On that part, @MarioLiebisch could you check my latest merge commit? There were many conflicts related to your changes for Android. I think it's ok but… just in case.
Sorry, I wanted to give it a shot this week but, you know, life... I'll try to get around to test it in the middle of next week.
SFMLBuildMaster: Build this please.
That being said, anyone else is welcome to try this PR! :)
Looks like CMake doesn't have the permission to install the FreeType framework on the CI. Does this need to be adjusted on the CI or in the setup?
I hope to have a look at this soon. Went to a C++ user group meeting this week and learned some stuff on modern CMake. 🙂
For the install issue this is what was discussed in https://github.com/SFML/SFML/pull/1335#issuecomment-362847863
There a small changes needed on CI side.
Thanks for checking the PR soon !
Can you open an issue on https://github.com/SFML/SFML-Buildbot with details so @binary1248 can fix it?
@eXpl0it3r No need to open an issue if I am just able to fix it like that. 😉
@binary1248 You should keep CMAKE_FRAMEWORK_INSTALL_PREFIX in cmake configure step as long as this PR is not merged, otherwise you break macOS build on other branches at the moment. For current branch it’ll be an unused cmake arg but it’s temporarily ok.
@Ceylo I like it better like this... more pressure to merge this faster. 😛
So… still blocked with testing for one more week. What do we do?
I can give it a try again on all platforms but that will only make sure it works in my use cases.
What do you think?
SFMLBuildMaster: Build this please.
Our CI install frameworks incorrectly ATM
[...]
-- Installing: /Users/SFML/Desktop/buildbot/osx-clang-el-capitan/install/./SFML.framework
[...]
-- Installing: /Users/SFML/Desktop/buildbot/osx-clang-el-capitan/install/./sfml-system.framework
[...]
They should go into [...]/install/Library/Frameworks.
@binary1248, could you append /Library/Frameworks to -DCMAKE_INSTALL_PREFIX current value? (According to https://github.com/SFML/SFML/pull/1335#issuecomment-362751934.)
@mantognini That doesn't sound right... CMAKE_INSTALL_PREFIX is used as a base directory for everything, not only frameworks. There are already a bunch of things installed into /Library/Frameworks which is specified via SFML_DEPENDENCIES_INSTALL_PREFIX, but I guess some things that should go in there as well aren't sent there by the install script.
It might looks a bit weird, but it's only for when building/installing frameworks. See above comments by @Ceylo. Other misc files are installed according to another variable.
@mantognini Should be fixed in https://ci.sfml-dev.org/#/builders/11/builds/31/steps/20/logs/stdio
Besides the Xcode template script issue, everything seems good on macOS. 👍
Thanks @binary1248, this looks good.
No, a bit more than that. The new script handles ressources being installed
elsewhere (according to the dependencies install prefix). Your change is
consistent with the rest so it's good. I've simply changes a few other
lines.
On Sun, 4 Mar 2018, 17:49 Ceylo<EMAIL_ADDRESS>wrote:
@Ceylo commented on this pull request.
In tools/xcode/templates/SFML/SFML App.xctemplate/TemplateInfo.plist.in
https://github.com/SFML/SFML/pull/1335#discussion_r172055425:
@@ -146,7 +146,7 @@ subject to the following restrictions:
If you're using static libraries (which is not recommended) you should remove this script from your project.
SETTINGS
-CMAKE_INSTALL_FRAMEWORK_PREFIX="@CMAKE_INSTALL_FRAMEWORK_PREFIX@"
+CMAKE_INSTALL_FRAMEWORK_PREFIX="@CMAKE_INSTALL_PREFIX@"
I suppose you mean rename the variable CMAKE_INSTALL_FRAMEWORK_PREFIX to
something else?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/SFML/SFML/pull/1335#discussion_r172055425, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AAqU6y-JJ_n5Ox05iprPUXqoF9216hgUks5tbBszgaJpZM4RPltw
.
Ah right, I had not seen the link to your gist. Committing your changes.
Had a look at SFML tutorials for building SFML with CMake and here are the changes I suggest in the table containing CMake variable description (bold is new contents):
[update] CMAKE_INSTALL_PREFIX: This is the install path. By default, it is set to the installation path that is most typical on the operating system ("/usr/local" for Linux and macOS, "C:\Program Files" for Windows, etc.). When building frameworks on macOS, you may want to change the value to "/Library/Frameworks".
[remove] CMAKE_INSTALL_FRAMEWORK_PREFIX
[add] SFML_DEPENDENCIES_INSTALL_PREFIX: This is the path where SFML's dependencies like Freetype and OpenAL are installed. By default it is the same as CMAKE_INSTALL_PREFIX, except on macOS where it defaults to "/Library/Frameworks", because dependencies on macOS are provided as frameworks. As stated above for CMAKE_INSTALL_PREFIX, it is not mandatory to install SFML after building it, but it is definitely cleaner to do so.
[add] SFML_MISC_INSTALL_PREFIX: This is the path where SFML examples, documentation, license and readme files are installed. On Windows it defaults to CMAKE_INSTALL_PREFIX, and to CMAKE_INSTALL_PREFIX/share/SFML on FreeBSD, Linux and macOS.
Also in https://www.sfml-dev.org/tutorials/2.4/compile-with-cmake.php I found many usages of the name "Mac OS X" which should be replaced with "macOS". "Mac OS X" isn't used for naming the desktop OS from Apple since 2016.
Tried it on Linux (Xubuntu 17.10).
CMake 3.10.2
Works fine for both static and dynamic linking, but I had to do this in my CMakeLists.txt:
cmake_policy(SET CMP0057 NEW)
Otherwise, the script failed:
if given arguments:
"window" "IN_LIST" "SFML_FIND_COMPONENTS"
Unknown arguments specified
Anyways, that's a very handy way of using SFML with CMake, so I'm looking forward to seeing it in SFML's master! It's very useful and makes everything neater.
Thanks a lot for testing and making me notice that it doesn't work with CMake < 3.3 (IN_LIST syntax). Will fix the installed config so that it works with 3.0.2 as the rest of SFML's CMake files.
SFMLBuildMaster: Build this please.
Had a look at SFML tutorials for building SFML with CMake and here are the changes I suggest in the table containing CMake variables descriptions
Thanks for your suggestions, I'll work on a quick PR for the tutorial in order to make this PR mergeable for the next release. :)
PS: my review accounts only for Mac. @eliasdaler can you confirm it works nicely now on Linux too? I leave it to others to double check the behaviour for the other platforms.
It works perfectly well on Linux now. :)
I finally got around to test this on a project of mine. I tried it out on MacOS for dynamic linking and it worked like a charm 👍
I've started SFML/SFML-Website#107 with the changes mentioned above.
Note that now there are some conflicts here... :/
Built with VS 2017 on Windows 10. Built (shared and static libs) without issues, and ran examples without issues!
@dabbertorres Just to make sure, you tested with your own project built against SFML with the find_package(SFML) stuff, right?
@mantognini Can you relaunch build? I fixed conflicts.
@MarioLiebisch @JonnyPtn @mantognini Is it ok for you in terms of code review? How many reviewers to we need to approve the changes?
I haven't been involved in this enough to make a fully informed decision, but I would say merge merge merge!
Is it ok for Android and iOS? Has it been tested with both makefile-based build systems and multi-configuration IDEs?
Haven't had a chance to test Windows yet, but Android projects (so far) shouldn't use CMake anyway. SFML is installed in the NDK's default directory for third party libraries, so you only have to reference them, no lookup required.
on iOS it doesn't find the openAL headers, but I'm struggling to establish why. OPENAL_INCLUDE_DIR seems correct and is set to the right folder...
SFMLBuildMaster: Build this please.
Is it ok for you in terms of code review?
I think it is for me. I went through the thing a few times as you added more commits. I can't pretend to understand everything without having to spend more time on it, but it seems to work (with the exception of iOS) and nothing in the code really made me jump. If you guys could find out why it fails for iOS that'd be great; otherwise, as far I'm concerned, it's ready to be merged.
How many reviewers do we need to approve the changes?
The more the better, as always, but I think we can move forward with this and adapt it if/when we get feedback (probably after official release).
Is it ok for Android and iOS?
Did FindSFML.cmake support Android and iOS? I supposed that it didn't according to
# detect the OS
if(${CMAKE_SYSTEM_NAME} MATCHES "Windows")
set(FIND_SFML_OS_WINDOWS 1)
elseif(${CMAKE_SYSTEM_NAME} MATCHES "Linux")
set(FIND_SFML_OS_LINUX 1)
elseif(${CMAKE_SYSTEM_NAME} MATCHES "FreeBSD")
set(FIND_SFML_OS_FREEBSD 1)
elseif(${CMAKE_SYSTEM_NAME} MATCHES "Darwin")
set(FIND_SFML_OS_MACOSX 1)
endif()
in the original FinSFML.cmake, so I didn't do anything specific on these platforms. And @MarioLiebisch's answer confirms this at least for Android.
Has it been tested with both makefile-based build systems and multi-configuration IDEs?
Yup. At least it does work nicely on macOS with Xcode generator and Unix Makefiles generator for the client project. I expect it to work the same on Windows because the CMake code isn't specific to any platform on this matter.
on iOS it doesn't find the openAL headers, but I'm struggling to establish why. OPENAL_INCLUDE_DIR seems correct and is set to the right folder...
Ok will check.
@JonnyPtn It's fixed. OpenAL was correctly found but OPENAL_INCLUDE_DIR wasn't used on iOS, because I previously relied on the "-framework OpenAL" flag.
@mantognini Can you relaunch the build?
I have tested it using CMake inside QtCreator.
Having some issues on iOS when trying to find SFML from an app, mainly with the dependencies.
I attempted a fix here which you might want to test. It compiles successfully even though It doesn't search for openGL (or openGLES). The app does then crash immediately, but again I'm not sure if that's a problem with the config or with the way I'm using it, or just a completely different bug.
I've also tested macOS static and debug libs which both worked excellently.
SFMLBuildMaster: Build this please.
@JonnyPtn Fix me if I’m wrong but original FindSFML.cmake didn’t support iOS too, so adding it now is out of scope I think.
Sure, I don’t want to hold this up for other platforms, I can do a PR for these later
@LaurentGomila @mantognini @eXpl0it3r Is there anything left to do apart from squashing commits and merge?
For me, it's ready! Feel free to squash commits in a way you find pertinent.
Let's merge it and release it in SFML 2.5. And then let's fix it in 2.5.1 😁
@mantognini @eXpl0it3r @LaurentGomila Up to you now!
Perfect!
SFMLBuildMaster: Build this please.
Just for notice: it looks ready to be merged.
Testing this on Windows with Visual Studio 2017 x64 project file generator, CMake 3.11.0-rc4 and Doxygen 1.8.10.
SFML_DEPENDENCIES_INSTALL_PREFIX isn't being aligned with the CMAKE_INSTALL_PREFIX.
It's nice that there's the option to install it to different places, but having to always adjust both prefixes is a bit cumbersome and unintuitive. My expectation would be that when I change CMAKE_INSTALL_PREFIX it will also change SFML_DEPENDENCIES_INSTALL_PREFIX but not the other way around.
The documentation build fails, even though everything went fine - it even installs the documentation, but for some reason it's reports it as failed.
1>CUSTOMBUILD : error : failed to run html help compiler on index.hhp
[...]
1>Running html help compiler...
1>lookup cache used 3614/65536 hits=22719 misses=3879
1>finished...
1>Done.
1>Building Custom Rule D:/Dev/SFML/doc/CMakeLists.txt
1>CMake does not need to re-run because D:/Dev/SFML/build/doc/CMakeFiles/generate.stamp is up-to-date.
1>Done building project "doc.vcxproj" -- FAILED.
*.pdb files (debug symbol files) are not being installed next to the libraries, but are installed to <install prefix>\lib\Debug. Additionally an empty Release directory is created (maybe because of the same reason). PDB files should really just be placed next to the libraries.
Right now I'm not sure exactly how the PDB files work. This implementation generates PDB files with the none-debug prefix (no -d) in debug mode and no PDB files in release mode. What I'm not sure is:
Do we need PDB files for every release?
Do we need PDB files only for release binaries since the debug libs contain the debug symbols?
Are the release PDB binaries generated in debug mode?
What's up with the lib\cmake\SFML directory that gets installed? Is this the correct location?
You should rebuild and install for all the configurations that you want to use (debug/release, dynamic/static/frameworks) to the same target installation directory. Then remove FindSFML.cmake from your project to make sure it's not used, and clear your CMake cache.
So how does CMake find SFML at my custom location? Previously, I'd set SFML_ROOT and CMAKE_MODULE_PATH and it would find it, how do I tell CMake now to find SFML? Do I need to place the SFMLConfig files somewhere?
CMakeLists.txt
cmake_minimum_required(VERSION 3.1)
project(SFMLTest)
set(CMAKE_CXX_STANDARD 11)
set(SFML_STATIC_LIBRARIES TRUE)
find_package(SFML 2.4 COMPONENTS graphics window system)
add_executable(SFMLTest main.cpp)
target_link_libraries(SFMLTest ${SFML_LIBRARIES})
So how does CMake find SFML at my custom location
Define a SFML_DIR variable that points to the location of the SFMLConfig.cmake file. This is exactly what CMake says in the error message that results from find_package, by the way 😛
I see. Yeah, it showed that message for years and it never applied to SFML, so I didn't think to try it. But this brings me back to the question from above whether <install-prefix>\lib\cmake\SFML is the correct location for the SFMLConfig.cmake file.
Using SFMLConfig.cmake it can find SFML just fine, unfortunately it doesn't set the include directory. Is this intended? Do I still have to use include_directories() if the headers aren't installed in some compiler toolchain known header location?
But this brings me back to the question from above whether \lib\cmake\SFML is the correct location for the SFMLConfig.cmake file.
Looks ok to me, at least that's what I've seen in other libs (Qt).
Using SFMLConfig.cmake it can find SFML just fine, unfortunately it doesn't set the include directory
The output of find_package(SFML ...) are "sfml-xxx" imported targets, so that's what you should pass to the target_link_libraries call. I don't think ${SFML_LIBRARIES} exists anymore now.
Right, I actually wanted to change it and then forgot.
A bit cumbersome that you have to specify the "library" twice, but otherwise it seems to work now.
@eXpl0it3r What do you mean by specifying the library twice? Here's an example I was using to test this: https://github.com/JonnyPtn/SFML-DOOM/blob/master/CMakeLists.txt#L51
First you specify which module you want to find and then you link each module. So you repeat it twice. Sure it might have a different meaning and all, but it's also a bit annoying, plus a source for people getting it wrong and the many questions that follow it.
find_package(SFML 2.4 COMPONENTS graphics window system)
[...]
target_link_libraries(SFMLTest sfml-graphics sfml-window sfml-system)
So the problem is specifying which components to link as opposed to just linking ${SFML_LIBRARIES}? Personally I prefer explicitly stating the components, and it simplifies having multiple projects in a CMakeLists which require different SFML components
First you specify which module you want to find and then you link each module
Listing the modules in find_package is optional (or it should). Plus, you can only specify sfml-graphics since the others are dependencies and will be automatically linked.
Okay, you can just ask for the graphics module and it will pull in the dependencies, but you seem to have to specify the components, otherwise CMake errors with:
find_package(SFML) called with no component
So here's a minimal CMakeLists.txt that works for me with a custom SFML installation.
cmake_minimum_required(VERSION 3.1)
project(SFMLTest)
set(SFML_DIR "<sfml root prefix>/lib/cmake/SFML")
find_package(SFML 2.4 COMPONENTS graphics)
add_executable(SFMLTest main.cpp)
target_link_libraries(SFMLTest sfml-graphics)
My expectation would be that when I change CMAKE_INSTALL_PREFIX it will also change SFML_DEPENDENCIES_INSTALL_PREFIX but not the other way around.
CMake cache system actually prevents from doing that. Once a cache entry is set, setting it in CMake code won't change it, unless you force updating the cache (but then you discard user choice).
The documentation build fails, even though everything went fine - it even installs the documentation, but for some reason it's reports it as failed.
My guess is that it is related to the "error :" that appears in logs. VS considers this a target failure even if the commands actually exit with status 0. I don't think I touched that part though. Is the error also happening on master branch?
*.pdb files (debug symbol files) are not being installed next to the libraries, but are installed to \lib\Debug. Additionally an empty Release directory is created (maybe because of the same reason). PDB files should really just be placed next to the libraries.
Booh… will check.
Right now I'm not sure exactly how the PDB files work. This implementation generates PDB files with the none-debug prefix (no -d) in debug mode and no PDB files in release mode.
Although it is interesting to get the answer, for this PR I will stick with the current behavior on master branch. The purpose of this PR isn't to change how/when PDB are generated.
What's up with the lib\cmake\SFML directory that gets installed? Is this the correct location?
This is one of the standard locations given by find_package()'s documentation. So I'd expect it to be correct :)
So how does CMake find SFML at my custom location (e.g. D:\Dev\SFML\install)? Previously, I'd set SFML_ROOT and CMAKE_MODULE_PATH and it would find it, how do I tell CMake now to find SFML? Do I need to place the SFMLConfig files somewhere?
As @LaurentGomila answered you need to set SFML_DIR instead, and you don't need to set CMAKE_MODULE_PATH because there isn't any FindSFML.cmake to find.
Listing the modules in find_package is optional (or it should). Plus, you can only specify sfml-graphics since the others are dependencies and will be automatically linked.
Did FindSFML.cmake previously work without any component given?
I followed what is done at https://github.com/SFML/SFML/pull/1335/files#diff-e40774a4a91e51bf8e2a254a86083435L126 which means, unless I missed something, that if no component is given, nothing is searched for by the FindSFML.cmake script.
@eXpl0it3r
I can't reproduce the issue about Doxygen. Here it generates the doc without any error log, and thus target succeeds. I'm using CMake 3.10.0, VS 2017 x64 generator and Doxygen 1.8.14. Didn't test with Doxygen 1.8.10 but if latest version works fine I guess it's ok. Does any of the CI builders generate the documentation? Can you test on your side with Doxygen 1.8.14?
As for where the PDB files go (ie. not next to libraries, and always in PREFIX/lib/Debug even for release builds), I checked and this is the same behavior as current master branch. Thus not going to change that for current PR.
The only difference I noticed about PDB files is that with my PR you now also get a PDB for sfml-main library.
All in all, if @LaurentGomila confirms expected behavior with find_package() and components, I guess current PR is still ok for merge. I'll just do a rebase against master to take into account latest commits.
CMake cache system actually prevents from doing that. Once a cache entry is set, setting it in CMake code won't change it, unless you force updating the cache (but then you discard user choice).
Yes, we musn't (can't) make one entry depend on the other. But what about using some default path if SFML_DEPENDENCIES_INSTALL_PREFIX is not set? Something relative to CMAKE_INSTALL_PREFIX.
set(SFML_DIR "<sfml root prefix>/lib/cmake/SFML")
In case someone takes your minimal example as a reference: the SFML_DIR variable should always be defined as a user variable, not directly in the CMakeLists.txt file 😃
Did FindSFML.cmake previously work without any component given?
Not sure about the current behaviour. But that was just an idea anyway. Do you think it's worth investigating?
CMake cache system actually prevents from doing that. Once a cache entry is set, setting it in CMake code won't change it, unless you force updating the cache (but then you discard user choice).
I thought so. For me it's just a different approach I'll have to take and define the prefix manually before clicking the configuration button. Unfortunately this will be an issue many people will fail to understand and we'll have to keep answering the same question. IMHO on Windows it never makes sense to install the dependencies somewhere else, as such this is really inconvenient for Windows users.
Yes, we musn't (can't) make one entry depend on the other. But what about using some default path if SFML_DEPENDENCIES_INSTALL_PREFIX is not set? Something relative to CMAKE_INSTALL_PREFIX.
By default it will use CMAKE_INSTALL_PREFIX. The issue is, that CMAKE_INSTALL_PREFIX has to be defined before you hit the configuration button in the CMake UI, similar to how you'd specify it in the command line the first time you call it.
Would it be an option to not set the SFML_DEPENDENCIES_INSTALL_PREFIX as long as SFML_DEPENDENCIES_INSTALL_PREFIX wasn't set by the user and instead always use CMAKE_INSTALL_PREFIX? That way CMAKE_INSTALL_PREFIX will always be used, unless the user explicitly specifies SFML_DEPENDENCIES_INSTALL_PREFIX. And since it's needed on macOS you could set it there directly. Maybe that's too many if-elses?
I can't reproduce the issue about Doxygen.
Checked again on master branch, it exists as well with my Doxygen version. The reason I stuck with the version is because newer versions don't generate the nice HTML output we want, but that's not important for this PR. Works fine with the latest Doxygen, so let's forget about this.
As for where the PDB files go (ie. not next to libraries, and always in PREFIX/lib/Debug even for release builds), I checked and this is the same behavior as current master branch. Thus not going to change that for current PR.
Odd, I could swear the behavior was different in the past, but yet, I see the same thing on master branch, so let's align this with another PR.
Did FindSFML.cmake previously work without any component given?
Not sure about the current behaviour. But that was just an idea anyway. Do you think it's worth investigating?
I think it's okay the way it is right now. I just suspect again that people will get it wrong and we'll have to keep answering the same question.
My suggestion would be, something that I kind of wanted for a while, to have a dedicated tutorial not on how to build SFML, but on how to use SFML with CMake. That way we can hopefully answer a lot of questions and have a reference for people.
Also we need to make sure to update the existing CMake tutorial for these changes.
But what about using some default path if SFML_DEPENDENCIES_INSTALL_PREFIX is not set? Something relative to CMAKE_INSTALL_PREFIX.
This is a good point indeed. Dunno why I didn't do that just like for SFML_MISC_INSTALL_PREFIX. It's committed.
Would it be an option to not set the SFML_DEPENDENCIES_INSTALL_PREFIX as long as SFML_DEPENDENCIES_INSTALL_PREFIX wasn't set by the user and instead always use CMAKE_INSTALL_PREFIX?
So with the relative path it's just fixed, not need for all these ifs :)
In case someone takes your minimal example as a reference: the SFML_DIR variable should always be defined as a user variable, not directly in the CMakeLists.txt file 😃
I don't think this is always true. For example if you take a project that depends on SFML and wants to embed all their dependencies in their repo (let's say with Git LFS). Setting this variable in the main project allows having a repo ready to use just with a single clone. So it depends on whether the developer wants to embed its dependencies or let its users provide them.
Not sure about the current behaviour. But that was just an idea anyway. Do you think it's worth investigating?
Just to make sure I tried using current FindSFML.cmake with no component. The result is actually worse than what I thought... with no component, it always consider SFML to be found even when it's not :/
Found SFML .. in SFML_INCLUDE_DIR-NOTFOUND
CMake Error: The following variables are used in this project, but they are set to NOTFOUND.
Please set them or make sure they are set and tested correctly in the CMake files:
SFML_INCLUDE_DIR
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples/Demo
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples/Demo
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples/Demo
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples/MinimalistDemo
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples/MinimalistDemo
used as include directory in directory C:/Users/Ceylo/Documents/sfeMovie/examples/MinimalistDemo
So no regression on that matter. As for investigating about support "no component given" case, I think this should be discussed and eventually changed later. I don't really have an opinion at the moment about whether this is a good idea.
My suggestion would be, something that I kind of wanted for a while, to have a dedicated tutorial not on how to build SFML, but on how to use SFML with CMake. That way we can hopefully answer a lot of questions and have a reference for people.
Currently I'd say that the best tutorial for this is the documentation in SFMLConfig.cmake file. But you have to find first, which is not quite obvious 😄 .
Also we need to make sure to update the existing CMake tutorial for these changes.
Do you mean more than https://github.com/SFML/SFML/pull/1335#issuecomment-370251236 ?
Currently I'd say that the best tutorial for this is the documentation in SFMLConfig.cmake file. But you have to find that file first, which is not quite obvious
Could be more or less a 1:1 copy, I don't like spreading documentation across multiple places, especially since it's not an obvious place to look.
Do you mean more than #1335 (comment) ?
If that covers all, then that's good. 😊
SFMLBuildMaster: Build this please.
Could be more or less a 1:1 copy, I don't like spreading documentation across multiple places, especially since it's not an obvious place to look.
A 1:1 copy would look weird to me. What about giving a link to SFMLConfig.cmake from SFML tutorials page? This way you can also make sure that the tutorial is consistent with what is shipped.
Ideally there would be some sort of Doxygen that generates formatted doc from the comments in CMake files, but I’m not aware of such tool.
By the way builds have finished now.
Merged in fc655f52b91d3e70ee43426dc872723418f2bbee
Thanks for the exhaustive work and the pushing forward, really appreciated! 🎉
Hooorrraaaaaay!!!!!!!
At last… :D
I would like to discuss with you about the PR process. Mainly about the testing part: it took almost 2 months between the day the PR was ready for testing and the day it got merged. This is way too long, even in the opensource world, and honestly if it always take this long I'm not gonna do other PRs. So I'd like to talk and find solutions with you.
Where can we do that? The PR comments don't feel like it's the right place, and on the forum (except maybe in section "SFML development" where I can't open topics) it's also too wide: this is mainly toward SFML development team & contributors.
Yeah, this is a big step forward! I'm really happy for this top-quality contribution of yours! Thanks again.
I understand the feeling, and share your point of view. But the fact is that the team alone cannot handle everything, at least not quickly enough to keep the flow. The only solution I see is asking the community to test things out. Now, in the last few months, several people have joined the effort, and I thank them. Hopefully, this will attract more testers as well. In the meantime, if you have ideas on how to improve the overall process (doc, how-to's, issue handling,...) I feel it would best fit on the forum. I'm pretty sure a moderator will agree to move your thread in the right category. :)
I think that people will be less likely to test if developers either provide detailed instructions on how to perform testing or make a test program if it's possible.
The easier it is for people to test, the more likely they're to do it.
I've opened https://en.sfml-dev.org/forums/index.php?topic=23841.0
|
2025-04-01T06:37:31.188554
| 2021-06-01T08:59:31
|
908070999
|
{
"authors": [
"jhoenicke"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2568",
"repo": "SMT-COMP/smt-comp.github.io",
"url": "https://github.com/SMT-COMP/smt-comp.github.io/pull/17"
}
|
gharchive/pull-request
|
Added participants page for 2021.
Added participants (mostly auto-generated from solver csv file).
Preview is here: http://smtcomp.jochen-hoenicke.de/2021/participants.html
|
2025-04-01T06:37:31.200729
| 2018-06-14T14:05:20
|
332418256
|
{
"authors": [
"Filnor",
"Undo1"
],
"license": "cc0-1.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2569",
"repo": "SOBotics/Redunda",
"url": "https://github.com/SOBotics/Redunda/issues/50"
}
|
gharchive/issue
|
Add an option to change the username
MS has a route that allows you to change your username without the need to ping someone with database access. The URL is https://metasmoke.erwaysoftware.com/users/username.
Redunda should have such a feature too, if a user wants to change their name.
Or that. Just a way to update it would be nice.
Am Do., 14. Juni 2018 um 17:24 Uhr schrieb Jed Fox<EMAIL_ADDRESS>
:
Alternatively, it could be updated from the user’s SE display name upon
login (or on a schedule).
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/SOBotics/Redunda/issues/50#issuecomment-397335573,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AG-cjN-cym8pyr7UoAaacj2KuXWwaieNks5t8oA_gaJpZM4UoBsU
.
It should now trigger an async username update (from SE) on every login. Pretty cheap call, no reason not to send it often.
|
2025-04-01T06:37:31.234234
| 2023-07-12T11:15:16
|
1800749997
|
{
"authors": [
"defagos"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2570",
"repo": "SRGSSR/pillarbox-apple",
"url": "https://github.com/SRGSSR/pillarbox-apple/issues/447"
}
|
gharchive/issue
|
Disable comScore analytics while playing video in the background
As a Mediapulse analyst I don't want videos to be tracked when played in the background.
Context
Mediapulse requires videos not to be tracked when < 50% of the video surface is visible. We can identify 3 matching scenarios:
The video is visible but only < 50% of its surface is actually presented on screen (e.g. covered by other views or played in a mini player bar).
The video is played in PiP, whose overlay can be put away by the user.
The video is played in the background (lock screen or app switching).
Due to the nature of the problem we cannot automatically cover several of these use cases:
< 50% area coverage is difficult to address automatically (we cannot really reliably determine how much of a view is visible).
We cannot know when the PiP overlay is visible or not.
We should therefore:
Deal with the cases that can be addressed automatically (app state transitions).
Document how app developers should manually address the other cases, using the available isTracked flag.
Remark
The isTracked flag is greedy and disables all trackers. If finer-grained control is required we should discuss https://github.com/SRGSSR/pillarbox-documentation/issues/38 as well.
Acceptance criteria
Background video playback is not tracked where automatically feasible.
Limitations and manual implementation requirements are documented.
Tasks
[ ] Disable comScore analytics while the application is in background.
[ ] Document required manual implementation inside an app (disabled for < 50% visibility).
Duplicate of #419. Visibility documentation added to #355. Closed.
|
2025-04-01T06:37:31.256600
| 2023-09-13T20:58:16
|
1895256890
|
{
"authors": [
"guyfleeman",
"joe-spall"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2571",
"repo": "SSL-A-Team/firmware",
"url": "https://github.com/SSL-A-Team/firmware/issues/40"
}
|
gharchive/issue
|
Refactor Communication
[ ] change ambiguous verb/noun names/functions
[ ] confirm battery voltage is sent back to ROS
[ ] confirm torques are sent back to ROS
[ ] IMU values
|
2025-04-01T06:37:31.291362
| 2024-12-18T08:43:36
|
2747172021
|
{
"authors": [
"StellarBot",
"shachar-a"
],
"license": "BSL-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2572",
"repo": "STEllAR-GROUP/hpx",
"url": "https://github.com/STEllAR-GROUP/hpx/pull/6593"
}
|
gharchive/pull-request
|
Fix outdated documentation and missing flags
Fixes #1868
Proposed Changes
Remove/add changed params
Add missing \cond \endcond attributes
Checklist
Not all points below apply to all pull requests.
[ ] I have added a new feature and have added tests to go along with it.
[ ] I have fixed a bug and have added a regression test.
[ ] I have added a test using random numbers; I have made sure it uses a seed, and that random numbers generated are valid inputs for the tests.
Can one of the admins verify this patch?
|
2025-04-01T06:37:31.456337
| 2016-04-08T07:23:41
|
146840195
|
{
"authors": [
"MaximilianMeister",
"jdsn",
"toabctl"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2575",
"repo": "SUSE-Cloud/automation",
"url": "https://github.com/SUSE-Cloud/automation/pull/944"
}
|
gharchive/pull-request
|
set testr with update alternatives
in an upgrade from 5to6 somehow testr points to /usr/bin/testr-2.6
which soes not exist. there is /usr/bin/testr-2.7 instead so we
need to set the right path with update-alternatives
17:10:42 + ./run_tempest.sh -N -t -s
17:10:42 + tee tempest.log
17:10:42 ./run_tempest.sh: line 90: testr: command not found
17:10:42 ./run_tempest.sh: line 107: testr: command not found
17:10:42 ./run_tempest.sh: line 107: subunit-2to1: command not found
not sure if that is the best solution, maybe we should run it only under certain conditions like "in an upgrade" or sth. does anyone have a better solution here, or can tell me the conditions under which we have to run this code?
hm. wasn't there a patch from @vuntz for that problem already?
hm. wasn't there a patch from @vuntz for that problem already?
i don't know. btw we have to do the same for subunit-2to1
This fix should go to somewhere else. Fixing it only for mkcloud runs seems odd. It is a quick workaround for sure, but it should be commented that this is a temporary workaround (with the corresponding bug number) that should be reverted once its fixed.
If we would keep it this PR hides the real issue from our ci runs.
This fix should go to somewhere else.
I guess that would be python-testrepository
or python-os-testr in D:C:6 in this case?
I just discovered https://github.com/crowbar/crowbar-core/pull/389 which could fix the issue
In that case lets close this PR
|
2025-04-01T06:37:31.471170
| 2018-04-11T08:38:19
|
313225129
|
{
"authors": [
"anandr781",
"felixPG",
"mssola",
"seanhoughton"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2576",
"repo": "SUSE/Portus",
"url": "https://github.com/SUSE/Portus/issues/1777"
}
|
gharchive/issue
|
Portus PORTUS_BACKGROUND=true and database empty
Hi All:
I enabled the background process with env PORTUS_BACKGROUND=true.
The problem is when the database are empty and NOT READY because PORTUS on bootstrap does not create the database yet, so, error:
[Database] Not ready yet. Waiting...
/srv/Portus/lib/portus/db.rb:40:in `wait_until': Timeout reached for 'ready' status
[Database] Timeout reached, exiting with error. Check the logs...
Database are not ready because PORTUS launch background proccess before the database creation on bootstrap.
How I could solve this?
Regards
You need to run two containers, one without the worker setting and one with it set to true. If you just run the worker you won't actually have a Portus server running to init the DB.
This issue sadly slipped through...
You need to run two containers, one without the worker setting and one with it set to true. If you just run the worker you won't actually have a Portus server running to init the DB.
Exactly :+1:
I'll close this issue now, but feel free to leave more comments if you have further doubts on this issue.
What flag should I use to docker run Portus in server mode so that it will bootstrap the database.
@anandr781 no flag, actually. See this examples
|
2025-04-01T06:37:31.483412
| 2019-05-28T10:07:51
|
449180963
|
{
"authors": [
"Falco20019",
"SuperSandro2000"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2577",
"repo": "SUSE/Portus",
"url": "https://github.com/SUSE/Portus/issues/2203"
}
|
gharchive/issue
|
"Invalid filter syntax" on LDAP team sync
Description
Adding new users to teams through LDAP fails with the message [ldap] Connection error: Invalid filter syntax..
Steps to reproduce
Add a team "LW-LI" (matching the LDAP group CN=LW-LI,OU=Universal,OU=Group,OU=NW,OU=DE,OU=Production,DC=my-company,DC=com)
Add a new user "debk0l" through first login of LDAP account
Seeing that the user can login and works
Watching the background taks logs, waiting for team addition
Expected behavior: The use should be added to the team
Actual behavior: [ldap] Connection error: Invalid filter syntax.
Logs don't show any valuable information:
[ldap] Looking up an LDAP group membership for 'debk0l'
User Load (0.4ms) SELECT `users`.*
Assumption
I assume, it's failing here when accessing the distinguished name. In our LDAP, there is no field dn, just distinguishedName. This will result in search.groups_from in a filter of the form (&(cn=*)(member=)) which is of course invalid.
I think, like with the uid, the dn attribute needs to be configurable.
I currently can't build the image myself to try it out, since our IT is using a man-in-the-middle proxy (ZScaler) with a self-signed certificate that parts of the your build chain are not trusting.
Deployment information
Deployment method:
Docker compose, pretty similar to the example.
Configuration:
Running docker-image opensuse/portus:head from today.
ldap:
enabled: true
hostname: "mos1d00001.my-company.com"
port: 636
timeout: 5
encryption:
method: "simple_tls"
base: "OU=Production,DC=my-company,DC=com"
group_base: "OU=Universal,OU=Group,OU=NW,OU=DE,OU=Production,DC=my-company,DC=com"
filter: "(&(objectCategory=person)(memberOf=CN=LW-LI,OU=Universal,OU=Group,OU=NW,OU=DE,OU=Production,DC=my-company,DC=com))"
uid: "sAMAccountName"
authentication:
enabled: true
bind_dn: "cn=ldap-user,ou=service,ou=user,ou=MB,ou=DE,ou=production,DC=my-company,DC=com"
group_sync:
enabled: true
default_role: "contributor"
guess_email:
enabled: true
attr: "userPrincipalName"
Portus version: 2.5.0-dev@a1b9f2ebfeb84680a9dcd5629195e4c52815735c
LDAP samples (relevant excerpt)
ldaps://mos1d00001.my-company.com:636/CN=Kraemer%5C,%20Benjamin,OU=LW-LI,OU=JLS,OU=Department,OU=People,OU=User,OU=MB,OU=DE,OU=Production,DC=my-company,DC=com
Field
Value
objectClass
person
cn
Kraemer, Benjamin
distinguishedName
CN=Kraemer, Benjamin,OU=LW-LI,OU=JLS,OU=Department,OU=People,OU=User,OU=MB,OU=DE,OU=Production,DC=my-company,DC=com
memberOf
CN=LW-LI,OU=Universal,OU=Group,OU=NW,OU=DE,OU=Production,DC=my-company,DC=com
sAMAccountName
dejhbk0l
userPrincipalName
<EMAIL_ADDRESS>
ldaps://mos1d00001.my-company.com:636/LW-LI,OU=Universal,OU=Group,OU=NW,OU=DE,OU=Production,DC=my-company,DC=com
Field
Value
objectClass
group
member
CN=Kraemer, Benjamin,OU=LW-LI,OU=JLS,OU=Department,OU=People,OU=User,OU=MB,OU=DE,OU=Production,DC=my-company,DC=com
Still waiting for any attention by the team.
Still a problem
bump
/unstale
|
2025-04-01T06:37:31.540611
| 2021-03-19T21:39:37
|
836377156
|
{
"authors": [
"AhmadShakerASH",
"Sadmansamee"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2578",
"repo": "SadaqaWorks/Quran-Flutter",
"url": "https://github.com/SadaqaWorks/Quran-Flutter/issues/19"
}
|
gharchive/issue
|
ERROR: [TAG] Failed to resolve variable '${junit.version}'
Launching lib/main.dart on Android SDK built for x86 in debug mode...
ERROR: [TAG] Failed to resolve variable '${junit.version}'
ERROR: [TAG] Failed to resolve variable '${animal.sniffer.version}'
@AhmadShakerASH Merge with latest change
|
2025-04-01T06:37:31.544365
| 2023-07-06T15:45:28
|
1791810113
|
{
"authors": [
"lars-reimann",
"zzril"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2579",
"repo": "Safe-DS/Stdlib",
"url": "https://github.com/Safe-DS/Stdlib/pull/427"
}
|
gharchive/pull-request
|
docs: Better documentation for developers
Closes #375.
Summary of Changes
Improved developer documentation about tests
Added guidelines about copying objects (rather than modifying them in-place)
Added code-style guidelines
Added code review guidelines
The code style section got somewhat small. Turns out we don't really have a common code style except for "do what the linter says".
Thought about alsl adding an FAQ for developers, but I came up with only like 3 questions, that didn't see emough, so I dropped the idea.
Don't have any other ideas, so let me mark this as ready for review.
:tada: This PR is included in version 0.15.0 :tada:
The release is available on:
v0.15.0
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:37:31.579272
| 2022-03-31T14:18:09
|
1188157841
|
{
"authors": [
"GiaJordan",
"milen-sage",
"rrchai",
"ychae"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2580",
"repo": "Sage-Bionetworks/data_curator",
"url": "https://github.com/Sage-Bionetworks/data_curator/issues/309"
}
|
gharchive/issue
|
Refactor the validation functions in DCA to incorporate changes from schematic
Incorporate backend updates from https://github.com/Sage-Bionetworks/schematic/pull/614
#315
[x] Ensure the updated validateModelManifest() works as expected
[x] Refactor related validation function in DCA to address the changes from schematic: e.g. validateModelManifest() now return two things: "error" and "warning". I think we just need first element (error) of output list in DCA for now
#322
[x] Test cross-manifest rules on HTAN data: https://github.com/ncihtan/data-models/pull/28
[ ] update UI error messages for cross-manifest errors
related to https://github.com/Sage-Bionetworks/schematic/issues/683:
One thing to note for DCA:
Each row represents each invalid value with error in the UI error table, so "matchExactlyOne" might not be fit for the error table since entire column/attribute is invalid. Thus, we might consider "matchExactlyOne" errors as a separated error type, I could just highlight the invalid source attribute and throw its errors.
Editting:
I have run some tests on the cross-manifest testing (MatchAtLeast) using the latest schematic:
testing manifest
testing data model
there is only one cross-manifest rule:
"matchAtLeastOne scRNA-seqLevel1.HTANParentBiospecimenID Biospecimen.HTANBiospecimenID"
Here is my benchmarking result for running time (by running validation 5 times with ^ testing files using hypefine:
Using the schematic before 700 is merged:
hyperfine --warmup 1 --runs 5 --ignore-failure 'schematic model -c schematic_config.yml validate -mp ~/Downloads/htan/test-cross-link.csv -dt ScRNA-seqLevel1'
Using the latest schematic:
hyperfine --warmup 1 --runs 5 --ignore-failure 'schematic model -c schematic_config.yml validate -mp ~/Downloads/htan/test-cross-link.csv -dt ScRNA-seqLevel1'
This is my downloading speed:
Testing download speed................................................................................
Download: 8.13 Mbit/s
Testing upload speed......................................................................................................
Upload: 11.04 Mbit/s
@rrchai this is exactly the kind of benchmarking we need more of! My take aways from it:
we'll need project scoping - and repeat the same benchmark with it
very likely, we'll still need filtering based on component annotation - that would require adding the component annotation to HTAN dataset manifests retrospectively
@GiaJordan when you're back, what are your thoughts on project scoping readiness for testing?
@ychae what do you think about who'd be best positioned to work on HTAN's retroactive component annotations of manifests?
@rrchai
In the development of this I noticed that when there were multiple manifests for a dataset, it would just take the first one and this wouldn't always be desired, so I added logic there for selection. It shouldn't affect datasets with only one manifest though.
The manifests currently on synapse will also need to have component annotations added either manually or by re-uploading for the expedited gathering to work, otherwise it will revert back to the old behavior.
@milen-sage
Project scoping is ready for testing now; I'm planning on merging #701 on Monday
@ychae what do you think about who'd be best positioned to work on HTAN's retroactive component annotations of manifests?
@milen-sage I think the data liaisons might be able to help with this task -- ideally each liaison would add the component annotation for their centers but I'll need to float this idea by Ashley to see how keen the liaisons would be to take this on.
@GiaJordan @milen-sage By validating on only HTAN center A using the latest schematic (merged 701, the speed is amazing now.
validation results
error: Manifest syn27331282 does not contain the value HTA111111_121324 from row 2 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn27331282 does not contain the value sdsadasd from row 3 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn27331282 does not contain the value HTA3_8001_1002 from row 4 of the attribute HTANParentBiospecimenID in the source manifest.
Since there is no flag for project_scope using cli, I replace default value of project_scope with centerA's synId for testing.
hyperfine --warmup 1 --runs 5 'schematic model -c schematic_config.yml validate -mp ~/Downloads/htan/test-cross-link.csv -dt ScRNA-seqLevel1'
Time (mean ± σ): 18.844 s ± 5.422 s [User: 4.838 s, System: 0.672 s]
Range (min … max): 11.972 s … 25.155 s 5 runs
@rrchai I'm not sure why there's different behavior for using the flag and chaning the default value. Does the master_fileview for both contain the CenterA project?
The warning for the unmet access restrictions is expected, though I suppose displaying a message that a censored version of the manifest will be downloaded instead would be helpful.
Very happy to hear about the times!
@GiaJordan ah, you are right. I forgot to use the HTAN's fileview. I will re-test it.
I forgot to use the HTAN's fileview. I will re-test it.
This might explain the speed up too. HTAN's fileview is larger.
Test HTAN fileview
validation results
error: Manifest syn27116926 does not contain the value HTA111111_121324 from row 2 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn27116926 does not contain the value sdsadasd from row 3 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn27116926 does not contain the value HTA3_8001_1002 from row 4 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn30560343 does not contain the value HTA111111_121324 from row 2 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn30560343 does not contain the value sdsadasd from row 3 of the attribute HTANParentBiospecimenID in the source manifest.
error: Manifest syn30560343 does not contain the value HTA3_8001_1002 from row 4 of the attribute HTANParentBiospecimenID in the source manifest.
hyperfine --warmup 1 --runs 5 'schematic model -c schematic_config.yml validate -mp ~/Downloads/htan/test-cross-link.csv -dt ScRNA-seqLevel1 -ps syn20977135'
Time (mean ± σ): 120.861 s ± 66.810 s [User: 16.036 s, System: 1.864 s]
Range (min … max): 54.659 s … 215.867 s 5 runs
The speed is definitely faster by using project scope. I think 2 min averagely for me is okay to use and test on one/two cross-manifests validation. I notice the most time it takes on my end is querying the fileview table?
Downloading [####################]100.00% 74.9MB/74.9MB (538.7kB/s) SYNAPSE_TABLE_QUERY_95623552.csv.synapse_download_95623552 Done...
Since the DCA has already stored the synapseStorage object at the beginning. @milen-sage I wonder if it possible to use the synapseStorage object directly as input to avoid downloading it again? I still can proceed to implement it in the DCA without this enhancement.
@GiaJordan I tested the develop-crossM-union-rule branch:
using "matchAtLeastOne Biospecimen.HTANBiospecimenID set":
It looks like it is validating individual values?
warning: Value(s) ['HTA111111_121324'] from row(s) [2] of the attribute htanparentbiospecimenid in the source manifest are missing. Manifest(s) ['syn27116926', 'syn30560343'] are missing the value(s).
Your manifest has been validated successfully. There are no errors in your manifest, and it can be submitted without any modifications.
using "matchAtLeastOne Biospecimen.HTANBiospecimenID value":
warning: Value(s) ['HTA111111_121324'] from row(s) [2] of the attribute htanparentbiospecimenid in the source manifest are missing. Manifest(s) ['syn27116926', 'syn30560343'] are missing the value(s).
Your manifest has been validated successfully. There are no errors in your manifest, and it can be submitted without any modifications.
FYI: currently, the DCA will only extract the error list (4 elements) from validation results and reformat it into a table:
Row
Column
Value
Error
1st element
2rd element
4th element
3rd element (anything start with "not")
Warning could also be used as long as it has the same format of errors. I found the almost all errors from schematic contains "not". To simplify the UI error message, only characters after "not" (including "not") are used. So the error message (3rd element of error output list) prefers to contain "not".
@rrchai yes, let's give reusing the fileview object a try! If this avoids the download time, it would be worth it for sure.
In this case it sounds like there are two items:
integration of the cross-manifest rule in the DCA and the dashboard
fileview object reuse in the dashboard
If the implementation of 2 doesn't affect 1 too much, you can start with 1 and then move on to 2.
On the other hand, if 2 will affect significantly how 1 is implemented, proceeding with 2 and testing it first would make more sense.
My intuition is that 2 shouldn't affect too much 1. What do you think?
@rrchai Thanks for the feedback. It will validate on the set or value level, but with either case it'll display which values are causing the invalidity. And while they've been switched to warnings, the format is still the same.
In response to the query being the limiting step currently, I added the project scope to the table query if one is provided. That should help to speed things up more. This change is still the cross manifest branch, so you can pull the latest version to try.
cc @milen-sage
okay, my internet issue has been fixed and the downloading speed is back to normal:
Testing download speed................................................................................
Download: 170.74 Mbit/s
Testing upload speed......................................................................................................
Upload: 9.92 Mbit/s
I retested the latest develop branch and downloading fileview table is not as slow as before with ^ speed 😄 :
Using the latest develop branch:
hyperfine --warmup 1 --runs 10 'schematic model -c schematic_config.yml validate -mp ~/Downloads/htan/test-cross-link.csv -dt ScRNA-seqLevel1 -ps syn20977135'
Time (mean ± σ): 42.877 s ± 5.390 s [User: 15.350 s, System: 1.220 s]
Range (min … max): 37.302 s … 51.442 s 10 runs
Using develop-crossM-union-rule branch:
hyperfine --warmup 1 --runs 10 'schematic model -c schematic_config.yml validate -mp ~/Downloads/htan/test-cross-link.csv -dt ScRNA-seqLevel1 -ps syn20977135'
Time (mean ± σ): 23.530 s ± 2.253 s [User: 5.753 s, System: 0.740 s]
Range (min … max): 20.351 s … 26.741 s 10 runs
@GiaJordan Thank you for the improvement. The new querying looks like that it reduces time significantly from 40ish to 20ish 🚀 !
My intuition is that 2 shouldn't affect too much 1. What do you think?
Based on the benchmarking results, I think downloading fileview object will affect the overall speed of validation. Since the @GiaJordan has already improve the query process and ~20s for one cross-manifest validation is good to me, I could start to implement the cross-manifest rules in the app now.
In short, by testing with one cross-manifest rule with my internet, the faster your downloading speed is and less projects/data to query/download, the faster the cross-manifest validation will be.
@rrchai Thanks for the benchmarking and updates!
In the farther future, we'll likely experiment with converting the synapse storage object to a sentinel object that can be reused within one validation run. This would expedite the validation further in cases where multiple attributes use the cross manifest validation rule.
Looks like we just need to review a PR and this will be ready to close?
@milen-sage can we then merge #751 as well?
|
2025-04-01T06:37:31.607491
| 2017-08-02T09:25:17
|
247325968
|
{
"authors": [
"SahebRoy92",
"vishaldeshai"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2581",
"repo": "SahebRoy92/SRPopView",
"url": "https://github.com/SahebRoy92/SRPopView/issues/2"
}
|
gharchive/issue
|
How can we add like a Pod?
Hello,
Can we add SRPopView using pod?
thanks in advance.
Vishal
Yes I am working on a swift version also with some enhancement, will update to a pod in a few days
Pod version is available
Ok thanks
I am adding to the trunk pod so it might take a few hours to take effect, so watch out for the update here
sure.
|
2025-04-01T06:37:31.618686
| 2024-04-08T18:32:16
|
2231847952
|
{
"authors": [
"codecov-commenter",
"sfmig"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2582",
"repo": "SainsburyWellcomeCentre/crabs-exploration",
"url": "https://github.com/SainsburyWellcomeCentre/crabs-exploration/pull/151"
}
|
gharchive/pull-request
|
Set persistent workers for validation dataloader
Following Lightning suggestion ⚡
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 31.07%. Comparing base (d0a8066) to head (d62facf).
Additional details and impacted files
@@ Coverage Diff @@
## main #151 +/- ##
=======================================
Coverage 31.07% 31.07%
=======================================
Files 19 19
Lines 1223 1223
=======================================
Hits 380 380
Misses 843 843
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
2025-04-01T06:37:31.619721
| 2023-06-23T03:28:41
|
1770718090
|
{
"authors": [
"SaishoVibes",
"gitdrug"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2583",
"repo": "SaishoVibes/Back-in-Classic-Fabric",
"url": "https://github.com/SaishoVibes/Back-in-Classic-Fabric/issues/2"
}
|
gharchive/issue
|
Update to 1.20x
Update to 1.20x
I will soon, just finalizing changes for 1.19.4 before thinking of dividing my work
|
2025-04-01T06:37:31.643860
| 2023-09-20T23:14:32
|
1905836482
|
{
"authors": [
"npsp-reedestockton"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2584",
"repo": "SalesforceFoundation/NPSP",
"url": "https://github.com/SalesforceFoundation/NPSP/pull/7193"
}
|
gharchive/pull-request
|
ACCT_ViewOverride_CTRL Security Fix
Critical Changes
Changes
Issues Closed
Community Ideas Delivered
Features Intended for Future Release
Features for Elevate Customers
New Metadata
Deleted Metadata
W-13641227
|
2025-04-01T06:37:31.654190
| 2020-04-30T20:19:01
|
610359262
|
{
"authors": [
"I1mran",
"SamSamskies"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2585",
"repo": "SamSamskies/react-map-gl-geocoder",
"url": "https://github.com/SamSamskies/react-map-gl-geocoder/issues/63"
}
|
gharchive/issue
|
Type issue while onHover and onClick triggers.
Hello, I am facing the issue of gl geocoder disable after adding to clicks on MapGL instance, on hover, and onClick
Let me know how I can resolve this.
Hi @I1mran, is it possible for you to provide a Code Sandbox (https://codesandbox.io/) that demonstrates this issue?
I don't have code right now for sharing.
|
2025-04-01T06:37:31.658599
| 2022-04-26T14:55:45
|
1216083483
|
{
"authors": [
"SamVerschueren",
"benasher44"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2586",
"repo": "SamVerschueren/listr-update-renderer",
"url": "https://github.com/SamVerschueren/listr-update-renderer/issues/31"
}
|
gharchive/issue
|
Open to new maintainers?
Hi @SamVerschueren! Would you be open to new maintainers? It seems like this dependency will forever be a part of our dep tree, so it'd be great to see its dependencies updated (keeps getting flagged by vulnerability scanners, even though we only use it in local, non-production contexts).
I'm totally down for new maintainers as I don't have much time myself anymore and have different priorities.
Great! I'll volunteer. How do you want to move forward?
|
2025-04-01T06:37:31.664753
| 2016-04-07T17:12:08
|
146687236
|
{
"authors": [
"FluenesHerre"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2587",
"repo": "Samfundet/Samfundet",
"url": "https://github.com/Samfundet/Samfundet/pull/58"
}
|
gharchive/pull-request
|
add csrf protection for resources available to everyone
Reservations made through sulten were previously vulnerable to csrf attacks.
This manually implements the default behaviour in rails 4.
https://github.com/rails/rails/blob/master/actionpack/lib/action_controller/metal/request_forgery_protection.rb#L194
This manually implements the behaviour in rails 4
|
2025-04-01T06:37:31.670953
| 2022-03-27T23:17:23
|
1182696960
|
{
"authors": [
"YongseopKim",
"hyunsik-yoon",
"jyoungyun"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2588",
"repo": "Samsung/ONE-vscode",
"url": "https://github.com/Samsung/ONE-vscode/issues/414"
}
|
gharchive/issue
|
Implementation of Installation of Compiling Env & Execution Env
Let's work on compiling env & execution env.
In this issue, only "local installation of debian package" will be handled.
[ ] designing UI for compiling env
[ ] designing UI for execution env
[ ] implementation of compiling env
[ ] implementation of execution env
[ ] implementation of auto detect (finding if any existing local env already exists)
find out the following
[ ] are they working
[ ] implementation of uninstall
[ ] implementation of install
[ ] implementation of install over existing one
Out of scope
remote installation
docker installation
supporting various backends
make it for existing backend first
consider various backends for easier refactoring later
but find clues for various backend support
/cc @jyoungyun
compiling env / execution env are very important concepts, I think. And I'm interested in them. Could you share your concrete concepts for impl them?
Related issue: #306
This issue was not done.
I think it's better to close this and reschedule after discussing with @jyoungyun.
|
2025-04-01T06:37:31.675789
| 2022-04-14T01:46:10
|
1203945042
|
{
"authors": [
"YongseopKim",
"dayo09",
"jyoungyun"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2589",
"repo": "Samsung/ONE-vscode",
"url": "https://github.com/Samsung/ONE-vscode/issues/476"
}
|
gharchive/issue
|
Let's define and schdule for ONE-context-explorer and ONE-context-explorer-detail views
[ ] Define this task: ~4/14(Thr)
[ ] Divide/Separate this task and schedule: ~4/15(Fri)
Maybe it's like the below things?
UI Tasks: task0(M1), task1(M2), ...
Internal Impl Tasks for UI: task0(M1), task1(M2), ...
[ ] Make issues and assign on our ONE-project
/cc @Samsung/one-vscode
(Suggestion)
About the naming for those two views we discussed,
how about 'explorer view' and 'detail view' ?
Things to concern
If we are going to regard all tflite, onnx, ... model files as a 'base model' and npu model file as 'target model', our extension will be not very extensible for other model compiler usages.
For example, a model compiler who basically want to do onnx to tflite jobs, it would not work properly.
Therefore, I suggest the concept of 'base model'. Let's make user to select a 'base model' extension and let the 'explorer' only handles them and the 'detail view' to collect other files.
We may start by blocking other options than 'tflite' to make our jobs rather easy.
First of all, we need to discuss whether we support to convert onnx to tflite in our ONE-vscode. How about talking on next Monday about this subject?
First of all, we need to discuss whether we support to convert onnx to tflite in our ONE-vscode.
My suggestion was based on its extensibility, rather than the actual support. I heard that our extension kinda meant to support the other compilers.
I agree. I am closing this on behalf of other members.
|
2025-04-01T06:37:31.677253
| 2023-09-20T10:35:19
|
1904697846
|
{
"authors": [
"jyoungyun"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2590",
"repo": "Samsung/ONE-vscode",
"url": "https://github.com/Samsung/ONE-vscode/pull/1662"
}
|
gharchive/pull-request
|
[Execute] Introduce DeviceViewNode by device type
This commit introduces DeviceViewNode by device type.
ONE-vscode-DCO-1.0-Signed-off-by: Jiyoung Yun<EMAIL_ADDRESS>
This PR works normally only when #1660 and #1661 PRs are merged together.
|
2025-04-01T06:37:31.732284
| 2022-02-11T17:26:27
|
1132889199
|
{
"authors": [
"Aniket-508",
"SaraswatGit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2591",
"repo": "SaraswatGit/PlanZap",
"url": "https://github.com/SaraswatGit/PlanZap/issues/44"
}
|
gharchive/issue
|
On Deleting a task confetti should not pop up
You must have observed a new confetti feature on our website. The thing is this appears even when a user deletes a task (clicking on a cross button) . To resolve this call the API call when user wants to delete and API call when user has completed the task should come inside different functions .
@SaraswatGit I want to work on this issue under JWoC.
@Aniket-508 I have assigned the issue to you, Best Wishes, for any problem reach out to me on discord .
|
2025-04-01T06:37:31.733489
| 2019-10-09T16:29:24
|
504760771
|
{
"authors": [
"SarathSantoshDamaraju",
"mrsupiri"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2592",
"repo": "SarathSantoshDamaraju/lazyGit",
"url": "https://github.com/SarathSantoshDamaraju/lazyGit/pull/20"
}
|
gharchive/pull-request
|
Added gyolo
when you type gyolo it will pull random commit message from the internet and commit changes with that message
@mrsupiri If you are interested, you can add something like Semantic or Conventional commit style.
|
2025-04-01T06:37:31.736766
| 2021-10-09T04:52:52
|
1021601405
|
{
"authors": [
"RAshid602",
"SarthakKeshari",
"sRahul-00"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2593",
"repo": "SarthakKeshari/Java-Questions-and-Solutions",
"url": "https://github.com/SarthakKeshari/Java-Questions-and-Solutions/issues/364"
}
|
gharchive/issue
|
MergeSort Without Taking Extra Space
Enter your question -
Merge Without Extra Space
Enter link to the question(if question belongs to any online platform) -
Merge Without Extra Space
Tags for the question(eg - Array, Basic, Stack, etc.) -
Array,Sorting
@RAshid602,
Kindly add your solution to "GeeksForGeeks" folder.
Deadline - 10/10/2021
Can I work on this?
@sRahul-00,
Its great to see your interest to solve the question. But since the deadline ain't over yet. You may refer to CONTRIBUTING.md for getting futher details and guidelines on how to contribute to this repo.
|
2025-04-01T06:37:31.760441
| 2021-09-30T16:02:37
|
1012357489
|
{
"authors": [
"benhammondmusic",
"jgonzalezmsm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2594",
"repo": "SatcherInstitute/health-equity-tracker",
"url": "https://github.com/SatcherInstitute/health-equity-tracker/issues/1184"
}
|
gharchive/issue
|
Need to standardize grammar, labels, language style
[ ] replace all straight/dumb quotation marks "" with fancy/smart quotes “” when needed as actual quotations. Terms and categories should be indicated using bold tags
[ ] use of HET vs Health Equity Tracker
[ ] need to define what is in the term itself, vs which is a logical combination multiple defined groups. &, and, or. When to use commas? When to use oxford commas? Josh noted CDC uses /
hold on this until larger website redesign and copy updates with Mahia
|
2025-04-01T06:37:31.767395
| 2023-06-09T18:15:37
|
1750311753
|
{
"authors": [
"XileHorizon",
"stavares843"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2595",
"repo": "Satellite-im/Uplink",
"url": "https://github.com/Satellite-im/Uplink/issues/874"
}
|
gharchive/issue
|
chore(chats): reactions position
STR:
go to chat
send message
right click
suggestion - when we add a reaction goes below the message but the reaction list is above the message, shouldn't both be either both in top or both in below?
macOS, m1
I think we should eventually use the emoji picker for the reactions or at least have options be in the context menu when you right click rather than be an additional action. Maybe we have a list of common reaction emoji and an extra button to view them all?
|
2025-04-01T06:37:31.774860
| 2024-07-02T14:50:23
|
2386474972
|
{
"authors": [
"lgmarchi",
"luisecm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2596",
"repo": "Satellite-im/UplinkWeb",
"url": "https://github.com/Satellite-im/UplinkWeb/pull/187"
}
|
gharchive/pull-request
|
update(Friend): Create toast message for friend request and delete modals
What this PR does 📖
1. Create toast messages on friend requests action, and delete modals
2. After send a friend request, if it is successful or not, it delete did from input.
3. Fix null problem on current dev with a if
4. Update rules for input on username, and add rules for status message
Which issue(s) this PR fixes 🔨
Resolve #181
Special notes for reviewers 🗒️
Additional comments 🎤
Tested the following
On Create Account screen:
Pin input empty - passed
Pin input less than 4 characters - passed
Pin input with spaces - passed
Pin input with non-alphanumeric chars - passed
Pin input with more than 32 chars - passed
Status input with more than 128 chars - passed
Noticed that create user button is always enabled even though there is an error input value (we might need a new ticket for this)
On Friend Requests
Modal is gone and now is toast notification displayed after sending friend request (sucess and to yourself) - Passed
Friend DID Input field is cleared out after sending friend request - Passed
Null Problem - Passed (app compiled correctly on branch)
Settings Profile
Same scenarios above passed
Noticed that Save/Cancel buttons are still displayed when there is a wrong input value. If user clicks on Save nothing happens
One more weird scenario found:
On settings profile, enter a new username - good value
Enter now a status exceeding 128 chars
Click on Save button
Profile updated toast notification will show
|
2025-04-01T06:37:31.821877
| 2018-12-28T11:36:23
|
394617296
|
{
"authors": [
"SaturnTeam",
"Zacknero",
"agiratech-kumaresanj",
"sancaruso"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2597",
"repo": "SaturnTeam/saturn-datepicker",
"url": "https://github.com/SaturnTeam/saturn-datepicker/issues/58"
}
|
gharchive/issue
|
Range: when resetting the form value, the values of begin and end are not reseted.
When in single selection mode, if we reset the form value driving the selected date, then the value of selectedValue is correctly reseted in the SatCalendarBody component.
But in range mode, this is not the case: the values of begin and end in SatCalendarBody remain unchanged.
I think this is a bug.
How do you reset value?
I tried with both ngModel and formControl (respectively resetting the form by setting the model to null and with form.reset()) and I have the same problem.
But indeed, I made more tests and the problem happens only when I reset the form immediately after the dateChange event.
Wrapping the reset in a setTimeout(() => {...}, 0) solves the problem.
Moreover, this only happens when the closeAfterSelection option is false.
So perhaps it could be solved with a markForCheck somewhere but I don't know where.
Here is a sample to reproduce the bug:
import { Component } from '@angular/core';
import {FormBuilder, FormGroup} from "@angular/forms";
@Component({
selector: 'app-root',
template: `
<form [formGroup]="form">
<mat-form-field>
<input matInput
placeholder="Choose a date"
[satDatepicker]="picker"
(dateChange)="dateChange($event)"
formControlName="date">
<sat-datepicker #picker [rangeMode]="true"
[closeAfterSelection]="false">
</sat-datepicker>
<sat-datepicker-toggle matSuffix [for]="picker"></sat-datepicker-toggle>
</mat-form-field>
</form>
`,
styleUrls: ['./app.component.scss']
})
export class AppComponent {
form: FormGroup;
constructor(fb: FormBuilder) {
this.form = fb.group({
date: [{ begin: new Date(2018, 7, 5), end: new Date(2018, 7, 25) }]
});
}
dateChange(event) {
// Do some stuff
this.form.reset();
}
}
Hope, maybe it will help you, To reset i have tried like this, passing | assigning the [beginDate] and [endDate] value as null on sat-calendar.
Hope, maybe it will help you, To reset i have tried like this, passing | assigning the [beginDate] and [endDate] value as null on sat-calendar.
Don't work this!
I tried in this mode but don't work same:
@ViewChild(SatCalendar, {static: true}) calendar: SatCalendar<Date> this.calendar.beginDateSelected = false; this.calendar.beginDate = null; this.calendar.maxDate = null; this.calendar.selected = null; this.calendar.startAt = null; ;
So it's a BUG of library and must to be fix or insert new solution
Hope, maybe it will help you, To reset i have tried like this, passing | assigning the [beginDate] and [endDate] value as null on sat-calendar.
Don't work this!
I tried in this mode but don't work same:
@ViewChild(SatCalendar, {static: true}) calendar: SatCalendar<Date> this.calendar.beginDateSelected = false; this.calendar.beginDate = null; this.calendar.maxDate = null; this.calendar.selected = null; this.calendar.startAt = null; ;
So it's a BUG of library and must to be fix or insert new solution
You have to pass the values(beginDate and endDate) like below in sat-calender element
<sat-calendar
[beginDate]="calendar.beginDate"
[endDate] = "calendar.maxDate">
The my setup is:
The unser click one day and the week is selected (from Mondato to Sunday). When click other day of week this reset week and re-click day for select week.
The bug is when the user (after clicked the day and selected week) re-click the same day or day in week selected the saturn fire (beginDateSelectedChange) and this work badly.
Example StackBlitz
So i would reset range or send to set null beginDate
Hope, maybe it will help you, To reset i have tried like this, passing | assigning the [beginDate] and [endDate] value as null on sat-calendar.
Don't work this!
I tried in this mode but don't work same:
@ViewChild(SatCalendar, {static: true}) calendar: SatCalendar<Date> this.calendar.beginDateSelected = false; this.calendar.beginDate = null; this.calendar.maxDate = null; this.calendar.selected = null; this.calendar.startAt = null; ;
So it's a BUG of library and must to be fix or insert new solution
You have to pass the values(beginDate and endDate) like below in sat-calender element
<sat-calendar
[beginDate]="calendar.beginDate"
[endDate] = "calendar.maxDate">
The my setup is:
The user click one day and the week is selected (from Monday to Sunday). When click other day of week this reset week and re-click day for select week.
The bug is when the user (after clicked the day and selected week) re-click the same day or day in week selected, the saturn fire (beginDateSelectedChange) and this work badly.
Example StackBlitz
So i would reset range of days or send to set beginDate = null
Hope, maybe it will help you, To reset i have tried like this, passing | assigning the [beginDate] and [endDate] value as null on sat-calendar.
Don't work this!
I tried in this mode but don't work same:
@ViewChild(SatCalendar, {static: true}) calendar: SatCalendar<Date> this.calendar.beginDateSelected = false; this.calendar.beginDate = null; this.calendar.maxDate = null; this.calendar.selected = null; this.calendar.startAt = null; ;
So it's a BUG of library and must to be fix or insert new solution
You have to pass the values(beginDate and endDate) like below in sat-calender element
<sat-calendar
[beginDate]="calendar.beginDate"
[endDate] = "calendar.maxDate">
The my setup is:
The user click one day and the week is selected (from Monday to Sunday). When click other day of week this reset week and re-click day for select week.
The bug is when the user (after clicked the day and selected week) re-click the same day or day in week selected, the saturn fire (beginDateSelectedChange) and this work badly.
Example StackBlitz
So i would reset range of days or send to set beginDate = null
NOTE: i don't konow because stackblitz don't work with momentjs. So you must try in local for try work example
Any solution boys??
Hope, maybe it will help you, To reset i have tried like this, passing | assigning the [beginDate] and [endDate] value as null on sat-calendar.
Don't work this!
I tried in this mode but don't work same:
@ViewChild(SatCalendar, {static: true}) calendar: SatCalendar<Date> this.calendar.beginDateSelected = false; this.calendar.beginDate = null; this.calendar.maxDate = null; this.calendar.selected = null; this.calendar.startAt = null; ;
So it's a BUG of library and must to be fix or insert new solution
You have to pass the values(beginDate and endDate) like below in sat-calender element
<sat-calendar
[beginDate]="calendar.beginDate"
[endDate] = "calendar.maxDate">
The my setup is:
The user click one day and the week is selected (from Monday to Sunday). When click other day of week this reset week and re-click day for select week.
The bug is when the user (after clicked the day and selected week) re-click the same day or day in week selected, the saturn fire (beginDateSelectedChange) and this work badly.
Example StackBlitz
So i would reset range of days or send to set beginDate = null
NOTE: i don't konow because stackblitz don't work with momentjs. So you must try in local for try work example
EMPORANY SOLUTION:
I Added SCSS/CSS attributes on root style (style.scss)
`.mat-calendar-body-selected {
pointer-events: none;
cursor: default;
text-decoration: none;
color: black;
}
.mat-calendar-body-begin-range:not(.mat-calendar-body-end-range) {
pointer-events: none;
cursor: default;
text-decoration: none;
color: black;
}
.mat-calendar-body-end-range:not(.mat-calendar-body-begin-range) {
pointer-events: none;
cursor: default;
text-decoration: none;
color: black;
}
.mat-calendar-cell-semi-selected {
pointer-events: none;
cursor: default;
text-decoration: none;
color: black;
}`
|
2025-04-01T06:37:31.837646
| 2024-06-23T21:39:59
|
2368900336
|
{
"authors": [
"Scarvy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2598",
"repo": "Scarvy/readwise-to-apple-notes",
"url": "https://github.com/Scarvy/readwise-to-apple-notes/issues/4"
}
|
gharchive/issue
|
export not exporting all highlights
I'm suspicious that the export_highlight function is not giving me all my highlights.
I ran it recently and only received 82 total notes in Apple Notes. I have way more than that...
I think it has something to do with the generator or the API wrapper I'm using.
def export_highlights(
updated_after: str = None, book_ids: str = None, token: str = None
) -> Generator[dict, None, None]:
"""Exports the highlights of books based on modification date and/or specific book IDs.
This function iterates over pages of highlights fetched from the client service,
filtering by update time and book IDs if provided, and yields each highlight.
Parameters:
updated_after (str, optional): The ISO 8601 date string to filter highlights
that were updated after a certain date. Defaults to None.
book_ids (str, optional): A comma-separated string of book IDs to filter
highlights by specific books. Defaults to None.
token (str): A Readwise API token. Default to None.
Yields:
dict: A dictionary representing a single book's highlight.
"""
client = get_client(token)
params = {}
if updated_after:
params["updatedAfter"] = updated_after
if book_ids:
params["ids"] = book_ids
for data in client.get_pagination_limit_20("/export/", params=params):
for book in data["results"]:
yield book
The issue was in the API wrapper (pyreadwise) not requesting the next page in the pagination. Based on the API documentation, the /export/ endpoint uses the parameter pageCursor while the other endpoints like /highlights/ use page.
pageCursor – (Optional) A string returned by a previous request to this endpoint. Use it to get the next page of books/highlights if there are too many for one request.
page – specify the pagination counter.
I am deciding whether to create a pull request in the original API wrapper repo or write my own. I'm leaning toward making a pull request.
|
2025-04-01T06:37:31.851690
| 2015-01-15T18:56:35
|
54488838
|
{
"authors": [
"fj",
"ronen",
"thmzlt"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2599",
"repo": "SchemaPlus/schema_plus",
"url": "https://github.com/SchemaPlus/schema_plus/issues/196"
}
|
gharchive/issue
|
Multiple indices on same column are ignored when loading schema on Postgres
When adding a table index with the same column of a foreign key index created by schema_plus (i.e. two indices on the same column with different names), loading the schema will ignore the first index declared in the create_table block for the table.
[This is on OS X 10.9, Postgres 9.3.5 and Ruby 2.1.5]
I have a test Rails application you can use to reproduce the problem. Clone the repo on the index branch, create the database, and try loading and then dumping the schema. (I used rake db:drop db:create db:schema:load db:schema:dump.)
The following diff comes up:
diff --git a/db/schema.rb b/db/schema.rb
index c2bbaf7..2cdbfff 100644
--- a/db/schema.rb
+++ b/db/schema.rb
@@ -28,7 +28,6 @@ ActiveRecord::Schema.define(version:<PHONE_NUMBER>4027) do
t.integer "airplane_id"
t.integer "pilot_id"
t.index ["airplane_id"], :name => "fk__airplanes_pilots_airplane_id"
- t.index ["pilot_id"], :name => "fk__airplanes_pilots_pilot_id"
t.index ["pilot_id"], :name => "index_airplanes_pilots_on_pilot_id", :unique => true
t.foreign_key ["airplane_id"], "airplanes", ["id"], :on_update => :no_action, :on_delete => :no_action, :name => "fk_airplanes_pilots_airplane_id"
t.foreign_key ["pilot_id"], "pilots", ["id"], :on_update => :no_action, :on_delete => :no_action, :name => "fk_airplanes_pilots_pilot_id"
If you manually change the order of the two indices on pilot_id, and load & dump the schema, then the diff becomes:
diff --git a/db/schema.rb b/db/schema.rb
index c2bbaf7..749bd9d 100644
--- a/db/schema.rb
+++ b/db/schema.rb
@@ -29,7 +29,6 @@ ActiveRecord::Schema.define(version:<PHONE_NUMBER>4027) do
t.integer "pilot_id"
t.index ["airplane_id"], :name => "fk__airplanes_pilots_airplane_id"
t.index ["pilot_id"], :name => "fk__airplanes_pilots_pilot_id"
- t.index ["pilot_id"], :name => "index_airplanes_pilots_on_pilot_id", :unique => true
t.foreign_key ["airplane_id"], "airplanes", ["id"], :on_update => :no_action, :on_delete => :no_action, :name => "fk_airplanes_pilots_airplane_id"
t.foreign_key ["pilot_id"], "pilots", ["id"], :on_update => :no_action, :on_delete => :no_action, :name => "fk_airplanes_pilots_pilot_id"
end
:+1:, I have this issue too.
The expected schema_plus behavior for me is that the schema does not have a diff after loading and dumping.
@thmzlt thanks for (yet another!) detailed bug report.
This is arguably a bug in rails: t.index only handles one index per column, see https://github.com/rails/rails/blob/4-1-stable/activerecord/lib/active_record/connection_adapters/abstract/schema_definitions.rb#L238. It's tickled by schema_plus though, because rails' schema dumper defines indexes using add_index outside the create_table block, whereas schema_plus tidies the dump by defining them using t.index.
That said, it's probably bad form to have two indexes on the same field anyway; the database will be doing extra unneeded work to maintain them both. No?
The reason you're getting two is that in your migration
create_table :airplanes_pilots do |t|
t.references :airplane
t.references :pilot
end
add_index :airplanes_pilots, [:pilot_id], unique: true
schema_plus (by default) auto-creates an index for the foreign key constraint; and the add_index statement creates a second one. In order to get just one index and have it be unique, you can do:
create_table :airplanes_pilots do |t|
t.references :airplane
t.references :pilot, index: { unique: true } # or just index: :unique
end
I could fix the bug by having schema_plus dump the extra indexes for a column using add_index outside the block. But since this seems like a bug only when doing something that probably shouldn't be done, and there's another way to achieve the result you want, I'm thinking I'll let this slide.
FWIW this is fixed in the schema_plus 2.0 prerelease, i.e. in the new schema_plus_indexes gem.
Thanks @ronen!
|
2025-04-01T06:37:31.855914
| 2017-12-31T19:29:19
|
285273097
|
{
"authors": [
"Schm1tz1",
"brainstain"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2600",
"repo": "Schm1tz1/aws-sdk-arduino-esp8266",
"url": "https://github.com/Schm1tz1/aws-sdk-arduino-esp8266/pull/2"
}
|
gharchive/pull-request
|
Update to enable DynamoDB
Added a more resilient response reader and changed DynamoDB to force HTTPS
Thanks for your contribution - after some testing I will tag a new release probably by end of this week.
|
2025-04-01T06:37:31.859306
| 2024-06-06T06:57:23
|
2337485095
|
{
"authors": [
"eifelmicha"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2601",
"repo": "Schmitzis/Uptime",
"url": "https://github.com/Schmitzis/Uptime/issues/1769"
}
|
gharchive/issue
|
⚠️ Jenkins Plugins Artifactory has degraded performance
In b8ee78e, Jenkins Plugins Artifactory (https://get.jenkins.io/plugins/artifactory/) experienced degraded performance:
HTTP code: 200
Response time: 1230 ms
Resolved: Jenkins Plugins Artifactory performance has improved in 37b41fe after 35 minutes.
|
2025-04-01T06:37:31.861186
| 2016-06-20T07:09:06
|
161129845
|
{
"authors": [
"postblue"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2602",
"repo": "Schoewilliam/Stitch-Blue",
"url": "https://github.com/Schoewilliam/Stitch-Blue/issues/30"
}
|
gharchive/issue
|
Add a subscribe page
And make it conditionnal to the @labs activation?
Here's the documentation: https://themes.ghost.org/docs/subscribers
Subscription form in Casper: https://github.com/TryGhost/Casper/blob/master/post.hbs
https://github.com/TryGhost/Casper/blob/master/post.hbs#L75
|
2025-04-01T06:37:31.865418
| 2024-05-21T18:59:18
|
2308920100
|
{
"authors": [
"EEvangelisti",
"rob-ferg"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2603",
"repo": "SchornacklabSLCU/amfinder",
"url": "https://github.com/SchornacklabSLCU/amfinder/issues/34"
}
|
gharchive/issue
|
Cannot select prediction file to import (amfbrowser)
Hello,
I've just finished installing amf and amfbrowser and was able to generate predictions for an image using amf.
But when I run amfbrowser and try to select the prediction file, I cannot click on the OK button. I've tried clicking, selecting with Enter, and just about every other key on my keyboard.
My lab mate thinks it is because I am trying to run this on Windows 11. Do you know if there is any way to run this without downgrading to Windows 10? Alternatively I could try installing on a system with Linux.
What is your recommendation?
Thanks,
Robbie
University of Ottawa
Dear Robbie,
Thank you for your interest in AMFinder.
Unfortunately, I have never heard of such an issue, so I can only guess it is a compatibility problem between the graphical system and the OS. The best is to install AMFinder on a Linux machine (it could be a virtual machine). I am happy to help with the installation if needed.
Best regards,
Edouard
|
2025-04-01T06:37:31.886453
| 2024-03-19T10:17:23
|
2194557498
|
{
"authors": [
"HenrZu"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2604",
"repo": "SciCompMod/memilio",
"url": "https://github.com/SciCompMod/memilio/pull/971"
}
|
gharchive/pull-request
|
55 Write mobility data
Changes and Information
Please briefly list the changes (main added features, changed items, or corrected bugs) made:
Save data about the commuters during mobility
Any Indices (or Groups of indices) can be chosen as input. Additionally, the total number of commuters is saved
Additionally, it is possible to save only selected edges
If need be, add additional information and what the reviewer should look out for in particular:
This functionality should be used carefully because if we save to many compartments from the edges, theres a large amount of storage necessary
Merge Request - Guideline Checklist
Please check our git workflow. Use the draft feature if the Pull Request is not yet ready to review.
Checks by code author
[x] Every addressed issue is linked (use the "Closes #ISSUE" keyword below)
[x] New code adheres to coding guidelines
[x] No large data files have been added (files should in sum not exceed 100 KB, avoid PDFs, Word docs, etc.)
[x] Tests are added for new functionality and a local test run was successful (with and without OpenMP)
[x] Appropriate documentation for new functionality has been added (Doxygen in the code and Markdown files if necessary)
[x] Proper attention to licenses, especially no new third-party software with conflicting license has been added
[ ] (For ABM development) Checked benchmark results and ran and posted a local test above from before and after development to ensure performance is monitored.
Checks by code reviewer(s)
[ ] Corresponding issue(s) is/are linked and addressed
[ ] Code is clean of development artifacts (no deactivated or commented code lines, no debugging printouts, etc.)
[ ] Appropriate unit tests have been added, CI passes, code coverage and performance is acceptable (did not decrease)
[ ] No large data files added in the whole history of commits(files should in sum not exceed 100 KB, avoid PDFs, Word docs, etc.)
[ ] On merge, add 2-5 lines with the changes (main added features, changed items, or corrected bugs) to the merge-commit-message. This can be taken from the briefly-list-the-changes above (best case) or the separate commit messages (worst case).
Closes #55
Using the setup in our simulation with ~30.000 edges and 50 simulation days, the size of the Edges.h file is 135mb.
There is an inconsistency in our project between the use of the terms migrated and mobility. I would leave them here for now and edit this PR after merging so that we only use mobility.
|
2025-04-01T06:37:31.891833
| 2016-09-14T07:56:28
|
176839718
|
{
"authors": [
"Sebastian-D",
"szilvajuhos"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2605",
"repo": "SciLifeLab/CAW",
"url": "https://github.com/SciLifeLab/CAW/pull/66"
}
|
gharchive/pull-request
|
MuTect1
@MaxUlysse and @Sebastian-D pls have a look - it is a duplication of MuTect2 practically. On long term we have to refactor these into one, especially the collate part.
Looks good, I see the influences ;).
I am doing something similar with HaplotypeCaller, the collateFiles function is bothering me however as I need to change it for normals and I am not proficient at groovy, yet.
|
2025-04-01T06:37:31.929598
| 2019-11-18T12:27:33
|
524336194
|
{
"authors": [
"trexfeathers"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2606",
"repo": "SciTools/iris",
"url": "https://github.com/SciTools/iris/issues/3543"
}
|
gharchive/issue
|
ASV - Document Airspeed Velocity
ASV configuration and a basic benchmark are now part of Iris (#3526). Having completed some basic experiments with ASV against Iris master branch I have some recommended minimum detail that will be needed in our developer documentation testing page:
[x] A new top level section called Performance Benchmarking
[x] The fact we use ASV
[x] The benchmark directory and prefix syntax for tests
[x] The configuration file location and its implications (e.g. the fact it aligns with the requirements for Iris test, ASV will be using conda to manage the environments it creates, etc...)
[x] A standard run configuration that will be used for official benchmarking - currently asv run v2.0.0..master --skip-existing-commits. OR a Nox session that encodes the same thing; this would probably be preferable.
Closed by #4621
|
2025-04-01T06:37:31.936587
| 2022-03-24T10:52:07
|
1179337692
|
{
"authors": [
"bsherratt",
"pp-mo",
"tkknight",
"trexfeathers"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2607",
"repo": "SciTools/iris",
"url": "https://github.com/SciTools/iris/issues/4657"
}
|
gharchive/issue
|
Iris API param not rendering
📚 Documentation
Whilst reviewing the dsocs for another task I noticed there are a few occurrences of :param: that is not rendering correctly due to alignment and maybe other syntax issues.
https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/pp.html?highlight=%3Aparam%3A#iris.fileformats.pp.save
https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/nimrod_load_rules.html?highlight=%3Aparam%3A#iris.fileformats.nimrod_load_rules.run
https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/dot.html?highlight=%3Aparam%3A#iris.fileformats.dot.save
There maybe other occurrences too, not dug any further.
I didn't even know that was supposed to be possible! Would be great if it worked 🙂
It definitely does work in many places.
For example : https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/coords.html#iris.coords.CellMethod.method
which comes from : https://github.com/SciTools/iris/blob/main/lib/iris/coords.py#L2974-L2975
However, I'm seeing that some other things are not so happy..
Sphinx seems to have started putting two of these on one line sometimes. #4660
Which may or not be related..
The consistent thing between those examples is that they are formatted like
Args:
* something - description
which is probably a few too many steps away from what napoleon expects. I would try:
removing the * list markers
removing the indentation
replacing the - with :
I have a branch primed to fix this issue but ideally want to use my read the docs test space to render them for all to see, thish is being used for another PR atm. Will create the PR next week sometime.
Related to https://github.com/SciTools/iris/issues/3840.
Fixed by https://github.com/SciTools/iris/issues/3840.
|
2025-04-01T06:37:31.938297
| 2016-04-25T15:29:36
|
150899771
|
{
"authors": [
"pp-mo",
"rhattersley"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2608",
"repo": "SciTools/iris",
"url": "https://github.com/SciTools/iris/pull/1984"
}
|
gharchive/pull-request
|
Remove obsolete import of iris.fileformats.dot
This explicit import should not be needed - the import is intended to be deferred.
Can't argue with that !
|
2025-04-01T06:37:31.941831
| 2020-09-15T16:36:52
|
702084577
|
{
"authors": [
"bjlittle",
"trexfeathers"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2609",
"repo": "SciTools/iris",
"url": "https://github.com/SciTools/iris/pull/3851"
}
|
gharchive/pull-request
|
whatsnew - branded section icons
🚀 Pull Request
Description
This PR adds icon branding to the whatsnew section headers.
Some of the icons align with those already chosen for the associated issues icon templates.
If and when #3850 is merged, I'll align the Announcements section appropriately.
To help reviewers, the rendered readthedocs changes of this PR are available:
latest whatsnew.
Contribution categories
Consult Iris pull request check list
Nice, we seem to have converted to the 'fun logo pack', which I am fully in favour of!
@trexfeathers Superb! Thanks :+1:
|
2025-04-01T06:37:31.953710
| 2024-07-21T02:00:15
|
2421170890
|
{
"authors": [
"moeshin"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2610",
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/pull/13671"
}
|
gharchive/pull-request
|
<EMAIL_ADDRESS>Fix install
Closes #13579
[x] I have read the Contributing Guide.
/verify
|
2025-04-01T06:37:31.955412
| 2024-11-02T12:17:36
|
2630427192
|
{
"authors": [
"Gitoffthelawn",
"aliesbelik"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2611",
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/pull/14347"
}
|
gharchive/pull-request
|
<EMAIL_ADDRESS>Fix homepage, add EOL note
Fixes homepage
Closes #14346
[x] I have read the Contributing Guide.
/verify
|
2025-04-01T06:37:31.966532
| 2023-02-08T17:28:22
|
1576526567
|
{
"authors": [
"Zliced13"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2612",
"repo": "ScoopInstaller/Versions",
"url": "https://github.com/ScoopInstaller/Versions/pull/992"
}
|
gharchive/pull-request
|
systeminformer-nightly: Enhance pre_uninstall script
Relates to https://github.com/ScoopInstaller/Versions/pull/885#issuecomment-1418137949
[x] I have read the Contributing Guide.
/verify
|
2025-04-01T06:37:31.969829
| 2023-10-13T17:32:11
|
1942356240
|
{
"authors": [
"Dantevg",
"Scotsguy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2613",
"repo": "Scotsguy/now-playing",
"url": "https://github.com/Scotsguy/now-playing/issues/34"
}
|
gharchive/issue
|
Crash on Fabric 1.20.1 for 1.4.4
The game crashes at the main menu in 1.20.1 (because music starts playing instantly). Version 1.4.3 worked fine, as does 1.4.4 on 1.20.2.
Mods:
Fabric API 0.90.0+1.20.1
Now Playing 1.4.4
crash-2023-10-13_19.20.38-client.txt
Sorry, looks like I mistagged that version. Use 1.4.3 for 1.20.1 and 1.4.4 for 1.20.2.
|
2025-04-01T06:37:31.972726
| 2019-02-21T15:29:13
|
412982885
|
{
"authors": [
"MattCline-SL",
"sl-slaing"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2614",
"repo": "ScottLogic/datahelix",
"url": "https://github.com/ScottLogic/datahelix/pull/622"
}
|
gharchive/pull-request
|
577 - Refactor to remove ProfileProvider
Description
Previous work was put in place to mark the ProfileProvider as deprecated and to get round to removing it. This goes as close as we currently can to removing it by replacing it with a much simpler CurrentProfileCache. This is now set at data generation stage and is only consumed once in HierarchalFieldFixStrategy.
I subsequently tried to remove this dependency too but the change was too large and risky to do (see PR #624
Changes
Replaced ProfileProvider with CurrentProfileCache.
Adjusted field fix strategies to use CurrentProfileCache or ProfileFields where appropriate.
Issue
Resolves #577
This change appears to simply move the problem back one stage. It certainly removes the Guice imports in code but I'm not sure it provides any material change.
On the face of it the change should be focused around removing the profile dependency from the FixFieldStrategy or something to this effect. This would then allow for a complete change.
Also, the tree is a representation of the profile, albeit an unsymmetrical adaptation. There should still be sufficient information here to support analysis of the profile (by way of its tree representation) to provide the order of fields to fix.
|
2025-04-01T06:37:31.982775
| 2021-08-07T05:37:08
|
963160674
|
{
"authors": [
"gigios",
"swharden",
"zeticabrian"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2615",
"repo": "ScottPlot/ScottPlot",
"url": "https://github.com/ScottPlot/ScottPlot/issues/1194"
}
|
gharchive/issue
|
Signal Plot: Is vertical orientation possible?
This issue will explore feasibility of creating vertical signal plots. The signal plot type is one of the most complex, with many features that would have to be refactored (e.g., color by density) making this task very challenging. An alternative option is to create a new VerticalSignal plot type with minimal features.
Discussed in https://github.com/ScottPlot/ScottPlot/discussions/1184
Originally posted by gigios August 5, 2021
First of all I apologize if the question has already been asked, but I have not found it 😊
I'm investigating the functionality of the library to be able to draw a graph with thousands of values (signal).
My problem is that I need to show 10 charts with the sample rate in the Y axis and values in X.
Exists a simple way to do this with the Signal plot?
What I need is similar to this photo
I need to replicate a similar visualization in my wpf app with the possibility to scroll all the signals in vertical and, if there are more than 10 charts, scroll also in horizontal (but the scroll I think is not a problem).
A possible solution could be rotate the Wpf Plot Control of 90 degree, but in this way I need also to rotate the Axis labels.
Thanks for any suggestion.
I have added, in the discussion, a first test of a possible workaround in the Wpf application, where it's very easy to rotate controls. I have only some troubles on place the axis labels in the correct position, but this is probably a my problem because it's the first time I'm using the 'ScottPlot' 😊
With the WPF workaround (LayoutTransform\RotateTransform in xaml) and the last version of the ScottPlot I'm able to 'simulate' a vertical orientation for the signal plot 👍
Now there is only a small problem related the position of the axis label because are too close to the chart. This probably depends how the rotation is applied to the label text.
This is the WpfSample, where I have rotated the control in the xaml (90 degree) and also the Axis Y2.
Rotated ScottPlot Control
<DockPanel>
<DockPanel.LayoutTransform>
<RotateTransform
CenterX="0.5"
CenterY="0.5"
Angle="90"/>
</DockPanel.LayoutTransform>
<ScottPlot:WpfPlot Name="WpfPlot1"/>
</DockPanel>
public MainWindow()
{
InitializeComponent();
WpfPlot1.Plot.AddSignal(ScottPlot.DataGen.Sin(51));
WpfPlot1.Plot.AddSignal(ScottPlot.DataGen.Cos(51));
// hide original vertical axis
WpfPlot1.Plot.YAxis.Ticks(false);
WpfPlot1.Plot.YAxis.Label("Title");
// vertical axis (when rotated)
WpfPlot1.Plot.XAxis.Ticks(true);
WpfPlot1.Plot.XAxis.TickLabelStyle(rotation: 90);
// horizontal axis (when rotated)
WpfPlot1.Plot.YAxis2.Ticks(true);
WpfPlot1.Plot.YAxis2.TickLabelStyle(rotation: 90);
WpfPlot1.Render();
}
Perfect, thanks.
Do you think to release a new nuget update in a short time? Otherwise I will try to use directly the source code to do my additional tests 👍
Perfect, thanks.
Sure thing! I'm happy with the outcome here.
Do you think to release a new nuget update in a short time?
I'll make a NuGet release after I finish #1242 which I anticipate completing today or tomorrow 👍
> // horizontal axis (when rotated)
> WpfPlot1.Plot.YAxis2.Ticks(true);
> WpfPlot1.Plot.YAxis2.TickLabelStyle(rotation: 90);
Its great that teh tick labels can be rotated, but unfortunately this leaves the title for YAxis2 (the new horizontal axis) upside down. Is there any means to rotate the axis titles?
My rotated implementation (left) compared with standard implementation (right):
Is there any means to rotate the axis titles?
Thanks for this question! Currently no, but it seems like an oversight. I created #1754 to address this deficiency and anticipate I can add this features in the next few days. I'll post updates on that issue page 👍
|
2025-04-01T06:37:32.104512
| 2019-03-21T01:45:58
|
423542609
|
{
"authors": [
"SeBassTian23",
"Xymph",
"excel-vba-course",
"nouranianfar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2616",
"repo": "SeBassTian23/CalendarHeatmap",
"url": "https://github.com/SeBassTian23/CalendarHeatmap/issues/1"
}
|
gharchive/issue
|
Update and Change Calendar data
Dear
I want to update calendar data real time with script.
My data if change I could not update calendar data.
thanks
I apologize for the long silence! Can you provide an example of your code?
Like this heatmap-variant more than others because months are shown separately and it has a nice set of customization options.
But I have the same problem. Script:
function refresh()
{
$('<my selector>').each( function() {
/*$('#act' + $(this).html()).html('');*/
$.ajax({
url: '<script returning JSON>.php',
type: 'GET',
data: {
year: $(this).html(),
<name1>: $('input[name=<inputname1>]').val(),
<name2>: $('input[name=<inputname2>]').val(),
},
dataType: 'json',
success: function(result) {
$('#act' + result.year).CalendarHeatmap(result.data, {
lastMonth: 12,
lastYear: result.year,
/* more cosmetic options */
});
}
});
});
}
$(function() {
refresh();
});
At page load, the heatmaps are rendered. There's a form with two input fields that also invoke 'refresh()' at the onchange event. I can see the Ajax calls returning the expected data, but the heatmaps aren't refreshed with the changed data sets. If I then reload the page (preserving form values), the updated heatmaps are rendered as yet.
If I enable the commented-out html('') call to clear the divs, the heatmaps are removed at onchange, but no new ones are rendered.
Could this be caused by the main "this.calendarHeatmap();" call being invoked only from the plugin's "init:" method?
Btw, in step 6 here, the tiles option with its values could be documented.
Thank you for the detailed description of the issue. I will try to update the plugin ASAP.
Regarding the link you have provided. I think the issue is that this service is not updating the Readme file. Please find everything documented including the tiles here.
That was encouraging to read -- any progress?
Even if my use of the heatmap plugin is only in a private, non-vital application, I'd love it if refreshing works as expected.
Hey mate - also was wondering about how to achieve updating of control with a different dataset. Do you have a workaround to achieve this ?
Thanks in advance
I found some time to work on the plugin and the functions to update the dates (data) and the options are available with the latest release (v.1.2.0) now.
I updated my application just now, and updateDates does the job. Thanks a lot!
I'm glad it is working for you now.
|
2025-04-01T06:37:32.105737
| 2015-10-20T11:26:20
|
112346502
|
{
"authors": [
"andreaturli",
"perezp"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2617",
"repo": "SeaCloudsEU/SeaCloudsPlatform",
"url": "https://github.com/SeaCloudsEU/SeaCloudsPlatform/pull/130"
}
|
gharchive/pull-request
|
increased debug verbosity and adding node templates in optimizer output
The information of cloud resources was not included in the output information of the optimizer, only the names of the cloud resources.
This PR adds such information to the ADPs created by the optimizer.
lgtm
|
2025-04-01T06:37:32.115671
| 2015-12-10T18:06:30
|
121544879
|
{
"authors": [
"MicheleGuerriero",
"adriannieto",
"andreaturli",
"codecov-io",
"jacopogiallo",
"kiuby88",
"perezp",
"rosogon",
"szenzaro"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2618",
"repo": "SeaCloudsEU/SeaCloudsPlatform",
"url": "https://github.com/SeaCloudsEU/SeaCloudsPlatform/pull/188"
}
|
gharchive/pull-request
|
Feature/group management
Groups are expected by Deployer as a topology_template's child.
@szenzaro @MicheleGuerriero please take a look to this commit. Probably, we could expect groups as a topology_template's child.
Current coverage is 32.17%
Merging #188 into master will increase coverage by +0.07% as of 8cb6bdf
@@ master #188 diff @@
======================================
Files 319 319
Stmts 11676 11676
Branches 1472 1472
Methods 0 0
======================================
+ Hit 3749 3757 +8
+ Partial 387 384 -3
+ Missed 7540 7535 -5
Review entire Coverage Diff as of 8cb6bdf
Powered by Codecov. Updated on successful CI builds.
+1 for me
@kiuby88 : This branch has conflicts. Could you resolve this?
@rosogon sorry for the late response.
Rebased
I think this PR will make the monitoring-dam-generator-core module to not work since it changes the format of the input and my ADP parser will fail. If it is needed to be compliant with the TOSCA specification I can fix my parser, but I have some doubts this doesn't affect only me. @kiuby88 can you please help me in clarify how to handle this point?
@MicheleGuerriero We can talk on monday and I can make clear this change. I agree with you that this might not affect only you.
I would keep the PR open until we solve these issues.
@kiuby88
Stop using just Strings!
@kiuby88 please refactor
public static String generateDam(String adp, String monitorGenURL, String monitorGenPort, String slaGenURL) {
into something Object Oriented like
public DAM generateDAM(ADP adp)
where ADP is probably just a4c's Topology and DAM is Topology + Policies, I think.
Do you think it is feasible?
@kiuby88 I'm doing something similar like @andreaturli proposes with SeaCloudsApplicationData.java (in https://github.com/adriannieto/SeaCloudsPlatform/blob/feature/move-orchestration-from-js-to-java/dashboard/src/main/java/eu/seaclouds/platform/dashboard/model/SeaCloudsApplicationData.java just in case you want to take a look), I agree with @andreaturli but @szenzaro must agree on this, because maybe it will involve a lot of changes.
@MicheleGuerriero , I did not know that your component used groups. I am sorry.
I have moved groups migration to generateDam method.
Of course, it is a TEMPORAL solution.
I have solved issues 198 and 199 on this PR as a temporal solution. These issues should be fixed from AAM generation.
@szenzaro , could you take a look?
We (UPI) agree on the updates by @kiuby88
Thanks a lot @jacopogiallo and @andreaturli
|
2025-04-01T06:37:32.139965
| 2021-05-20T15:51:32
|
897016407
|
{
"authors": [
"Shreya-18",
"knekrasov",
"mssawant"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2619",
"repo": "Seagate/cortx-hare",
"url": "https://github.com/Seagate/cortx-hare/pull/1622"
}
|
gharchive/pull-request
|
EOS-19656: Notify M0_NC_FAILED for motr configuration objects on server node failure
Description
Jira - EOS-19656
On server node failure notification, Hare needs to report M0_NC_FAILED for the corresponding motr configuration objects, viz: node, enclosure (post EOS-17599), controller, and drives to all the Motr processes including s3servers.
Solution
In broadcast_ha_states(), added a check for process failures.
fetch node from the failed process fid
fetch fids of all configuration objects
fetch node fid
fetch enclosure fid
fetch controller fid
append all states (fid: status) to HaNoteStruct notes
which is further notified to motr processes.
Note: how to be sure that process failure is due to node failure?
We don't need to. Logically node failure and process failure will have the same effect.
If the node fails devices are not accessible, or even if the process fails we lose access to devices. Without ioservice, we cannot access the devices on that node so they are effectively failed.
@Shreya-18 Can you please re-phrase your first two commits (or probably squash them if they don't have a historical value per se)?
@Shreya-18 Can you please re-phrase your first two commits (or probably squash them if they don't have a historical value per se)?
Done, and rebased too
@Shreya-18 , @knekrasov , please resolve the comments if addressed to satisfaction.
@Shreya-18 , we need to make following changes,
diff --git a/hax/hax/handler.py b/hax/hax/handler.py
index d79a9ce..b5815f1 100644
--- a/hax/hax/handler.py
+++ b/hax/hax/handler.py
@@ -98,8 +98,6 @@ class ConsumerThread(StoppableThread):
# notifications, it may cause delay in cleanup
# activities.
continue
- else:
- current_status = ServiceHealth.STOPPED
if current_status == ServiceHealth.UNKNOWN:
# We got service status as UNKNOWN, that means hax was
# notified about process failure but hax couldn't
diff --git a/hax/hax/motr/__init__.py b/hax/hax/motr/__init__.py
index e6b7b33..ec58740 100644
--- a/hax/hax/motr/__init__.py
+++ b/hax/hax/motr/__init__.py
@@ -277,18 +277,13 @@ class Motr:
continue
note = HaNoteStruct(st.fid.to_c(), ha_obj_state(st))
notes.append(note)
- notify_node = False
if (st.fid.container == ObjT.PROCESS.value and
st.status == ServiceHealth.STOPPED):
notify_devices = False
# For process failure, we report failure for the corresponding
# node (enclosure) and CVGs.
- notify_node = True
notes += self._generate_sub_services(note, self.consul_util,
notify_devices)
- if notify_node:
- notes += self.notify_node_failure(note)
-
if not notes:
return []
message_ids: List[MessageId] = self._ffi.ha_broadcast(
@@ -368,6 +363,7 @@ class Motr:
service_notes = [HaNoteStruct(no_id=x.fid.to_c(), no_state=new_state)
for x in service_list]
if notify_devices:
+ service_notes += self.notify_node_failure(note)
service_notes += self._generate_sub_disks(note, service_list, cns)
return service_notes
ctrl, node and enclosure failure notifications need to be sent in case the process status is FAILED.
And there's another change which was prohibiting FAILED status to be escalated. I have included both the changes in the patch.
Will post the details of degraded IO steps.
@Shreya-18 , we need to make following changes,
ctrl, node and enclosure failure notifications need to be sent in case the process status is FAILED.
And there's another change which was prohibiting FAILED status to be escalated. I have included both the changes in the patch.
Will post the details of degraded IO steps.
@mssawant I have a question
when a node goes down, what should be the service health of the failed node's processes -- stopped or failed?
My observations are when I reboot a node, we get notifications of the process's health as STOPPED.
@Shreya-18 , we need to make following changes,
diff --git a/hax/hax/handler.py b/hax/hax/handler.py
index d79a9ce..b5815f1 100644
--- a/hax/hax/handler.py
+++ b/hax/hax/handler.py
@@ -98,8 +98,6 @@ class ConsumerThread(StoppableThread):
# notifications, it may cause delay in cleanup
# activities.
continue
- else:
- current_status = ServiceHealth.STOPPED
if current_status == ServiceHealth.UNKNOWN:
# We got service status as UNKNOWN, that means hax was
# notified about process failure but hax couldn't
diff --git a/hax/hax/motr/__init__.py b/hax/hax/motr/__init__.py
index e6b7b33..ec58740 100644
--- a/hax/hax/motr/__init__.py
+++ b/hax/hax/motr/__init__.py
@@ -277,18 +277,13 @@ class Motr:
continue
note = HaNoteStruct(st.fid.to_c(), ha_obj_state(st))
notes.append(note)
- notify_node = False
if (st.fid.container == ObjT.PROCESS.value and
st.status == ServiceHealth.STOPPED):
notify_devices = False
# For process failure, we report failure for the corresponding
# node (enclosure) and CVGs.
- notify_node = True
notes += self._generate_sub_services(note, self.consul_util,
notify_devices)
- if notify_node:
- notes += self.notify_node_failure(note)
-
if not notes:
return []
message_ids: List[MessageId] = self._ffi.ha_broadcast(
@@ -368,6 +363,7 @@ class Motr:
service_notes = [HaNoteStruct(no_id=x.fid.to_c(), no_state=new_state)
for x in service_list]
if notify_devices:
+ service_notes += self.notify_node_failure(note)
service_notes += self._generate_sub_disks(note, service_list, cns)
return service_notes
Done, updated the patch with changes.
@Shreya-18, the patch looks good, degraded IO was successful with a minor change to cfgen.
[root@ssc-vm-1623:root] m0client -l '<IP_ADDRESS>@tcp:12345:4:1' -H '<IP_ADDRESS>@tcp:12345:1:1' -p '<0x7000000000000001:0xb6>' -P '<0x7200000000000001:0x25>'
motr[119715]: 9930 ERROR [conf/confc.c:557:m0_confc_init_wait] <! rc=-110 confc=0x7f25a40086d0 confd_addr=<IP_ADDRESS>@tcp:12345:2:1
motr[119715]: 9930 ERROR [conf/confc.c:557:m0_confc_init_wait] <! rc=-110 confc=0x7f25a4009e50 confd_addr=<IP_ADDRESS>@tcp:12345:2:1
motr[119715]: 9a80 WARN [conf/rconfc.c:2419:rconfc_quorum_is_possible] No chance left to reach the quorum
motr[119715]: 9bd0 ERROR [conf/rconfc.c:2571:rconfc_version_elected] <! rc=-71 re-election started
m0client >>write 1048680 /root/read_1048680_1 4096 200 50
m0client >>motr[119715]: 1380 ERROR [pool/pool_machine.c:472:spare_usage_arr_update] <7600000000000001:70>: No free spare space slot is found, this pool version is in DUD state; event_index=16 event_state=2
motr[119715]: 1380 ERROR [pool/pool_machine.c:472:spare_usage_arr_update] <7600000000000001:70>: No free spare space slot is found, this pool version is in DUD state; event_index=17 event_state=2
motr[119715]: a890 ERROR [rpc/frmops.c:474:item_done] packet 0x7f25e00540b0, item 0x7f25e0053b40[36] failed with ri_error=-5
motr[119715]: a890 ERROR [rpc/frmops.c:474:item_done] packet 0x7f25e00540b0, item 0x7f25e0053b40[38] failed with ri_error=-5
Help:
touch OBJ_ID
write OBJ_ID SRC_FILE BLOCK_SIZE BLOCK_COUNT BLOCKS_PER_IO UPDATE_FLAG OFFSET
read OBJ_ID DEST_FILE BLOCK_SIZE BLOCK_COUNT BLOCKS_PER_IO
delete OBJ_ID
help
quit
m0client >>read 1048680 /root/read_1048680_node_fail_degraded 4096 200 50
motr[119715]: 16d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 16d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 16d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 16d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 1790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: 1790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: 1a80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f26380ed250 rc=-125
motr[119715]: 1a80 ERROR [io_req_fop.c:210:io_bottom_half] ioo=0x14f5d00 from=<IP_ADDRESS>@tcp:12345:2:2 rc=-125 ti_rc=-125 @<4300001100000000:100068>
motr[119715]: 1a80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f26380eb6b0 rc=-125
motr[119715]: 1950 WARN [io_req.c:1484:ioreq_dgmode_read] Process failed parity groups in dgmode/read ioo=0x14f5d00 dgmap_nr=18
motr[119715]: 36d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 36d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 36d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 36d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: 3790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: 3790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: 3a80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f2624047f40 rc=-125
motr[119715]: 3a80 ERROR [io_req_fop.c:210:io_bottom_half] ioo=0x14f5d00 from=<IP_ADDRESS>@tcp:12345:2:2 rc=-125 ti_rc=-125 @<4300001100000000:100068>
motr[119715]: 3a80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f2624046660 rc=-125
motr[119715]: 3950 WARN [io_req.c:1484:ioreq_dgmode_read] Process failed parity groups in dgmode/read ioo=0x14f5d00 dgmap_nr=17
motr[119715]: e6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: e6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: e6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: e6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: e790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: e790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: ea80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f261c0612c0 rc=-125
motr[119715]: ea80 ERROR [io_req_fop.c:210:io_bottom_half] ioo=0x14f5d00 from=<IP_ADDRESS>@tcp:12345:2:2 rc=-125 ti_rc=-125 @<4300001100000000:100068>
motr[119715]: ea80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f261c05ee30 rc=-125
motr[119715]: e950 WARN [io_req.c:1484:ioreq_dgmode_read] Process failed parity groups in dgmode/read ioo=0x14f5d00 dgmap_nr=18
motr[119715]: c6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: c6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: c6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: c6d0 ERROR [rpc/session.c:578:m0_rpc_session_validate] <! rc=-125 Cancelled session
motr[119715]: c790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: c790 ERROR [rpc/rpc.c:126:m0_rpc__post_locked] <! rc=-125
motr[119715]: ca80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f261405cc30 rc=-125
motr[119715]: ca80 ERROR [io_req_fop.c:210:io_bottom_half] ioo=0x14f5d00 from=<IP_ADDRESS>@tcp:12345:2:2 rc=-125 ti_rc=-125 @<4300001000000000:100068>
motr[119715]: ca80 ERROR [io_req_fop.c:161:io_bottom_half] [0x14f5d00] rpc item 0x7f261405b910 rc=-125
motr[119715]: c950 WARN [io_req.c:1484:ioreq_dgmode_read] Process failed parity groups in dgmode/read ioo=0x14f5d00 dgmap_nr=17
m0client >>
[root@ssc-vm-1623:root] ls -l /root/read_1048680_node_fail_degraded
-rw-r--r-- 1 root root 819200 Jun 14 17:40 /root/read_1048680_node_fail_degraded
[root@ssc-vm-1623:root] diff /root/read_1048680_node_fail_degraded /root/read_1048680_1
[root@ssc-vm-1623:root]
|
2025-04-01T06:37:32.149915
| 2022-12-07T22:51:55
|
1483034503
|
{
"authors": [
"MartinSimangoDaemn",
"SeanJxie"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2620",
"repo": "SeanJxie/polygo",
"url": "https://github.com/SeanJxie/polygo/issues/8"
}
|
gharchive/issue
|
FindRootsWithin not always returning correct results
Hi @SeanJxie
I found that FindRootsWithin doesn't always return the correct roots.
Here is an example of a polynomial i used:
x^21 - 86400x +86399=0
When I used FindRootsWithin(1,100) I get the following roots: [14.912790 1.000000] : (14.912790 is not a valid root)
When I used FindRootsWithin(0,100) I got the following roots: [1.000000 1.687851] - which are correct.
When I used FindRootsWithin(-100,100) I got the following roots: [-14.765139 1.000000 1.687851] of which -14.765139 is not a valid root.
Hello @MartinSimangoDaemn
Thank you for pointing this bug out. It has been fixed as a result of the Polygo rewrite.
p := NewPolyFromString("x^21 - 86400x + 86399")
s := NewSolverDefault()
fmt.Println(s.FindRootsWithin(p, 1, 100))
fmt.Println(s.FindRootsWithin(p, 0, 100))
fmt.Println(s.FindRootsWithin(p, -100, 100))
Now, you can also find all roots at once:
fmt.Println(s.FindRoots(p))
Cheers,
Sean
|
2025-04-01T06:37:32.158490
| 2024-05-16T19:01:52
|
2301135981
|
{
"authors": [
"MaddasaHatter",
"MuhammadNSC",
"brinkbrink",
"kaimanasse15",
"taylorpapke"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2622",
"repo": "SeattleColleges/nsc-events-nextjs",
"url": "https://github.com/SeattleColleges/nsc-events-nextjs/pull/344"
}
|
gharchive/pull-request
|
nsc-events-nextjs_6_341_unit-test-for-event-detail-page
This is the Unit test for Event-detail Page
Issue: #341
I have an error with this part and still not figure out yet.
Not all the tests passed. I did a bit of research to see what the issue here might be and to my knowledge i think to resolve this issue, you need to mock the Next.js router in your test. This can be achieved using jest.mock to create a mock implementation of the useRouter hook. Other classmates might have better solutions to help fix the issue
@kaimanasse15 Did you try @MuhammadNSC's advice?
@kaimanasse15 Did you try @MuhammadNSC's advice?
It doesn't work for me. I think I need someone else to follow this. Thanks, @brinkbrink
It's still failing were you able to try other resources for this?
Closing due to inactivity. Reopen if needed.
|
2025-04-01T06:37:32.163689
| 2024-05-23T20:18:29
|
2313759968
|
{
"authors": [
"Diego-Cano",
"MaddasaHatter",
"brinkbrink",
"kaimanasse15",
"nwm516",
"tinpham5614"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2623",
"repo": "SeattleColleges/nsc-events-nextjs",
"url": "https://github.com/SeattleColleges/nsc-events-nextjs/pull/364"
}
|
gharchive/pull-request
|
nsc-events-nextjs_7_295_arrow_icons
(Help is needed on this User Story/Work in progress)
Relates to #293
[ ] Add arrow icons to the left and right side of the event.
When the right arrow is clicked, the event being displayed switches to the next one. If there are no more events, the right arrow should disappear.
When the left arrow is clicked, the event being displayed switches to the previous one. If there are no more previous events, the left arrow should disappear.
I need to hand off this User Story to another capable developer. This is beyond my scope of knowledge, so any assistance is greatly appreciated. Thank you.
Hello i think this will be a great imporvement for the UI on the app. I agree that the left arrow should disappear if there are no previous events, and the right arrow should do the same if there are no more upcoming events. The changes should be made in the page.tsx file.
I noticed that clicking the arrows doesn't change the event being displayed yet, so that part still needs fixing. On the plus side, the date format issue is sorted, and dates now show up correctly without the timestamp, Good Job. This improvement makes the event details page look much better.
For more info on navigation features in React I found this link : https://reactrouter.com/en/main/components/navigate
Looks good. I'm able to see the dates in the proper format mm-dd-yyyy format without the timestamp. I'm also able to see the arrow on the event-detail page but when I click the arrow it doesn't change the event. I can pick up where you left off on this user story/issue
Thank you, Robel!
There are conflicts with the main branch.
So far it looks good the function is there and I can really see a difference in usability with these additions. keep up the good work. I did notice though there may be some merge conflicts.
The arrows looks good but I am not able to create an event but I think some data is conflicinting and not your error.
|
2025-04-01T06:37:32.218722
| 2022-05-03T08:58:48
|
1223843071
|
{
"authors": [
"SeedSigner",
"blackcoffeexbt",
"newtonick"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2624",
"repo": "SeedSigner/seedsigner",
"url": "https://github.com/SeedSigner/seedsigner/pull/201"
}
|
gharchive/pull-request
|
Add Simple Pill enclosure stl files
💊💊💊
LGTM. I don't personally have a setup to verify/print these designs. But I have no issue adding these to the repo. @SeedSigner thoughts?
I am extremely grateful for @blackcoffeexbt for making this PR and for his very high quality design. My only question is on the installation location for the camera. Have any modifications been made to widen the front-to-back spacing for the stock camera board? I have printed the simple pill and done a couple of builds, and each time I have had to use a dremel to shave down the camera board to be able to install it. Having assembled hundreds of SeedSigners, I feel as though the current open pill has hit the sweet spot for this front-to-back clearance. The camera board doesn't necessarily have to snap in tightly, as it will generally be held in place by the Pi Zero. I wish there wasn't so much variance in the sizing of those camera boards, but feel like we should accommodate for it without users needing to potentially try to modify their camera board. Open to your thoughts @blackcoffeexbt and thank you again for being willing to contribute this design, a lot of people are going to get a ton of use and enjoyment out of it.
I just completed a test print and the fit is perfect. Immensely appreciate @blackcoffeexbt for open sourcing this design.
|
2025-04-01T06:37:32.226213
| 2023-05-29T19:34:56
|
1731165638
|
{
"authors": [
"Seeker14491",
"tibordp"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2625",
"repo": "Seeker14491/opener",
"url": "https://github.com/Seeker14491/opener/issues/25"
}
|
gharchive/issue
|
opener::open does not fail at all, even when e.g. file is missing on Linux
Fedora 38 + Gnome (Wayland)
Repro:
> xdg-open foo.sav
gio: file:///home/tibordp/foo.sav: Error when getting information for file “/home/tibordp/foo.sav”: No such file or directory
> echo $?
4
> cat repro.rs
#!/usr/bin/env rust-script
//! ```cargo
//! [dependencies]
//! opener = "0.6.1"
//! ```
fn main() {
const FILE: &str = "foo.sav";
opener::open(FILE).unwrap();
println!("{} opened successfully!", FILE);
}
> ./repro.rs
foo.sav opened successfully!
> echo $?
0
When the system xdg-open fails, as in this case, we fall back to our internal xdg-open script:
https://github.com/Seeker14491/opener/blob/ff6e2599f1360332dac2b87877f9e1694ad654c3/opener/src/linux_and_more.rs#L52-L58
It seems the internal xdg-open erroneously succeeds in this case, so no error is reported. To fix this, maybe we shouldn't be using the fallback in the case where the system xdg-open exists.
Tried with the bundled xdg-open and it fails as expected
tibordp@mnemosyne:~ $ ./bundled-xdg-open.sh foo.sav
gio: file:///home/tibordp/foo.sav: Error when getting information for file “/home/tibordp/foo.sav”: No such file or directory
tibordp@mnemosyne:~ $ echo $?
4
The way I'm seeing is the problem is that the library just checks if the spawning fails, but does not actually check the exit code of the spawned process.
https://github.com/Seeker14491/opener/blob/master/opener/src/linux_and_more.rs#L69-L98
Ah, you're right, though that is intentional. The issue is that we don't want to block on the open() call, so we don't wait to check the exit code.
Interesting - is there a reason the crate waits on Mac then? At least on my system, xdg-open exits once the file has been opened and does not wait for the process to terminate.
Yes, on Mac the system-provided open command is used, which doesn't block (unless you pass in the -W flag which waits for the program to exit). On Linux, we can't say in general whether xdg-open will or will not block. In the past, this crate actually did wait for xdg-open's status code so we could detect failure, but we encountered the issue of the command blocking in some environments, so it was changed.
|
2025-04-01T06:37:32.244391
| 2018-09-05T16:07:07
|
357300336
|
{
"authors": [
"broxen",
"dracc",
"eos428"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2626",
"repo": "Segs/Segs",
"url": "https://github.com/Segs/Segs/pull/548"
}
|
gharchive/pull-request
|
Creating a Semi-Standalone EmailHandler that derives from EventProcessor (Part 2)
Closes #367 .
Continuation of the previous PR that has since been pulled down because I changed its name.
At the moment it runs, but the actual MapInstance/DataHelper-EmailHandler interaction is not there yet.
Remaining things to do:
Parse EmailSend correctly in slashCommands. It seems to send the name of the recipient instead of the db_id which does complicate things a bit...
EmailRead and EmailDelete should be alright, since you'd just need the email id and the handlers will deal the rest
Create Email column in Database and integrate it with EmailHandler, DbHandler and DbContext
At the moment these are the major things to do. I expect further issues to arise, and then there's the matter of polishing it too.
Just a note here. As emails are sure to be updated less frequently than CharData, maybe we should stick them in their own table linked to char by db_id as pk?
That's the plan from the beginning. The question is how do I make them?
Once the email table is up and running and integrated with the email handler, I'd say we'd be really close to being done here :D
Some questions I have based on my latest commit (And yeah, it's a bit messy and plenty of placeholders but at least it can run):
Getting emails through sender_id and recipient_id would return multiple emails, should I send an entire vector containing all the emails over a single event?
There are two cases where we'd update an email: To mark the email as read, and when the sender character and/or account is deleted.
Should we put two separate events for this (Something like EmailMarkAsReadMessage and EmailUpdateOnCharacterDeleteMessage)?
And should we put m_is_read_by_recipient outside the blob, since sending the entire email data just to mark it as read feels a little cumbersome.
EmailRemove should be okay, since the only case we'd actually delete the email is from the recipient deleting it from his/her inbox.
That's all for now. I am getting confident we'd finish this before 0.6.0 releases :D
Apart from the very tiny suggested changes, the SQL part of this PR looks good to me. :+1:
Should be working fine except for sending email to names where some other character might have a more extended name of the recipient, such as admin2 sending an email to admin, which instead finds the recipient_id of admin2.
And, I think we can send an email to the same character. It was my oversight of not storing newly-created emails to m_stored_email_datas.
Hi eoS! Thank you for tackling the Email system!
I'm getting some errors on this PR though.
Recipient of email doesn't receive email in their inbox
Sender appears to have gotten email in their inbox instead? Read receipt fires immediately
Sender client crashes when clicking on reply
Writing an email to an offline character causes segfault, and the sender's ID is stored in the database for recipient id.
Debug : Email Sent to recipient: EmailGuy1
authserver: /Segs/Projects/CoX/Common/Servers/ClientManager.h:118: SESSION_CLASS& ClientSessionStore<SESSION_CLASS>::session_from_token(uint64_t) [with SESSION_CLASS = MapClientSession; uint64_t = long unsigned int]: Assertion `iter != m_token_to_session.end()' failed.
Thank you eoS! Great PR!
Most of the fixes seemed to work, but I received a segfault when opening an email received while offline.
Debug : Opening Email ID: 1
Debug : Opening Email ID: 5
authserver: /Segs/Projects/CoX/Common/Servers/ClientManager.h:118: SESSION_CLASS& ClientSessionStore<SESSION_CLASS>::session_from_token(uint64_t) [with SESSION_CLASS = MapClientSession; uint64_t = long unsigned int]: Assertion `iter != m_token_to_session.end()' failed.
crashing when session is used here:
void MapInstance::on_email_read_by_recipient(EmailWasReadByRecipientMessage *msg)
{
MapClientSession &map_session(m_session_store.session_from_token(msg->session_token()));
Steps to reproduce:
Create two characters on two separate accounts. Send emails back and forth. All OK.
Log out char-2, create new char (char-3). Send email from char-3 to char-2, so far, all OK.
Log out of char-3, and log in char-2. Select new email to read: segfault =(
Notes
Database looks correct, in terms of sender_id and recipient_id equating to the db_id of the characters
Other issues noted:
As mentioned on discord, sending to multiple recipients doesn't seem to work as intended, but I'm comfortable with leaving that for a future PR.
Also, as you've mentioned before, timestamps aren't working, but can be in a followup PR as well.
|
2025-04-01T06:37:32.251094
| 2018-10-26T09:20:58
|
374297453
|
{
"authors": [
"JulienBreux",
"a-rose"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2627",
"repo": "Seize/joy4",
"url": "https://github.com/Seize/joy4/pull/9"
}
|
gharchive/pull-request
|
[fpsconv] fix framerate increasing
PR Type
[ ] Feature
[x] Bug fix
[ ] Docs
Description
Freeing output images caused crashes when increasing the framerate (for example 20 fps to 25 fps)
This PR reworks the free mechanism to properly manage increased fps.
Notes for your reviewer
I also improved debug traces
Yes
On Fri 26 Oct 2018 at 3:22 PM, Antoine R<EMAIL_ADDRESS>wrote:
ping @JulienBreux https://github.com/JulienBreux, can I merge this ? 😃
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/Seize/joy4/pull/9#issuecomment-433405371, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AA626hkoG5X1ZnrykJLli0ywk4sqkqWVks5uowyAgaJpZM4X72S5
.
--
[image: photo]
Julien Breux
+33 687 764 228 | https://www.julienbreux.uk
CTO @SeizeTV https://seize.tv?utm_source=email_jbreux
Founder @Trofify https://trofify.io/?utm_source=email_jbreux
http://us.linkedin.com/in/JulienBreux http://twitter.com/JulienBreux
http://github.com/JulienBreux
|
2025-04-01T06:37:32.320988
| 2023-01-10T14:56:13
|
1527537027
|
{
"authors": [
"daidaiworm",
"userKer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2629",
"repo": "Sellix/python-sdk",
"url": "https://github.com/Sellix/python-sdk/issues/2"
}
|
gharchive/issue
|
requests.exceptions.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
Python: 3.10.9
Sellix: sellix-python-sdk 1.0.5
replaced <YOUR_API_KEY> and <MERCHANT_NAME> with my own.
encounter this issue:
requests.exceptions.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
from sellix import Sellix
# pass <MERCHANT_NAME> only if you need to be authenticated as an additional store
client = Sellix("<YOUR_API_KEY>", "<MERCHANT_NAME>")
try:
products = client.get_products()
except Sellix.SellixException as e:
print(e)
aslo ImportError: cannot import name 'Sellix' from partially initialized module 'sellix' (most likely due to a circular import)
|
2025-04-01T06:37:32.322429
| 2016-12-18T02:25:10
|
196254132
|
{
"authors": [
"lednhatkhanh",
"levithomason"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2630",
"repo": "Semantic-Org/Semantic-UI-React",
"url": "https://github.com/Semantic-Org/Semantic-UI-React/issues/1048"
}
|
gharchive/issue
|
Sidebar docs's broken
Here's the log when I go to sidebar docs:
Thanks for the report. There's a new deploying going out with the fix.
|
2025-04-01T06:37:32.324781
| 2017-02-20T03:12:18
|
208774074
|
{
"authors": [
"levithomason",
"tarang9211"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2631",
"repo": "Semantic-Org/Semantic-UI-React",
"url": "https://github.com/Semantic-Org/Semantic-UI-React/issues/1351"
}
|
gharchive/issue
|
Add label docs usage example for Progress
The Progress component can take a label enum prop to control the type of label displayed within the Progres bar:
ratio
<Progress value='3' total='5' label='ratio' />
percent
<Progress value='3' total='5' label='percent' />
PR is #1356
|
2025-04-01T06:37:32.334317
| 2015-11-05T15:34:24
|
115309264
|
{
"authors": [
"danr1979",
"larsbo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2632",
"repo": "Semantic-Org/Semantic-UI",
"url": "https://github.com/Semantic-Org/Semantic-UI/issues/3290"
}
|
gharchive/issue
|
Help with SUI Upgrade on Windows
Hi.
I've been using SUI for around a year now (it's fantastic!), but in order to streamline the upgrade process for our team I want to use npm. So far I have done:
npm install semantic-ui
gulp build
and have a working installation which I have customised for our needs to build to different output folders.
However, I have recently tried to upgrade from version 2.1.4 to 2.1.5 and I'm not sure things are working for me as they should. I did:
npm update
from within the semantic-ui folder where I originally did the install. This runs through fine and when I check the semantic version number in the node_modules\semantic-ui folder I see 2.1.5 so that's great. The problem is that this version is not mirrored in the semantic-ui installation directory because the semantic.json file in there still reads 2.1.4. I tried doing:
npm install
from inside the semantic-ui installation folder but I get a load of errors relating to a missing package.json file. I appreciate that the package.json file is sitting in the node_modules/semantic-ui folder and not in the semantic-ui install folder and that is why npm is reporting an error so I assume what I have attempted here is wrong.
I have read all of the documentation on installs/upgrades here:
http://learnsemantic.com/guide/expert.html#running-installer
but they aren't clear on what to do in my specific case to ensure that semantic.json shows the correct version telling me that my installation is running the latest version.
Could someone give me a hand understanding what I am doing wrong and what I need to do to upgrade cleanly please?
Many thanks.
In my case running npm update didn't update semantic ui, my package.json still contained "semantic-ui": "^2.1.4".
So I tried npm install<EMAIL_ADDRESS>which worked fine.
After switching into semantic/ and gulp build I got the latest semantic ui version running.
Hope this helps :)
Fantastic! Thanks, that did it!
|
2025-04-01T06:37:32.339248
| 2016-11-16T09:08:25
|
189637837
|
{
"authors": [
"Banandrew",
"KeZhimin",
"bukaOK"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2633",
"repo": "Semantic-Org/Semantic-UI",
"url": "https://github.com/Semantic-Org/Semantic-UI/issues/4751"
}
|
gharchive/issue
|
How to change content when touch different tab
the code like this:
<div class="ui container" style="margin-top: 10px;">
<div class="ui tabular menu">
<a class="active item" data-tab="first">First</a>
<a class="item" data-tab="second">Second</a>
</div>
<div class="ui bottom attached tab segment" data-tab="first">
</div>
<div class="ui bottom attached tab segment" data-tab="second">
</div>
</div>
<script type="text/javascript">
$('.menu .item').tab({
cache: false,
// faking API request
apiSettings: {
loadingDuration : 300,
mockResponse : function(settings) {
var response = {
first : 'query1.json',
second : 'query2.json'
};
return response[settings.urlData.tab];
}
},
context : 'parent',
path : '/'
});
</script>
when I click the First tab, i want get the content from the url 'query1.json', when i click the the Secode tab, i want get the content from the url 'query2.json', the code like that dont work.
@KeZhimin maybe you need something like this..
$('.menu .item').tab(); $('.tab[data-tab="first"]').api({ url: 'query1.json', //Another settings }); $('.tab[data-tab="second"]').api({ url: 'query2.json', //Another settings });
Hi @KeZhimin,
Thank you for posting, but although it’s a valid usage question, we’ve limited GitHub Issues to bug reports and feature requests, keeping the board more manageable for maintainers; see the contributing guidelines for more information on what kind of posts should find themselves into the GitHub Issues board.
To get answers or feedback that might allow you to repost this issue, please use one of our other support resources:
Gitter chatroom
Official forums
StackOverflow
|
2025-04-01T06:37:32.406104
| 2018-08-10T15:01:06
|
349549594
|
{
"authors": [
"asger-semmle",
"xiemaisi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2634",
"repo": "Semmle/ql",
"url": "https://github.com/Semmle/ql/pull/48"
}
|
gharchive/pull-request
|
JavaScript: Add WebView-related taint sinks for CodeInjection, DomBasedXss and ServerSideUrlRedirect.
I did a small evaluation on four projects that use WebView including VSCode and TinyMCE. Performance was unchanged and there were no new results. I can run a larger-scale evaluation over the weekend if desired.
I was going to suggest that we check for the flag javaScriptEnabled which can be used to turn off JavaScript in the web view. On the other hand, allowing non-scripted html injection can still be pretty bad, and looking at the docs now, it seems the flag only works on Android(?!). What do you think?
Otherwise LGTM.
Yes, I saw that, but as you say it only seems to apply to Android. I didn't really understand whether it was just always enabled on iOS or whether you could turn it off there as well (in which case it would be the same as on Android, so that's probably not it). In the end, I decided to just ignore it for now.
|
2025-04-01T06:37:32.429401
| 2020-04-27T14:57:08
|
607618014
|
{
"authors": [
"PhilZ-cwm6"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2635",
"repo": "Sentaroh/SMBSync2",
"url": "https://github.com/Sentaroh/SMBSync2/pull/48"
}
|
gharchive/pull-request
|
Hide network advanced options when pertinent
If neither Target nor Master are SMB shares, hide "Limit SMB I/O write buffer to 16 KB" and "Retry on network error" options
They still were showing even when no SMB share was selected
Note: give me a few days to validate the french menus, I will push a commit when done
After that, it will take me a few more time/weeks to finish fixing english help docs + translate them to french
Keep up the good work 👍
there is one more smb option left...
|
2025-04-01T06:37:32.438594
| 2022-07-27T13:31:20
|
1319571831
|
{
"authors": [
"roncewind"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2636",
"repo": "Senzing/aws-cloudformation-database-cluster",
"url": "https://github.com/Senzing/aws-cloudformation-database-cluster/pull/53"
}
|
gharchive/pull-request
|
Issue 52.ron.0
Pull request questions
Which issue does this address
Issue number: #52
Why was change needed
???
What does change improve
???
No changes needed, closing PR
|
2025-04-01T06:37:32.443341
| 2022-03-02T21:49:52
|
1157729524
|
{
"authors": [
"SeparateRecords",
"drakebrian"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2638",
"repo": "SeparateRecords/apple_device_identifiers",
"url": "https://github.com/SeparateRecords/apple_device_identifiers/issues/13"
}
|
gharchive/issue
|
HomePod models
Here are the model numbers for the current HomePod lineup.
{
"HomePod": "AudioAccessory1,1",
"HomePod": "AudioAccessory1,2",
"HomePod Mini": "AudioAccessory5,1",
}
Thanks for the great resource!
Thanks!! Since it's JSON the keys need to be different - couldn't find a hardware difference between AudioAccessory1,1 and AudioAccessory1,2.
Maybe something like "HomePod" and "HomePod (Revision)"?
Good catch, I can't find any meaningful differences either. The model number for 1,1 is B238aAP while 1,2 is B238AP.
I like (Revision) better than (Unknown Model) that I see elsewhere. I'll give contributing a go, thanks!
|
2025-04-01T06:37:32.517370
| 2019-08-19T05:45:39
|
482116150
|
{
"authors": [
"SergioCrisostomo",
"webistomin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2640",
"repo": "SergioCrisostomo/vue-codemods",
"url": "https://github.com/SergioCrisostomo/vue-codemods/issues/1"
}
|
gharchive/issue
|
Add codemod to extract methods from instance that are not instance specific
If a method does not use this we should extract it from the instance and have it as a global function assignment const functionName = () => {... etc
Fixed in 051f97c36fabd6f5381ba7fecefcd439f0197ffb
Can you explain the benefits of this approach?
Hi @webistomin!
There are different reasons for wanting to remove methods from a Vue component that are not dependent on the instance. I do not mean we should always do it, but by doing it we can:
prevent creating "fat" instances with methods that could be outside the instance and do not need to be in memory of every object/instance
identify possible utils functions and extract them to another file to keep the file size manageable
better test functions not related to the instance
If a method does not use the this then it can live outside the component. In some cases it makes sense to keep it so all the logic is in one place. This codemod allows to identify these and extract them.
Hi @webistomin!
There are different reasons for wanting to remove methods from a Vue component that are not dependent on the instance. I do not mean we should always do it, but by doing it we can:
prevent creating "fat" instances with methods that could be outside the instance and do not need to be in memory of every object/instance
identify possible utils functions and extract them to another file to keep the file size manageable
better test functions not related to the instance
If a method does not use the this then it can live outside the component. In some cases it makes sense to keep it so all the logic is in one place. This codemod allows to identify these and extract them.
Thanks!
|
2025-04-01T06:37:32.533544
| 2016-03-16T13:10:30
|
141269775
|
{
"authors": [
"MrCircuit",
"nzain"
],
"license": "MS-PL",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2641",
"repo": "SeriousM/WPFLocalizationExtension",
"url": "https://github.com/SeriousM/WPFLocalizationExtension/issues/95"
}
|
gharchive/issue
|
Localized KeyGesture
I'm looking into ways to localize a tooltip and KeyGesture (think of a button with keyboard shortcut), with the desired result
[en] This button does something (ctrl + H)
[de] Macht irgendwas (strg + H)
Note that the modifiers are localized, too. I know this is related to the every-once-and-then-raised question for localized texts with bindings (now gap text may offer a solution), but this is not exactly the same problem. System.Windows.Input.KeyGesture has a function GetDisplayStringForCulture(CultureInfo c) which returns the nicely formatted "ctrl + H" thing. Tricky part is how to integrate this with WPF-LocEx.
Different (unsatisfying) solutions:
I could hardcode key + modifier into the localization resources.
Drawback: The key/modifier must not change. Wenn key or modifier change, all localizations have to be updated and we have some customers that translate stuff themselves.
I could localize the tooltip message with WPF-LocEx and localize the gesture with the builtin function.
Drawback a) the latter is not sensitive to culture changes. I think one could work around this with weak references and a callback, though.
Drawback b) Cumbersome to use. The code here is terrible for such a small functionality.
<Button Content="{lex:LocText MyButton}">
<Button.ToolTip>
<TextBlock>
<Run Text="{lex:LocText MyButtonTooltip}"/>
<Run Text="("/><Run Text="{Binding KeyGestureText, Mode=OneWay}"/><Run Text=")"/>
</TextBlock>
</Button.ToolTip>
</Button>
Drawback c) The position of the gesture hint is hard coded.
I could not make this work with StringFormat of a Binding
<TextBlock Text="{Binding KeyGestureText, StringFormat={lex:LocText KeyToFormattedTooltip}}"/>
Although the formatted tooltip is retreived (checked via debugger), it has no effect (not sure why, but I guess it is a known issue).
I was thinking of a similar solution to EnumRun, but could not make it work. Any directions?
Ok, it seems the function KeyGesture.GetDisplayStringForCulture does not, what I expected. On my system it always returns "ctrl" regardless of the culture argument. Kind of useless (?). I iterated over all available cultures using CultureInfo.GetCultures()... result is invariant.
Anyways, I've put together a UserControl (single TextBlock only) that takes a (localization dictionary) key and a KeyGesture. Much like the EnumRun it updates its text on every property change, but additionally registers in a custom WeakReference manager class that notifies all interested (alive) references about culture changes. So... theoretically, the gesture text would update dynamically too - practically, the above mentioned function does not work. Thus, I'm not sure if this has any relevance for other users. Furthermore, it is still cumbersome to use, mhh.
<Button Command="{Binding ToggleUnitCommand}"
Content="{lex:LocText MainWindow/ChangeUnit}">
<Button.ToolTip>
<ctrl:GestureTooltip Key="MainWindow/ChangeUnitTooltip"
Gesture="{Binding ToggleUnitCommand.KeyGesture}" />
</Button.ToolTip>
</Button>
For your interest:
https://msdn.microsoft.com/en-us/library/system.windows.input.keygesture.getdisplaystringforculture(v=vs.110).aspx
https://stackoverflow.com/questions/35341543/show-inputgesturetext-as-a-tooltip-for-a-button
Close unless you're interested.
Thanks for the investigations - I did not believe it but I also got the same string for all cultures for KeyGesture.GetDisplayStringForCulture. This is really weird.
In general - if the function would have worked - one could think of a callback functionality (e.g. an assignable Func<string, CultureInfo>), that could be overwritten by particular implementations, e.g. by deriving from the LocExtension. This would leave space for different tasks. Please open a feature request, if this is what you could need in future. Otherwise, I would recommend using the multiple-Run approach you proposed or the Cloze.
|
2025-04-01T06:37:32.567476
| 2017-10-26T10:59:10
|
268716192
|
{
"authors": [
"coveralls",
"seanyinx"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2643",
"repo": "ServiceComb/ServiceComb-Saga",
"url": "https://github.com/ServiceComb/ServiceComb-Saga/pull/50"
}
|
gharchive/pull-request
|
JAV-232 used specific service center version since it has breaking ch…
…anges
Signed-off-by: seanyinx<EMAIL_ADDRESS>
Coverage remained the same at 94.758% when pulling 3f693310b74ae78f5102d589acd3b390eef7d722 on JAV-232_service_center_downgrade into 53e1460289ee7ea6091841a585ee30f65bcc91f6 on master.
|
2025-04-01T06:37:32.581706
| 2018-11-14T01:41:01
|
380499073
|
{
"authors": [
"bweick",
"coveralls"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2644",
"repo": "SetProtocol/set-protocol-contracts",
"url": "https://github.com/SetProtocol/set-protocol-contracts/pull/284"
}
|
gharchive/pull-request
|
Updated rebalancing set token with internal audit fixes. Added furthe…
…r testing for fee distribution.
Pull Request Test Coverage Report for Build 2777
12 of 15 (80.0%) changed or added relevant lines in 1 file are covered.
1 unchanged line in 1 file lost coverage.
Overall coverage decreased (-0.9%) to 99.14%
Changes Missing Coverage
Covered Lines
Changed/Added Lines
%
contracts/core/RebalancingSetToken.sol
12
15
80.0%
Files with Coverage Reduction
New Missed Lines
%
contracts/core/RebalancingSetToken.sol
1
97.0%
Totals
Change from base Build 2763:
-0.9%
Covered Lines:
578
Relevant Lines:
582
💛 - Coveralls
|
2025-04-01T06:37:32.654619
| 2020-05-21T13:21:43
|
622495204
|
{
"authors": [
"Het-Shah",
"NeelayS",
"khizirsiddiqui"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2645",
"repo": "SforAiDl/KD_Lib",
"url": "https://github.com/SforAiDl/KD_Lib/issues/22"
}
|
gharchive/issue
|
Refactor code
Refactor all the following papers:
[ ] TAKD (@Het-Shah)
[ ] Attention (@Het-Shah)
[ ] Noisy
Assign yourselves. Let's get this done quickly and then we can continue adding more papers.
@NeelayS if it is okay to, I will refactor Noisy.
@NeelayS if it is okay to, I will refactor Noisy.
Yeah, sure.
|
2025-04-01T06:37:32.714876
| 2022-11-02T20:12:31
|
1433725951
|
{
"authors": [
"Chidubemmo",
"Flux2011",
"nennneko5787"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2646",
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/11097"
}
|
gharchive/issue
|
Psych Engine Source Code Problem (Please Help)
Describe your problem here.
Hi i'm having a problem with FNF Source Code someone please help because i don't know how to fix this.
Are you modding a build from source or with Lua?
Source
What is your build target?
Windows x64
Did you edit anything in this build? If so, mention or summarize your changes.
No i didn't.
You have failed to load lime.ndll.
Please make sure that lime.ndll is in the folder where PsychEngine.exe is located in the export folder of the source code folder. (For Release builds, export/release/; for debug builds, export/debug/)
But I already have the lime.ndll file and it still doesn't work
I think I once got something similr.
Try reinstalling haxe and all the libraries
|
2025-04-01T06:37:32.718906
| 2024-10-10T18:51:05
|
2579592984
|
{
"authors": [
"Maxi2022gt",
"SantiagoCalebe",
"Shadow06735"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2647",
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/15658"
}
|
gharchive/issue
|
Game not starting
Describe your bug here.
I am just installing this because of fun, but then i encountered an error with SScript and then looking thru issues i found an archive of it. I tried it and it worked but when i did lime test windows the game doesn't start and instead throws an error with HScript.hx.
Command Prompt/Terminal logs (if existing)
source/psychlua/HScript.hx:291: characters 89-96 : Type not found : TeaCall
Are you modding a build from source or with Lua?
Source
What is your build target?
Windows
Did you edit anything in this build? If so, mention or summarize your changes.
No
Btw source version is 0.7.2.
wait i just fixed it but theres another error
Warning: Could not find Visual Studio VsDevCmd
Missing HXCPP_VARS
Error: Could not automatically setup MSVC
The most obvious error.
Change 'Teacall' to 'Tea'
wait i just fixed it but theres another error
Warning: Could not find Visual Studio VsDevCmd
Missing HXCPP_VARS
Error: Could not automatically setup MSVC
you need to download visual studio
i did
also i fixed it so i can close it ig
|
2025-04-01T06:37:32.720879
| 2019-03-01T10:07:09
|
416029031
|
{
"authors": [
"NillerMedDild",
"Shadows-of-Fire"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2648",
"repo": "Shadows-of-Fire/HitWithAxe",
"url": "https://github.com/Shadows-of-Fire/HitWithAxe/issues/7"
}
|
gharchive/issue
|
Nature's Aura Compatibility
Hello!
I would like to request compatibility for Nature's Aura's only tree :)
I made the textures:
ancient_tree.zip
Done. Will be released in 1.4.0. Default script wasn't updated, but I trust you can figure that out.
Thank you!
|
2025-04-01T06:37:32.728990
| 2024-10-04T05:21:14
|
2565471413
|
{
"authors": [
"ShaneD20",
"aljex13"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2649",
"repo": "ShaneD20/EquipmentRandomizerER",
"url": "https://github.com/ShaneD20/EquipmentRandomizerER/pull/50"
}
|
gharchive/pull-request
|
Smithing Stone Update
Smithing Weapon Stone Cost is now...
Three per level for +1 to +9
Two per level for +10 to +24
Removal of unused code.
I couldn't quite figure out how/where to comment on this but I personally wanted to change the costs to 1 stone at all levels but don't know where to start.
|
2025-04-01T06:37:32.732270
| 2019-11-08T10:50:36
|
519946288
|
{
"authors": [
"andrewconnell",
"srinutamada"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2650",
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/issues/4877"
}
|
gharchive/issue
|
Can I hide default "Approve/Reject" item property?
I got a requirement to hide default "Approve/Reject", Publish, Cancel Approval options from item properties.
Requirement:
Users uploads documents to the document library. All these document should appear after review and approval process. The contribute user or approve user should not get an option to publish or approves the document using OOTB operations. These actions should be happen using Flow because of business requirements.
SharePoint online + SPFx
This isn't supported... you can't change the default menu options without using DOM manipulation which isn't supported.
|
2025-04-01T06:37:32.735416
| 2017-05-22T18:43:35
|
230483687
|
{
"authors": [
"VesaJuvonen"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2651",
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/issues/595"
}
|
gharchive/issue
|
Documentation formatting issue at dev.office.com
Category
[ ] Question
[ ] Typo
[ X ] Bug
[ ] Additional article idea
Observed Behavior
Documentation formatting is not working properly at dev.office.com for at least following URL - https://dev.office.com/sharepoint/docs/apis/rest/get-to-know-the-sharepoint-rest-service
Rendered properly in the dev.office.com now. Closing issue.
|
2025-04-01T06:37:32.739528
| 2021-03-05T13:27:40
|
823083191
|
{
"authors": [
"SubbaReddi"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2652",
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/issues/6790"
}
|
gharchive/issue
|
SharePoint site with Traditional Chinese language - 404 not found error
Created site with Traditional chinese language and created new page.
As page title is in chinese, tried to add the same page as personal tab.
But then 404 not found error message is displaying in Teams tab.
Is it a known issue?
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 8ef9b82d-5475-c5fa-ac7a-7005558e06f1
Version Independent ID: 1d9d4923-c44f-b00b-a553-60c73ed8b8e6
Content: Embedding modern SharePoint pages in Microsoft Teams as personal apps (preview)
Content Source: docs/features/embed-pages-to-teams.md
Product: sharepoint
GitHub Login: @VesaJuvonen
Microsoft Alias: vesaj
Above mentioned issue is resolved. Working fine now.
|
2025-04-01T06:37:32.749101
| 2022-09-23T17:01:51
|
1384074350
|
{
"authors": [
"JoanneHendrickson",
"VesaJuvonen",
"andrewconnell",
"gittig23441"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2653",
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/pull/8455"
}
|
gharchive/pull-request
|
Update migration-api-whats-new.md
Category
[X] Content fix/update
What's in this Pull Request?
Added notation of new requirement, the mandatory use of SourceType when calling the Migration API. updated date.
Docs Build status updates of commit e9352e6:
:white_check_mark: Validation status: passed
File
Status
Preview URL
Details
docs/apis/migration-api-whats-new.md
:white_check_mark:Succeeded
View
For more details, please refer to the build report.
Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report.
For any questions, please:Try searching the docs.microsoft.com contributor guidesPost your question in the Docs support channel
@andrewconnell I need to add this article (which is an existing topic) to the TOC but cannot find the file to edit or that has the node for the Sharepoint migration api.. Can you point me to it? thanks joanne
https://github.com/SharePoint/sp-dev-docs/blob/main/docs/toc.yml#L2076
@andrewconnell When can we expect this to go live? thanks!
¯_(ツ)_/¯ ... I think that's a scheduled process, but it's managed by MS... @VesaJuvonen might have insight into the schedule for golive pushes...
@andrewconnell -- thanks and good to know -- If we could know the managed schedule publishing days are (weekly?) I can set expectations with my PMs regarding timing.
Unfortunately, I can't answer those questions... I just help in reviewing PR submissions and update the docs on a weekly basis. Only MS employees (IOW, not me) can push doc updates to prod.
@VesaJuvonen @andrewconnell Vesa: Can you predict when this pr will go live? If you could let me know the managed publishing schedule I can let the PMs know and set expectations. A communication to our ISVs is on hold until this goes live. Thanks!
@VesaJuvonen Yes -- and thank you!
Policy.edit group
|
2025-04-01T06:37:32.754053
| 2020-11-24T13:32:23
|
749710294
|
{
"authors": [
"githubsteveb"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2654",
"repo": "SharePoint/sp-provisioning-service",
"url": "https://github.com/SharePoint/sp-provisioning-service/issues/346"
}
|
gharchive/issue
|
Your request to provision "New Employee Onboarding Hub" template failed!
Describe the issue
Trying to deploy the "Microsoft New Employee Onboarding hub" via MS Lookbook fails.
Expected behavior
A successful deployment
Screenshots
Here's the automated email:
As a bit more info, just before this failed, I deployed the "Microsoft 365 learning pathways" template successfully:
Details (please complete the following information):
PnP Correlation ID: 66b07e81-5a29-4e42-972a-15a8a610d24a
Reference Template: https://lookbook.microsoft.com/details/75e60a32-9849-4ed4-b83e-b2b08983ad19
Primary language of the target tenant: English
Are you using a Tenant Global Admin account to provision the template: Yes
Additional context
Add any other context about the problem here.
Any ideas on this one?
I retried deploying the template again this time using a different name and was successful.
|
2025-04-01T06:37:32.773797
| 2017-10-17T10:19:25
|
266070202
|
{
"authors": [
"FDiskas",
"Sharlaan"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2655",
"repo": "Sharlaan/material-ui-superselectfield",
"url": "https://github.com/Sharlaan/material-ui-superselectfield/issues/75"
}
|
gharchive/issue
|
Multiselect - missing property to sort by selected items
If the list is very big, for the UX - selected items should be placed at the top of the list - I mean sorted by selection.
Nice one! thanks
After some thoughts and by common agreeness with other contributors, we think this should not be the responsibility of the component to manipulate data.
SelectField is just a presenting component.
I'm aware this is kinda borderline, but this request should be done on the developper side
this request should be done on the developper side, superSelectField will automatically detect if children changed -by sorting- and will rerender.
:pear:
|
2025-04-01T06:37:32.788802
| 2018-10-08T02:53:56
|
367624168
|
{
"authors": [
"SheetJSDev",
"ZSH-HSZ",
"tangdw"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2656",
"repo": "SheetJS/js-xlsx",
"url": "https://github.com/SheetJS/js-xlsx/issues/1302"
}
|
gharchive/issue
|
ge is not a constructor
There is no problem when NPM is used without CDN, but after CDN is used the export will report an error, ge is not a constructor
浏览器里需要使用完整版本:xlsx.full.min.js 而不是 xlsx.min.js
@ZSH-HSZ what CDN? can you reproduce using a service like jsfiddle?
|
2025-04-01T06:37:32.795402
| 2019-12-22T18:06:21
|
541483999
|
{
"authors": [
"Shfty"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2657",
"repo": "Shfty/qodot-plugin",
"url": "https://github.com/Shfty/qodot-plugin/issues/59"
}
|
gharchive/issue
|
Clearer / more intuitive download setup
Right now the .gitattributes is making it a bit confusing for users to get their hands on the TrenchBroom plugin and example content, since Clone or download > Download ZIP respects the .gitattributes file and excludes everything except the plugin.
Need to check if there's a more granular way to control this through .gitattributes (ex. "don't include in commit / tagged zips, but do include in branch zips")
If not, need to consider moving some of the content to a different repo. Dependencies and submodule limitations will make this difficult for the example content itself, but it's feasible to move the TB plugin.
Fixed in the repo split.
|
2025-04-01T06:37:32.808578
| 2023-06-25T04:32:15
|
1773051941
|
{
"authors": [
"Anajrim01",
"tvl83"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2658",
"repo": "ShipBit/slickgpt",
"url": "https://github.com/ShipBit/slickgpt/issues/59"
}
|
gharchive/issue
|
Feature Request: Import chat histories from chatgpt @ openai and aichat's cli tool
Personally I am a digital pack-rat. I like to keep my history of chatgpt chats from openai.com. It would be nice to import my previous chat history into this interface.
I also use the cli application aichat https://github.com/sigoden/aichat that I would like to import if possible. They store everything in a single .md file though so I think that would be a lot more difficult.
openai has a export data function that gives you a json file with the conversations.
the aichat cli application does not seem to have an export function but writes everything that happens to a MD file. i will ask over there to have a way to save chats to a json format similar to openai's format so there's only one schema for the data.
Thanks.
I can possibly implement the code for importing the chats. The frontend design will still be need to done by someone else.
|
2025-04-01T06:37:32.813697
| 2018-02-23T11:01:35
|
299674680
|
{
"authors": [
"jatindogra"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2659",
"repo": "Shippable/admiral",
"url": "https://github.com/Shippable/admiral/issues/1967"
}
|
gharchive/issue
|
Use {arch}/{os}/remote/setupNode.sh to install db on a remote machine
https://github.com/Shippable/pm/issues/10242
Completed
|
2025-04-01T06:37:32.861232
| 2022-04-12T22:08:25
|
1202453048
|
{
"authors": [
"Shockah"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2660",
"repo": "Shockah/Stardew-Valley-Mods",
"url": "https://github.com/Shockah/Stardew-Valley-Mods/issues/49"
}
|
gharchive/issue
|
GMCM integration: tooltips not translated to other languages
A user reported that the GMCM integration did not have translated tooltips in non-English languages (the language was Hungarian, but the game still used the default, English translation).
Fixed in dded8a82776c782e8233fae94a1b4427b03ca718.
|
2025-04-01T06:37:32.878296
| 2023-06-14T10:34:01
|
1756574269
|
{
"authors": [
"mohitmoengage"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2661",
"repo": "Shopify/cli",
"url": "https://github.com/Shopify/cli/issues/2169"
}
|
gharchive/issue
|
[Bug]: Adding private package to shopify app
Please confirm that you have:
[X] Searched existing issues to see if your issue is a duplicate. (If you’ve found a duplicate issue, feel free to add additional information in a comment on it.)
[X] Reproduced the issue in the latest CLI version.
In which of these areas are you experiencing a problem?
App
Expected behavior
I am trying to add a private package to web/frontend package.json.
It is able to install locally as my local machine has .npmrc global file which contains the authToken but when i do npm run dev.
It is deploying/redirecting to shopify cloudflare machine which is not able to find the package which fails the whole UI flow
Actual behavior
Shopify is able to find the private package.
Verbose output
N/A
Reproduction steps
Add a private package to frontend
npm run dev
Operating System
macOS Venture version 13.2.1 8-Core Intel Core i9
Shopify CLI version (check your project's package.json if you're not sure)
3.46.3
Shell
No response
Node version (run node -v if you're not sure)
No response
What language and version are you using in your application?
No response
I guess this is related log for this
2023-06-17 13:31:21 │ frontend │ Sat, 17 Jun 2023 13:31:21 GMT connect:dispatcher viteTimeMiddleware : /node_modules/.vite/deps/@moengage_commons.js?v=ef1836f4
2023-06-17 13:31:21 │ frontend │ Sat, 17 Jun 2023 13:31:21 GMT connect:dispatcher corsMiddleware : /node_modules/.vite/deps/@moengage_commons.js?v=ef1836f4
2023-06-17 13:31:21 │ frontend │ Sat, 17 Jun 2023 13:31:21 GMT connect:dispatcher viteProxyMiddleware : /node_modules/.vite/deps/@moengage_commons.js?v=ef1836f4
2023-06-17 13:31:21 │ frontend │ Sat, 17 Jun 2023 13:31:21 GMT connect:dispatcher viteHMRPingMiddleware : /node_modules/.vite/deps/@moengage_commons.js?v=ef1836f4
2023-06-17 13:31:21 │ frontend │ Sat, 17 Jun 2023 13:31:21 GMT connect:dispatcher viteServePublicMiddleware : /node_modules/.vite/deps/@moengage_commons.js?v=ef1836f4
2023-06-17 13:31:21 │ frontend │ Sat, 17 Jun 2023 13:31:21 GMT connect:dispatcher viteTransformMiddleware : /node_modules/.vite/deps/@moengage_commons.js?v=ef1836f4
The js file for this keeps waiting until the socket times out.
|
2025-04-01T06:37:32.942931
| 2024-07-27T00:14:09
|
2433119432
|
{
"authors": [
"adil-tanveer888"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2662",
"repo": "Shopify/react-native-skia",
"url": "https://github.com/Shopify/react-native-skia/issues/2549"
}
|
gharchive/issue
|
React Native Skia Headless Documentation
Description
Are there any working examples of React Native Skia on Node that can be linked in the documentation? I have referred the documentation here: https://shopify.github.io/react-native-skia/docs/getting-started/headless#hello-world but it is not clear how the JSX syntax for the drawings given to the drawOffscreen function will work in node environment.
It would be great if we can give a working example that would work in a plain new nodejs environment without needing any setup.
It was not clear how the existing example in the documentation could be leveraged the same way since it had JSX. Not sure if it will work even with Babel setup.
I ended up using the following modified code. I am putting it here as reference.
import { getSkiaExports } from<EMAIL_ADDRESS>import { LoadSkiaWeb } from<EMAIL_ADDRESS>import * as fs from "fs";
(async () => {
const width = 256;
const height = 256;
// Load CanvasKit
await LoadSkiaWeb();
// Once that CanvasKit is loaded, you can access Skia via getSkiaExports()
const { Skia } = getSkiaExports();
const surface = Skia.Surface.MakeOffscreen(width, height);
if (!surface) {
throw new Error("Could not create surface");
}
const canvas = surface.getCanvas();
const paint = Skia.Paint();
paint.setColor(Skia.Color("#c7522ad0"));
paint.setShader(
Skia.Shader.MakeRadialGradient(
{ x: width / 2, y: height / 2 },
Math.min(width, height) / 4,
[Skia.Color("#c7522ad0"), Skia.Color("#ff8531d0")],
null,
0
)
);
canvas.drawCircle(width / 2, height / 2, Math.min(width, height) / 4, paint);
paint.setColor(Skia.Color("white"));
paint.setBlendMode(0);
canvas.drawCircle(width / 2 + 2, height / 2 + 2, 4, paint);
surface.flush();
const image = surface.makeImageSnapshot({ x: 0, y: 0, width, height });
fs.writeFile(
"./img.png",
Buffer.from(image.encodeToBase64(), "base64"),
(err) => {
if (err) {
console.error(err);
} else {
// file written successfully
console.log("File written successfully.");
}
}
);
image.dispose();
surface.dispose();
})();
In my opinion, we should have something like this which somebody can just copy paste and play around in a plain nodejs environment with just the required dependencies.
|
2025-04-01T06:37:32.944668
| 2023-10-13T10:08:02
|
1941653305
|
{
"authors": [
"masonle2x2",
"wcandillon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2663",
"repo": "Shopify/react-native-skia",
"url": "https://github.com/Shopify/react-native-skia/pull/1917"
}
|
gharchive/pull-request
|
Add ellipsisText method
const font = useFont(require('Roboto.ttf'), 14);
const clip = font.ellipsisText({ text: "Hello World", type: 'clip', width: 60 });
const head = font.ellipsisText({ text: "Hello World", type: 'head', width: 60 });
const middle = font.ellipsisText({ text: "Hello World", type: 'middle', width: 60 });
const tail = font.ellipsisText({ text: "Hello World", type: 'tail', width: 60 });
Result:
sounds like you are working on something really cool :)
I will close this for now but let us know as always if you have any feedback questions.
|
2025-04-01T06:37:32.979457
| 2018-07-06T08:45:19
|
338855641
|
{
"authors": [
"steffatstellaviecom"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2664",
"repo": "Shopify/slate",
"url": "https://github.com/Shopify/slate/issues/666"
}
|
gharchive/issue
|
npm start, Hostname/IP doesn't match certificate's altnames?
Getting started with Slate 1.x. Stumbling at "npm start" right after setting up the .env file. The theme is complied successfully, but followed by this error: Hostname/IP doesn't match certificate's altnames: "Host: EXAMPLE.myshopify.com. is not in the cert's altnames: DNS:*.myshopify.com, DNS:myshopify.com"
Error: Hostname/IP doesn't match certificate's altnames: "Host: example.myshopify.com. is not in the cert's altnames: DNS:*.myshopify.com, DNS:myshopify.com"
at Object.checkServerIdentity (tls.js:223:17)
at TLSSocket.<anonymous> (_tls_wrap.js:1111:29)
at emitNone (events.js:106:13)
at TLSSocket.emit (events.js:208:7)
at TLSSocket._finishInit (_tls_wrap.js:639:8)
at TLSWrap.ssl.onhandshakedone (_tls_wrap.js:469:38)
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR<EMAIL_ADDRESS>start: `slate-tools start`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the<EMAIL_ADDRESS>start script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
Solved.
|
2025-04-01T06:37:32.983823
| 2019-03-14T16:36:51
|
421126319
|
{
"authors": [
"Maurice-JB",
"andyexeter",
"wizardlyhel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2665",
"repo": "Shopify/slate",
"url": "https://github.com/Shopify/slate/pull/1008"
}
|
gharchive/pull-request
|
Fix lax checking of filename against genericTemplateName in sections plugin
The sections plugin was performing a lax check on the section filename against the genericTemplateName option, using the includes() method to see if the relativePathFromSections argument includes the genericTemplateName option value anywhere within it.
This meant files such as src/sections/collection-template.liquid were compiled to dist/sections/collection-template.liquid.liquid and were not being updated on Shopify when changes were made locally with yarn start running.
This fix compares the basename of the relativePathFromSections argument instead.
Checklist
For contributors:
[ ] I have updated the docs to reflect these changes, if applicable.
For maintainers:
[ ] I have :tophat:'d these changes.
I've signed the CLA - not sure what I need to do to get the check to run again.
Just checking -- what is the status of this PR? I can see it's been merged into Master but running
yarn upgrade --latest @shopify/slate-tools
doesn't seem to update the package.json or update any dependencies...
@Maurice-JB Working on it (This week). I'm fiddling with auto deploy configurations.
|
2025-04-01T06:37:32.998387
| 2023-01-19T13:23:21
|
1549152080
|
{
"authors": [
"Milly",
"Shougo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2666",
"repo": "Shougo/ddu.vim",
"url": "https://github.com/Shougo/ddu.vim/pull/45"
}
|
gharchive/pull-request
|
fix types in help
Fix types in help on multiple commits.
If there are unnecessary changes, I will delete it.
Merged.
|
2025-04-01T06:37:33.017406
| 2022-02-24T18:42:27
|
1149626562
|
{
"authors": [
"Shresht7"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2667",
"repo": "Shresht7/Screen-Recorder",
"url": "https://github.com/Shresht7/Screen-Recorder/issues/42"
}
|
gharchive/issue
|
Add FFmpeg WASM support
[x] Add FFmpeg.wasm
[ ] Video File-Converter (GIF)
[ ] Video Editing Tools (trim etc)
SharedArrayBuffer is not available by default.
The cost of implementing this library is way higher than the benefits it provides. Maybe a separate web-app?
Shelved for now.
|
2025-04-01T06:37:33.018247
| 2024-04-29T00:01:52
|
2267889076
|
{
"authors": [
"Shu-AFK",
"hvostov"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2668",
"repo": "Shu-AFK/Datatransmission",
"url": "https://github.com/Shu-AFK/Datatransmission/pull/44"
}
|
gharchive/pull-request
|
Multiple clients, changed thread func.
Now it should work. Also added things you said in comments to the prev pr.
I'll check when I'm home
|
2025-04-01T06:37:33.022063
| 2020-10-17T05:51:50
|
723677739
|
{
"authors": [
"Shubham0812",
"fredmurakawa"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2669",
"repo": "Shubham0812/Animatify-ios",
"url": "https://github.com/Shubham0812/Animatify-ios/pull/35"
}
|
gharchive/pull-request
|
Add the capability to open Github Repository inside app #28
This PR solves issue #28 and add improvements to CodeSymbolView
@fredmurakawa Can you do a pull from the dev branch and push again, there is a conflict due to the color changes.
Also, the text in the Open Github repo is too light and hard to read, do fix that too 👍🏻
Good job @fredmurakawa Merged the code. 😄👍🏻
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.