added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:37:28.585408
| 2023-12-17T11:33:08
|
2045182311
|
{
"authors": [
"Madouura",
"PIPIPIG233666"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2448",
"repo": "ROCmSoftwarePlatform/MIOpenGEMM",
"url": "https://github.com/ROCmSoftwarePlatform/MIOpenGEMM/issues/62"
}
|
gharchive/issue
|
Is this deprecated?
From what I can see, this hasn't been moved to the ROCm organization like the others, nor has it been updated past 5.5.0.
The last commit was also more than a year ago.
https://github.com/ROCm/MIOpen/commit/dd23f4e8c0783027128f95320c5af58982296d23#diff-b335630551682c19a781afebcf4d07bf978fb1f8ac04c6bf87428ed5106870f5
looks like the answer is yes
|
2025-04-01T06:37:28.586868
| 2018-05-14T18:45:44
|
322935642
|
{
"authors": [
"cbiest"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2449",
"repo": "ROIcoin/ROIcoin",
"url": "https://github.com/ROIcoin/ROIcoin/issues/63"
}
|
gharchive/issue
|
problem setting up pool
Hey I followed the instructions from here. https://www.youtube.com/watch?v=dR2fgqCSJ4Y
I followed the directions, made a text file copied and pasted the binary code from the optiminer website, pasted my wallet address where It said to put it then I saved it as a .bat file. What am I missing here?
Look g hobson I'm sorry I set you off, but I can't understand code because I'm not a coder. it's not hard to understand it. It just took finding out where I needed to run the dang code.
|
2025-04-01T06:37:28.590333
| 2022-03-09T16:23:19
|
1164158622
|
{
"authors": [
"EvilAnsh",
"rasenss"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2450",
"repo": "ROM-builders/temporary",
"url": "https://github.com/ROM-builders/temporary/pull/7477"
}
|
gharchive/pull-request
|
santoni: Build Spark-Rom (Bringup)
Build Spark-Rom for Santoni.
Don't build 2 rom in one time
How many time i asking to you
I'm sorry, I will not do it again
On Thu, Mar 10, 2022, 11:32 rasenss @.***> wrote:
Don't build 2 rom in one time
How many time i asking to you
—
Reply to this email directly, view it on GitHub
https://github.com/ROM-builders/temporary/pull/7477#issuecomment-1063694784,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AVHPLFROMWYNUSFZFZS726DU7GF5VANCNFSM5QKBC2UA
.
You are receiving this because you authored the thread.Message ID:
@.***>
|
2025-04-01T06:37:28.664427
| 2019-07-27T14:01:49
|
473641100
|
{
"authors": [
"johnnygroovy",
"teromene"
],
"license": "unlicense",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2451",
"repo": "RSS-Bridge/rss-bridge",
"url": "https://github.com/RSS-Bridge/rss-bridge/issues/1241"
}
|
gharchive/issue
|
Bridge request for Dave's Trailer
Hi !
I've made a bridge for Dave's Trailer, but I'm not a programmer so it's probably written badly. Plus I'm not used to github so I've no idea how to submit a new bridge. I guess it's better to put it here than to keep it for myself, but if it's not, feel free to erase this message ! :)
https://pastebin.com/XHL8Q9U2
Hello,
Thanks for the bridge ! I can merge it manually, or you can open a pull request and I will be happy to include it in the project .
Since there's a beginning for everything, I think I did my first pull request ever.
Closed by #1246
|
2025-04-01T06:37:28.709215
| 2023-10-07T12:53:00
|
1931341210
|
{
"authors": [
"KittoHub",
"RVC-Boss",
"vorob1",
"xden2331"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2452",
"repo": "RVC-Project/Retrieval-based-Voice-Conversion-WebUI",
"url": "https://github.com/RVC-Project/Retrieval-based-Voice-Conversion-WebUI/issues/1392"
}
|
gharchive/issue
|
Unable to clean audio
I'm getting this error:
20231007_154555008.wav.reformatted.wav->Traceback (most recent call last):
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\core\audio.py", line 155, in load
context = sf.SoundFile(path)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\soundfile.py", line 658, in init
self._file = self._open(file, mode_int, closefd)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\soundfile.py", line 1216, in _open
raise LibsndfileError(err, prefix="Error opening {0!r}: ".format(self.name))
soundfile.LibsndfileError: Error opening 'D:\Voice\TEMP/20231007_154555008.wav.reformatted.wav': System error.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\Voice\infer\modules\uvr5\modules.py", line 73, in uvr
pre_fun.path_audio(
File "D:\Voice\infer\modules\uvr5\preprocess.py", line 61, in path_audio
) = librosa.core.load( # 理论上librosa读取可能对某些音频有bug,应该上ffmpeg读取,但是太麻烦了弃坑
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\util\decorators.py", line 104, in inner_f
return f(**kwargs)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\core\audio.py", line 174, in load
y, sr_native = __audioread_load(path, offset, duration, dtype)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\core\audio.py", line 198, in _audioread_load
with audioread.audio_open(path) as input_file:
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\audioread_init.py", line 127, in audio_open
return BackendClass(path)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\audioread\rawread.py", line 59, in init
self._fh = open(filename, 'rb')
FileNotFoundError: [Errno 2] No such file or directory: 'D:\Voice\TEMP/20231007_154555008.wav.reformatted.wav'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\core\audio.py", line 155, in load
context = sf.SoundFile(path)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\soundfile.py", line 658, in init
self._file = self._open(file, mode_int, closefd)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\soundfile.py", line 1216, in _open
raise LibsndfileError(err, prefix="Error opening {0!r}: ".format(self.name))
soundfile.LibsndfileError: Error opening 'D:\Voice\TEMP/20231007_154555008.wav.reformatted.wav': System error.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\Voice\infer\modules\uvr5\modules.py", line 81, in uvr
pre_fun.path_audio(
File "D:\Voice\infer\modules\uvr5\preprocess.py", line 61, in path_audio
) = librosa.core.load( # 理论上librosa读取可能对某些音频有bug,应该上ffmpeg读取,但是太麻烦了弃坑
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\util\decorators.py", line 104, in inner_f
return f(**kwargs)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\core\audio.py", line 174, in load
y, sr_native = __audioread_load(path, offset, duration, dtype)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\librosa\core\audio.py", line 198, in _audioread_load
with audioread.audio_open(path) as input_file:
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\audioread_init.py", line 127, in audio_open
return BackendClass(path)
File "C:\Users\babud\AppData\Local\Programs\Python\Python310\lib\site-packages\audioread\rawread.py", line 59, in init
self._fh = open(filename, 'rb')
FileNotFoundError: [Errno 2] No such file or directory: 'D:\Voice\TEMP/20231007_154555008.wav.reformatted.wav'
would you mind uploading the audio that you want to get the clean version for us for checking?
sg10406F.zip
For example this one, want to clean echo from it.
@vorob1 @RVC-Boss
I had the same issue error I have been getting. I found the reason why its not working.
You have to make sure the input file name and input folder name does not have space. I put underscore between the space of the input mp3 file and folder and voila sucess.
Example:
Current file name: ABC Songs.mp3
Change file name to: ABC_Songs.mp3 (put a underscore or remove spaces)
Do the same with the main input folder if you have just remove it or add underscore then it works.
For those using Mac, try
brew install ffmpeg fprobe
Works on M2.
|
2025-04-01T06:37:28.720298
| 2017-03-22T17:36:36
|
216150552
|
{
"authors": [
"lubebul",
"tmylk"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2453",
"repo": "RaRe-Technologies/smart_open",
"url": "https://github.com/RaRe-Technologies/smart_open/pull/117"
}
|
gharchive/pull-request
|
Make #107 compatible and found HttpReadStream resource leak
Hello, I'm Ruo-Chun Tzeng, who'd like to participant in GSoC 2017.
As a test, I was assigned to fix #112.
There're 2 changes:
I add unit tests for checking file-descriptor leak of local compressed files
and caught gzip, bz2 are not properly closed on python 2.6.9 (failed my file-closing test)
I do a quick fix to make #107 compatible with make_closing function. so that HttpReadStream work with compressed formats
However, I found the HttpReadStream is not properly closed, the contents is still readable after closing the stream.with smart_open.smart_open('https://github.com/RaRe-Technologies/smart_open/blob/master/smart_open/tests/test_data/crlf_at_1k_boundary.warc.gz', 'r') as fin:
print(fin.read())
fin.read()
Thanks for investigating closing but the most important of this PR is that new unit tests need to be added when some gzipped/bz2 file is opened via http.
OK, my bad, I've add unit tests for read/write mode tests and close tests for bz/gz2 over http.
Also, since there's file-handle leak on
local bz/gz2(python2.6.9) and
bz/gz2 over http,
the unit test for file-handle leak tests are commented.
@lubebul Please see tests and code in #112
Fixed in #112
|
2025-04-01T06:37:28.723518
| 2014-10-21T11:09:03
|
46380971
|
{
"authors": [
"GLloydSC",
"RaYell"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2454",
"repo": "RaYell/selenium-webdriver-extensions",
"url": "https://github.com/RaYell/selenium-webdriver-extensions/issues/15"
}
|
gharchive/issue
|
IWebElement.FindElement(By by) method can't take Selenium.WebDriver.Extentions.JQuerySelector
I think the title says it all:
IWebDriver.FindElement(Selenium.WebDriver.Extentions.JQuerySelector by) is ok
IWebElement.FindElement(Selenium.WebDriver.Extentions.JQuerySelector by) is not ok
I've found a possible workaround that would allow to implement the similar behavior as invoking FindElement/FindElements on IWebElement.
The way this could work is I could add a wrapper type on RemoteWebElement (the base type implementing IWebElement interface) and on this wrapper store also information about the selector that has been used to locate it. Having this information it should be possible to generate a new selector taking into account that the scope for the new search has been narrowed by previous query.
The functionality is implemened as part of the https://github.com/RaYell/selenium-webdriver-extensions/pull/21
It's scheduled to be released with version 1.2. It comes with several limitations which you can find in the pull request comments.
|
2025-04-01T06:37:28.736996
| 2018-07-19T20:39:10
|
342880785
|
{
"authors": [
"alexey-morozov",
"gstoner",
"jlgreathouse"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2455",
"repo": "RadeonOpenCompute/ROCm",
"url": "https://github.com/RadeonOpenCompute/ROCm/issues/462"
}
|
gharchive/issue
|
OpenCL shared library overrides signal handling
Hi,
we have a problem to use ROCm (1.8.151) OpenCL shared library (Ubuntu 16.04.4 LTS) in our own development environment where we make use of POSIX signal handlers. When calling clGetPlatformIDs function the signal handling of our application process gets overridden that causes problems for the application and finally makes it crashing. A simple CPP example below demonstrates the problem. Are there any specific reasons for the shared library to override the signal handling of the loading process? How can we avoid that?
Thanks for help.
// test.cpp
#include <dlfcn.h>
#include <signal.h>
#include <stdio.h>
int main()
{
auto lib = dlopen("/opt/rocm/opencl/lib/x86_64/libOpenCL.so", RTLD_LAZY);
auto function = reinterpret_cast<int (*)(unsigned, void *, unsigned *)>(dlsym(lib, "clGetPlatformIDs"));
unsigned n = 0;
auto result = function(0, 0, &n); // this function call overrides the current process signal handling
printf("generating int3 trap\r\n");
asm("int3"); // normally this induces a trap but here the process will exit silently
}
// compilation: g++ -std=c++11 -o test test.cpp -ldl
// execution: ./test
I have engineering looking at this. They will be getting back to you about this
any news on the issue?
We're tracking this internally. I believe we've identified the issue, and I have a potential workaround that has not gone through review, validation, or testing.
Would you be willing to do a custom build of our OpenCL runtime, apply a highly experimental patch, and see if it fixes your issue without causing you any other problems?
Hi @alexey-morozov
Could you try running the following commands to make a custom OpenCL build with a simple patch in place? Sorry for using sed to do this patch, :)
mkdir ~/test_opencl
cd ~/test_opencl
mkdir -p ~/bin
curl https://storage.googleapis.com/git-repo-downloads/repo > ~/bin/repo
chmod a+x ~/bin/repo
~/bin/repo init -u https://github.com/RadeonOpenCompute/ROCm-OpenCL-Runtime.git -b master -m opencl.xml
~/bin/repo sync
cd opencl
sudo apt-get install ocaml ocaml-findlib python-z3 git-svn
mkdir -p build && cd build
sed -i '353i // In OpenCL applications that call LC, the end-user may want to receive' ../opencl/compiler/llvm/lib/Support/Unix/Signals.inc
sed -i '354i // a signal that is sent asynchronously, and thus would not be re-raised' ../opencl/compiler/llvm/lib/Support/Unix/Signals.inc
sed -i '355i // even after returning from the compilation path. Raise it here so that' ../opencl/compiler/llvm/lib/Support/Unix/Signals.inc
sed -i '356i // the original handler is called.' ../opencl/compiler/llvm/lib/Support/Unix/Signals.inc
sed -i '357i raise(Sig);' ../opencl/compiler/llvm/lib/Support/Unix/Signals.inc
cmake -DCMAKE_BUILD_TYPE=RelWithDebInfo ..
make -j `nproc`
At this point, you can try using this OpenCL runtime by putting the ~/test_opencl/opencl/build/lib directory in your LD_LIBRARY_PATH. For example: LD_LIBRARY_PATH=~/test_opencl/opencl/build/lib ./test
Hi @alexey-morozov
Have you had a chance to try this yet?
Hi @jlgreathouse
Yes, I've tried it and our IDE still crashed. But we've found a workaround and I would consider the issue closed.
Thanks in any case!
|
2025-04-01T06:37:28.768428
| 2022-12-15T10:00:04
|
1498134574
|
{
"authors": [
"RafaelGB",
"zeltak"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2456",
"repo": "RafaelGB/obsidian-db-folder",
"url": "https://github.com/RafaelGB/obsidian-db-folder/issues/668"
}
|
gharchive/issue
|
[Bug]: illegal character error in a tag column
Contact Details
No response
What happened?
I get the following error after defining a tag column
any clue?
What platform were you using?
Desktop
Version of the plugin
3.0.1
Relevant log output
No response
Are you using the latest version of the plugin in your Obsidian vault?
[X] I have verified that I am on the latest version
Are you check if there is a similar issue?
[X] I have verified that there are not similar issues
Hello @zeltak , I am glad the the new error handler print those errors =)
You have an invalid character with the value of your label
Open your markdown file and modify
label: "\-" with label "-"
If the error persist I will investigate it better
thx @RafaelGB !
yeah that did the trick! still dont know which note had that invalid character (I tried searching :)) but now all is well
thx so much a bunch for you continuous amazing work that makes my (and many others!) life that much more enjoyable :D
best
Z
Thanks for your feedback too! it makes this plugin so much easier to maintain
I will try to check this kind of illegal characters with the plugin, so you dont have to worry about
Commit d6dd0df9
|
2025-04-01T06:37:28.772131
| 2023-02-02T11:54:22
|
1567904857
|
{
"authors": [
"RafaelGB",
"Shaeeeeeee"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2457",
"repo": "RafaelGB/obsidian-db-folder",
"url": "https://github.com/RafaelGB/obsidian-db-folder/issues/803"
}
|
gharchive/issue
|
[Bug]:
Contact Details
No response
What happened?
There is always a blank line in front of the bullet list. Even though I try to delete the comma in the front, it's useless because the comma soon comes back.
What platform were you using?
Desktop
Version of the plugin
3.3.0
Relevant log output
No response
Are you using the latest version of the plugin in your Obsidian vault?
[X] I have verified that I am on the latest version
Are you check if there is a similar issue?
[X] I have verified that there are not similar issues
this issue will be fixed with the next version :)
|
2025-04-01T06:37:28.807395
| 2022-01-03T13:25:56
|
1092511012
|
{
"authors": [
"Rahil627"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2458",
"repo": "Rahil627/rathewolf",
"url": "https://github.com/Rahil627/rathewolf/issues/1"
}
|
gharchive/issue
|
test github issues and determine if it’s worth using
is it worth it? or stick to a plain text file?
can assign things to issues: people, labels, projects, milestones, pull requests(!)
i definitely don’t need assignees—it’s just me!
i don’t need labels either, that’s wayyyy too much!
categorizing issues by projects... hmmmm... i’d rather have a separate github repo for each project
no milestones neither, just free-flow~~
linking pull requests, i think, is for collaborative projects...
soooo, no. no thank you.
...but, you know, it’s cool to have this sorta dashboard, that notifies or shows how many issues, or project plans, etc... Maybe that’s an app I need to make: a text-file based simple dashboard
|
2025-04-01T06:37:28.816809
| 2019-08-06T10:40:49
|
477304658
|
{
"authors": [
"RahulShaw",
"hila3000",
"newtrojan",
"satrox28",
"techguyj"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2459",
"repo": "RahulShaw/LinuxAcademy-DL",
"url": "https://github.com/RahulShaw/LinuxAcademy-DL/issues/16"
}
|
gharchive/issue
|
The same video is repeatedly downloaded
This worked on August 4th and now suddenly stopped working. Tested from multiple computers.
PowerShell_transcript.qfpEhXoA.20190806132901.txt
$ ll -lah
total 61M
drwxr-xr-x 1 user 197121 0 Aug 6 13:32 ./
drwxr-xr-x 1 user 197121 0 Aug 6 13:30 ../
-rw-r--r-- 1 user 197121 21M Aug 6 13:31 '1. About the Course Author.mp4'
-rw-r--r-- 1 user 197121 21M Aug 6 13:31 '2. How to Use the Linux Academy Cloud Playground for this Course.mp4'
-rw-r--r-- 1 user 197121 21M Aug 6 13:32 '3. Course `Syllabus.mp4'
will it be fixed soon?. I am also facing the same issue.
Currently, I do not have an active LinuxAcademy subscription. If someone can voluntarily share their subscription with me, I shall work on the issues/requests. I'm sorry :(
Currently, I do not have an active LinuxAcademy subscription. If someone can voluntarily share their subscription with me, I shall work on the issues/requests. I'm sorry :(
Is there a way to send a PM
You may
Currently, I do not have an active LinuxAcademy subscription. If someone can voluntarily share their subscription with me, I shall work on the issues/requests. I'm sorry :(
Is there a way to send a PM
You may reach me here https://www.facebook.com/TheRahulShaw
@RahulShaw I've sent you a message on facebook, please check.
Please update the youtube-dl package and that will fix the problem.
pip install youtube-dl --upgrade
|
2025-04-01T06:37:28.824939
| 2024-08-19T10:49:50
|
2473057468
|
{
"authors": [
"RahulVadisetty91"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2460",
"repo": "RahulVadisetty91/Free-Auto-GPT",
"url": "https://github.com/RahulVadisetty91/Free-Auto-GPT/pull/1"
}
|
gharchive/pull-request
|
bug_fixes.py
Fix various issues and enhance code stability
Resolved the following issues:
Import Errors:
pydantic Import Error: Added the pydantic module to the project’s dependencies to resolve the import error. This allows the code to correctly utilize pydantic for data validation and settings management.
faiss Import Error: Ensured faiss is properly installed and included in the project. This fixes the issue with the faiss import, which is used for vector search functionalities.
playwright.async_api Import Error: Installed the playwright package and fixed the import path for playwright.async_api, ensuring compatibility with the code that uses Playwright for web scraping.
Import Path Corrections:
langchain.experimental.autonomous_agents.autogpt.agent Import Error: Corrected the import path for the AutoGPT class from the langchain library. This resolves issues with module resolution, ensuring that the AutoGPT class is properly imported and used in the code.
Code Duplication:
Defined Constant for File Name: Replaced multiple instances of the hardcoded file name "cookiesBing.json" with a constant. This improves code maintainability and readability by avoiding magic strings and making future changes easier.
This pull request addresses several issues to enhance the stability and maintainability of the code:
Resolved Import Errors:
pydantic and faiss: Added the missing dependencies for pydantic and faiss to ensure proper module resolution and functionality. This fixes the import errors and allows the code to use these modules effectively.
playwright.async_api: Installed and correctly configured the playwright package. This resolves the import issue and ensures that the Playwright-based web scraping functionality works as expected.
Fixed Import Path:
AutoGPT Import Path: Updated the import path for the AutoGPT class from the langchain library. This corrects the issue with module resolution and ensures that AutoGPT is properly imported and used in the code.
Improved Code Maintainability:
Defined Constant for File Name: Replaced multiple instances of the hardcoded file name "cookiesBing.json" with a constant. This improves code readability and maintainability, making future modifications easier and avoiding magic strings.
Cleaned Up Code:
Fixed a redundant max_iterations parameter. This resolves a SonarLint warning and cleans up the code, enhancing its clarity and quality.
These changes aim to improve the overall code quality, ensure proper functionality, and make the codebase more maintainable.
|
2025-04-01T06:37:28.833758
| 2020-09-27T12:56:50
|
709735180
|
{
"authors": [
"RailKill"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2461",
"repo": "RailKill/tetraho",
"url": "https://github.com/RailKill/tetraho/pull/26"
}
|
gharchive/pull-request
|
Merging for v1.1.0 release
Minor patch which addresses the bugs and design problems from #mizjam1 with a bunch of optimizations, refactoring and automated testing. There are no new features since v1.0.0, no change to level design or gameplay. This PR is solely to fix and patch things that were supposed to be working or be implemented in the first place. By solving these fundamental issues, it allows the project to be exported as a release build with the proper framework for future development.
also closes #25
|
2025-04-01T06:37:28.848616
| 2024-12-05T09:36:21
|
2719872171
|
{
"authors": [
"Nicostratus",
"mnaoumov"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2462",
"repo": "RainCat1998/obsidian-custom-attachment-location",
"url": "https://github.com/RainCat1998/obsidian-custom-attachment-location/issues/123"
}
|
gharchive/issue
|
[BUG] The action “Collect attachments...” does bypass the “renameOnlyImages” setting.
Description
First of all, I would like to thank you for creating this plugin.
It is simply awesome.
Recently while organizing my vault I noticed the following thing, I am not sure if it is intended:
Despite setting the following configuration:
“renameOnlyImages": true,
“renameCollectedFiles": true,
When you run the “Collect attachments for current note” action, the plugin renames all attachments, while theoretically it should rename only images.
Other attachments for example “zip”, “pdf” etc should have the file name unchanged
Steps to Reproduce
Set following settings : “renameOnlyImages": true, “renameCollectedFiles": true,
Run “Collect attachments for current note” action on note that contains ZIP or PDF attachments
Notice that all files will be renamed
Expected Behavior
Other attachments for example “zip”, “pdf” etc should have the file name unchanged
Actual Behavior
The plugin renames all attachments, while theoretically it should rename only images.
Environment Information
Plugin Version: v4.29.1
Obsidian Version: v1.7.7
Operating System: Windows 11
Attachments
Please attach a video showing the issue
Please attach a sample vault where the issue can be reproduced
Confirmations
[X] I attached a video showing the issue
[X] I attached a sample vault where the issue can be reproduced
[X] I have tested the issue with the latest version
[X] I have checked for existing issues
Fixed in https://github.com/RainCat1998/obsidian-custom-attachment-location/releases/tag/4.31.1
|
2025-04-01T06:37:28.849593
| 2014-07-25T17:59:47
|
38754942
|
{
"authors": [
"Offerel",
"christian-putzke"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2463",
"repo": "RainLoop/rainloop-webmail",
"url": "https://github.com/RainLoop/rainloop-webmail/issues/255"
}
|
gharchive/issue
|
CardDAV contact images
It would be great if you could add the CardDAV contact images in addition to the gravatar option.
+1
|
2025-04-01T06:37:28.886445
| 2020-12-12T16:56:44
|
764050224
|
{
"authors": [
"200sc",
"AndrewMD5",
"lynn"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2465",
"repo": "RainwayApp/bebop",
"url": "https://github.com/RainwayApp/bebop/issues/71"
}
|
gharchive/issue
|
Clarification on byte vs uint8
Intuitively byte == uint8, however in Writing Bops, uint8 is not listed as a type, so its not allowed. So far so good.
The msgpack comparison example uses the uint8 type: i.e. uint8 iNT0; // "int0": 0,.
Is uint8 a valid builtin type? Is the example outdated? Do the docs just need a clarification 'byte' or 'uint8'?
Bebop supports type aliases so you can use either byte or uint8 and both represent an unsigned 8-bit integer.
So to bring this to resolution, should uint8 be added to the Writing Bops guide? Or should a doc be written explaining type aliases (which I assume are a builtin only feature, in which case it would ideally list the built in aliases that exist)?
For additional context, it looks like in the existing codebase the -only- type alias is uint8 = byte (https://github.com/RainwayApp/bebop/blob/8c65bcb1271168b67ea308ab52556e34c488fe66/Core/Parser/Extensions/TypeExtensions.cs#L16)
I updated the schema language documentation to note that uint8 is an alias for byte.
There is another kind of alias-like behavior, where you can write either array[T] or T[] to name the “array of T” type. Indeed, uint8 = byte is the only base type alias right now, and there is not yet any way to specify custom type aliases.
This alias exists because sometimes it makes sense to talk about bytes as “the smallest memory unit”, like when talking about a byte[], but other times it makes sense to think about a byte as an “8-bit unsigned numeric type”, and in that case the name uint8 is provided to match all the other ones.
|
2025-04-01T06:37:28.898134
| 2024-01-21T13:21:22
|
2092583842
|
{
"authors": [
"Rakesh9100",
"amansgith"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2466",
"repo": "Rakesh9100/CalcDiverse",
"url": "https://github.com/Rakesh9100/CalcDiverse/issues/421"
}
|
gharchive/issue
|
[Bug] Font Family inconsistentcy between Home and Contributor Page
Description
Problem
I have checked the Font Family of the navbar elements of both home page and Contributor page
And i've found that there is difference in font family applied of the navigation bar elements.
Screenshots
Home Page
Contributors Page
I have underlined the elements with red color
Checklist
[X] I have checked the existing issues.
[X] I follow Contributing Guidelines of this project.
[ ] I am a IWOC'24 contributor.
[ ] I am a JWOC'24 contributor.
[X] I am a SWOC'24 contributor.
Code of Conduct
[X] I agree to follow this project's "Code of Conduct".
Please assign this issue to me because i have figured out the problem and can fix it quickly
Thank you
Proceed @amansgith
|
2025-04-01T06:37:28.902947
| 2024-02-19T15:18:32
|
2142610403
|
{
"authors": [
"AftabMankapure",
"Rakesh9100"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2467",
"repo": "Rakesh9100/CalcDiverse",
"url": "https://github.com/Rakesh9100/CalcDiverse/issues/598"
}
|
gharchive/issue
|
[Bug] Remove border radius from navbar
Description
Remove border radius form navbar
Screenshots
Checklist
[X] I have checked the existing issues.
[X] I follow Contributing Guidelines of this project.
[ ] I am a IWOC'24 contributor.
[X] I am a JWOC'24 contributor.
[ ] I am a SWOC'24 contributor.
Code of Conduct
[X] I agree to follow this project's "Code of Conduct".
Pls upload the issue supporting screenshot properly @AftabMankapure
Current
New
Proceed @AftabMankapure
|
2025-04-01T06:37:28.918274
| 2024-05-29T10:58:44
|
2323022993
|
{
"authors": [
"finanalyst",
"thoughtstream"
],
"license": "Artistic-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2468",
"repo": "Raku/RakuDoc-GAMMA",
"url": "https://github.com/Raku/RakuDoc-GAMMA/issues/30"
}
|
gharchive/issue
|
Embedded Format Codes in Definitions
@thoughtstream @lizmat tl;dr Can a =defn have format codes embedded in the term term?
format code is a markup code without side-effects.
I am implementing =defn and =numdefn and run across a limitation I think is not natural.
(RakuDoc v1 using $=pod did not implement =defn well)
At a minimum, we have
=defn A term
The expansion
At the present, we can have embedded format codes in the expansion but not in the term, that is:
=defn A term
The expansion may have U<embedded> format codes at present.
Currently we cannot have
=defn A term with B<embedded> format codes
The expansion with or without format codes
@thoughtstream was it intended that the term should be without format codes?
@lizmat would it be possible to have numdefn have the same AST as defn? Currently, numdefn is treated as a normal block, meaning that a series of tests on the Paragraph elements needs to be run to determine what is in the term and what is in the expansion.
FYI I propose that for =alias TERM EXPANSION TERM is most definitely without format codes!
@lizmat I propose that even if the term and expansion parts are Str, the AST for a defn & numdefn is
RakuAST::Doc::Block.new(
type => "defn",
abbreviated => True,
paragraphs => (
RakuAST::Doc::Paragraph.new( ... )
RakuAST::Doc::Paragraph.new( ... )
)
...
My expectation was that the "term" component of a =defn should not have formatting.
I can see why formatting might be useful within a term, and it may seem odd to end-users that
it's one of the few places where you can write ordinary text, but can't use formatting codes.
However, the term component also acts as a target for L<defn:...> links, and it would seem better
if that target is unambiguous.
I mean, if we allow:
=defn A term with B<embedded> formatting codes
The definition with or without formatting codes
...then does the corresponding L<defn:...> also have to be: L<defn:"A term with B<embedded> formatting codes>"?
It seems simpler (at least initially) just to specify that the term of a =defn is treated as a if it were a V<> literal string,
with any apparent formatting codes ignored (and perhaps warned about). The same rule should probably also apply to the term component of an inline D<>.
Besides, the idea was that the term component automatically gets special formatting anyway. That is:
=defn A term
The definition with or without formatting codes
A D<placement directive> inserts something into the document at that point.
...should be rendered in such a way that the term is easily distinguished as the thing being defined.
Something like:
A term
The definition with or without formatting codes
A placement directive inserts something into the document at that point.
So the TL;DR is: No, I didn't intend the term component of a =defn to allow formatting codes.
But, as always, I'm happy to discuss the issue further.
@thoughtstream & @lizmat That's the way =defn / D<> are implemented. Closing as resolved. I'll add a clarification as suggested
closing as resolved
|
2025-04-01T06:37:28.971065
| 2019-06-20T06:59:53
|
458384744
|
{
"authors": [
"Highwall2016",
"akelad"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2469",
"repo": "RasaHQ/rasa",
"url": "https://github.com/RasaHQ/rasa/issues/3828"
}
|
gharchive/issue
|
Could I only use MappingPolicy in my config ?
in my config.yml just like:
Configuration for Rasa NLU.
language: "zh"
pipeline:
name: "MitieNLP"
model: "total_word_feature_extractor.dat"
name: "MitieTokenizer"
name: "MitieEntityExtractor"
name: "EntitySynonymMapper"
name: "RegexFeaturizer"
name: "MitieFeaturizer"
name: "SklearnIntentClassifier"
name: "MitieIntentClassifier"
Configuration for Rasa Core.
policies:
name: MappingPolicy
Please ask these kind of questions in our forum
|
2025-04-01T06:37:28.974152
| 2020-12-02T23:29:46
|
755688025
|
{
"authors": [
"rgstephens",
"wochinge"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2470",
"repo": "RasaHQ/rasa",
"url": "https://github.com/RasaHQ/rasa/pull/7440"
}
|
gharchive/pull-request
|
publish retries
Proposed changes:
Added publish retry code similar to Rabbit
Status (please check what you already did):
[ ] added some tests for the functionality
[ ] updated the documentation
[ ] updated the changelog (please check changelog for instructions)
[ ] reformat files using black (please check Readme for instructions)
@rgstephens Can we get this merged?
@wochinge Have you seen this gitleaks secrets warning before?
|
2025-04-01T06:37:28.977450
| 2018-09-11T11:51:19
|
359010887
|
{
"authors": [
"akelad",
"neerajb1"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2471",
"repo": "RasaHQ/rasa_core",
"url": "https://github.com/RasaHQ/rasa_core/issues/977"
}
|
gharchive/issue
|
How to retrieve chat conversasion from mongo tracker
Rasa Core version:.10.4
Python version: 3.5
Operating system (windows, osx, ...):windows
Issue:
I am able to save tracker on mongo, but data store in mongo is only sender id and tracker_id. I want tos store conversation history on mongo. Please help me to get chat conversation history
Content of domain file (if used & relevant):
This is more of a question for the forum, and I believe you've already posted there too? In any case, it is saving the conversation history, that's what the tracker is: https://github.com/RasaHQ/rasa_core/blob/master/rasa_core/tracker_store.py#L144
|
2025-04-01T06:37:28.984589
| 2018-11-27T11:23:31
|
384747911
|
{
"authors": [
"Ghostvv",
"akelad",
"jstriebel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2472",
"repo": "RasaHQ/rasa_nlu",
"url": "https://github.com/RasaHQ/rasa_nlu/pull/1555"
}
|
gharchive/pull-request
|
EmbeddingIntentClassifier: allow config overwrite
Previous problem:
Setting epochs for training via the config.yml was not possible.
Proposed changes:
This PR refactors the use of config parameters in EmbeddingIntentClassifier and allows to use the component config (e.g. via config.yml). The previous _load_params implementation seemed quite doubled, as the superclass already does the logic of merging the config with the defaults. Also doubling all settings as direct attributes seemed superfluous to me. Therefore all usages of them are refactored to use self.component_config. The only exceptions and notable occurances are:
self.num_neg
This is updated during training. Not sure of a variable in train would be fine, I sticked to the safe version and kept it as an attribute.
self.intent_tokenization_flag
To keep the early warning message I kept this in _load_params.
evaluate_every_num_epochs
The logic moved from _load_visual_params directly to _train_tf, as no side-effects happen.
Status (please check what you already did):
[x] made PR ready for code review
[ ] added some tests for the functionality
[ ] updated the documentation
[ ] updated the changelog
Not sure what to do about the long lines. I tested this locally, worked for me.
PS: An alternative would be to simply call self._load_params(self.component_config), but I thought it would be a good case to remove the redundancies.
@jstriebel what do you mean setting epochs wasn't possible? I've definitely done that in the config.yml file before and it worked fine
Hi @akelad, it surely worked before, but it is not possible with version 0.14.0a1. I think this regression was introduced in #1456. There are many possibilities how to resolve this, I found this version to be the cleanest.
For reference the nlu config I tested this with:
language: "de"
pipeline:
- name: "intent_featurizer_count_vectors"
- name: "intent_classifier_tensorflow_embedding"
embed_dim: 20
num_neg: 20
epochs: 100
Thanks a lot for the fix, but there is another PR that fixes it by passing component_config to load_params method: https://github.com/RasaHQ/rasa_nlu/pull/1547
My bad, I was thinking of Core btw 🙃
|
2025-04-01T06:37:29.036379
| 2018-08-04T10:06:54
|
347606607
|
{
"authors": [
"FeiCoding",
"Rayhane-mamah"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2473",
"repo": "Rayhane-mamah/Tacotron-2",
"url": "https://github.com/Rayhane-mamah/Tacotron-2/issues/120"
}
|
gharchive/issue
|
Eva data of WaveNet is really bad, and the average error always keep around 7 for 140k steps
Is this normal?
or maybe some bugs in Wavenet?
dose anyone get the similar problems?
I use GTX 1080 to train,
the wavenet_batch_size = 2
Hello @FeiCoding thanks for reaching out.
I fixed this in last commit, it is merely a synthesis time bug. Also please keep in mind that for "raw" mode, you need to train your model for about 1 million steps to get human like speech. (yeah it takes time to converge..)
This is a similar issue to yours:
https://github.com/Rayhane-mamah/Tacotron-2/issues/57
So I am closing this now, if your problem still persists with latest commit, feel free to reopen.
|
2025-04-01T06:37:29.043342
| 2017-04-22T02:35:51
|
223534603
|
{
"authors": [
"LucasPMagno",
"Rayzr522",
"abyssvi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2474",
"repo": "Rayzr522/SharpBot",
"url": "https://github.com/Rayzr522/SharpBot/issues/32"
}
|
gharchive/issue
|
Random and funny outputs
something like this:
https://cdn.discordapp.com/attachments/281141184108167178/305169127222280204/unknown.png
https://cdn.discordapp.com/attachments/281141184108167178/305169006950744066/unknown.png
https://cdn.discordapp.com/attachments/281141184108167178/305168920145428481/unknown.png
@abyssvi what're your thoughts on this?
I'm not in favor of it for personal flavor reasons. I prefer consistent and obvious outputs. Worth noting that my usecase is largely utilitarian.
This would probably be best handled by only doing it on silly/fun commands, and leaving the serious commands to be serious.
I'm a total grouch, I know xD
There are much much better things to be done with SharpBot. I think this and #30 should be closed indefinitely.
|
2025-04-01T06:37:29.048475
| 2021-01-03T17:31:02
|
777682151
|
{
"authors": [
"llndqvst"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2475",
"repo": "RblSb/SyncTube",
"url": "https://github.com/RblSb/SyncTube/pull/15"
}
|
gharchive/pull-request
|
Add simple dockerfile
This just adds a really simple Dockerfile for those who wants to deploy SyncTube on docker.
Test by running docker build -t localhost/synctube . and docker run --rm -it -p 4200:4200 localhost/synctube
Didn't see that there already exists a pull request for adding a Dockerfile
#11
Didn't see that there already exists a pull request for adding a Dockerfile
#11
|
2025-04-01T06:37:29.105347
| 2020-06-27T08:46:02
|
646640078
|
{
"authors": [
"pankaj-ch",
"timdorr",
"tpenaranda"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2476",
"repo": "ReactTraining/react-router",
"url": "https://github.com/ReactTraining/react-router/issues/7457"
}
|
gharchive/issue
|
useHistory() return history object different for
Version
v5.2.0
Test Case
https://codesandbox.io/s/react-router-689yc
Steps to reproduce
Instead of BrowserRouter, use Router with customized history
then use the hook useHistory() and see the different object return by history
Expected Behavior
to return history object with properties
length
action
location
createHref
push
replace
go
goBack
goForward
block
listen
Actual Behavior
return properties from history object
action
location
createHref
push
replace
go
back
forward
listen
block
TLDR:
goBack and goForward methods are undefined
for history object when using custom history with Router component
Remove 'history' dependency from package.json file and that will do the trick (at least on your test case). @pankaj-ch
Only history 4.x works with Router 5.x.
|
2025-04-01T06:37:29.131389
| 2021-11-09T16:11:09
|
1048790875
|
{
"authors": [
"cursorsdottsx"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2477",
"repo": "ReactiveX/rxjs",
"url": "https://github.com/ReactiveX/rxjs/issues/6670"
}
|
gharchive/issue
|
Possibly better RxJS pipe typings?
I know it's unlikely for someone to use 9+ operators in a pipe... but this is really bugging me...
pipe.ts' typings seems pretty huge, redundant, and unnecessary, and I want to help reduce it so:
It's not huge, redundant, and unnecessary
It can take 9+ operators and still produce the correct type
I made a simple prototype with this working in this gist (updated)
Currently, it takes in any amount of unary functions, and does the following:
Validates if the unary functions are composable together (parameter and return types match)
Spits out an unary function of the first one's parameter and the last one's return type
If no arguments are present it spits out the identity function
If only one argument is present it returns that argument
Thoughts, anyone?
Let me know if this can be possibly added.
Uh oh, I've been testing my prototype and it seems to not handle the obvious case when there is only one operator 😅
No matter, I think it's an easy fix though
Possibly related issue: #4177 ?
|
2025-04-01T06:37:29.243063
| 2020-11-24T22:45:16
|
750121286
|
{
"authors": [
"Josca",
"jpavlick"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2478",
"repo": "RedHatInsights/expandjsonsmt",
"url": "https://github.com/RedHatInsights/expandjsonsmt/issues/9"
}
|
gharchive/issue
|
Support for mixed types in arrays of objects
Hello! I'm trying to use this SMT to transform some data I have in a Mongo Database into a Connect record. It works very well overall, but I have one field that looks like this:
[
{ "tagName" : "SourceFile", "tagValue" : "string", "tagGroup" : "none" },
{ "tagName" : "SourceFile", "tagValue" : 199, "tagGroup" : "none" },
{ "tagName" : "SourceFile", "tagValue" : [ "tag1", "tag2 ], "tagGroup" : "none" }
]
This fails with the following error:
org.apache.kafka.connect.errors.DataException: Invalid Java object for schema type STRING: class java.lang.Double for field: "tagValue"
at org.apache.kafka.connect.data.ConnectSchema.validateValue(ConnectSchema.java:245)
at org.apache.kafka.connect.data.Struct.put(Struct.java:216)
at org.apache.kafka.connect.data.Struct.put(Struct.java:203)
at com.redhat.insights.expandjsonsmt.DataConverter.convertFieldValue(DataConverter.java:35)
...
It seems to set the schema type for tagValue based on whatever it sees first. For example, if the array was the first thing it saw, I would get an error like Invalid Java obect for schema type ARRAY: class java.lang.Double....
Is this expected for how this SMT is currently designed? If not, is it something that it's possible to add support for? Thanks!
Hello @jpavlick, nice to see our tool is useful for you. It's some time ago I've implemented that. If I remember correctly, yes, schema is created dynamically from the first item, what is quite convenient in most cases, I think. I don't think there is any simple workaround how to support such "multi-type" field.
But to be honest, I don't think it's a good idea to have such data in database. I would rather define "tagValue" as a array in general so you can store: ["string"], ["199"] and ["tag1", "tag2"]. Then you will avoid this issue and probably some others, I guess. Likely it's not the only place where you are going to process this data and it's always much easier when you can expect the same type for each record. Regrds.
Thanks for the reply, and sorry it took me awhile to get back to you. We don't actually have access to the source data to be able to change the data at the source, and since it's valid json it seems like perhaps this SMT should be able to handle that use case.
That said, I recognize why it's difficult to do in Java, so I decided to create my own SMT to handle this specific case and coerce the singletons into arrays of strings, as per your suggestion. That fixed our issue, so I'll go ahead and close this issue. Thanks for the response!
|
2025-04-01T06:37:29.289634
| 2020-06-26T13:17:18
|
646256862
|
{
"authors": [
"joselsegura",
"quarckster"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2479",
"repo": "RedHatInsights/insights-results-smart-proxy",
"url": "https://github.com/RedHatInsights/insights-results-smart-proxy/issues/67"
}
|
gharchive/issue
|
GET /clusters/{clusterId}/report returns 404
smart-proxy returns 404 on GET /clusters/{clusterId}/report. The log contains an error:
{"level":"info","time":"2020-06-26T13:10:22Z","message":"Request received - URI: /api/insights-results-aggregator/v1/organizations/11789772/clusters, Method: GET"}
{"level":"info","time":"2020-06-26T13:10:22Z","message":"Handling response as a proxy"}
{"level":"debug","time":"2020-06-26T13:10:22Z","message":"Connecting to http://insights-results-aggregator:8080/api/insights-results-aggregator/v1/organizations/11789772/clusters "}
{"level":"info","time":"2020-06-26T13:10:23Z","message":"Request received - URI: /api/insights-results-aggregator/v1/clusters/254633ef-bd50-4504-91f7-3acf8d2cc83b/report, Method: GET"}
{"level":"error","error":"Item with ID ccx_rules_ocp.external.rules.nodes_kubelet_version_check.report/NODE_KUBELET_VERSION was not found in the storage","time":"2020-06-26T13:10:23Z","message":"handleServerError()"}
insights-results-aggregator receives a proxied request and replies without errors:
{"level":"info","time":"2020-06-26T13:10:23Z","message":"Request received - URI: /api/insights-results-aggregator/v1/organizations/11789772/clusters/254633ef-bd50-4504-91f7-3acf8d2cc83b/users/6089719/report, Method: GET"}
Apparently the problem is related to the lack of static content for this specific rule (ccx_rules_ocp.external.rules.nodes_kubelet_version_check).
Are you OK closing this issue?
Sure
|
2025-04-01T06:37:29.300498
| 2022-11-21T08:34:48
|
1457530673
|
{
"authors": [
"JiriPapousek",
"codecov-commenter"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2480",
"repo": "RedHatInsights/insights-results-smart-proxy",
"url": "https://github.com/RedHatInsights/insights-results-smart-proxy/pull/962"
}
|
gharchive/pull-request
|
Add missing token message when using jwt tokens
Description
Show correct status for missing token when authenticating with JWT tokens.
Fixes #736
Type of change
Bug fix (non-breaking change which fixes an issue)
Testing steps
Run the service locally.
Checklist
[ ] make before_commit passes
[ ] updated documentation wherever necessary
[ ] added or modified tests if necessary
[ ] updated schemas and validators in insights-data-schemas in case of input/output change
No unit tests?
I added them.
Codecov Report
Base: 60.43% // Head: 60.56% // Increases project coverage by +0.12% :tada:
Coverage data is based on head (fb408e1) compared to base (c6e05d6).
Patch coverage: 66.66% of modified lines in pull request are covered.
Additional details and impacted files
@@ Coverage Diff @@
## master #962 +/- ##
==========================================
+ Coverage 60.43% 60.56% +0.12%
==========================================
Files 23 23
Lines 3228 3231 +3
==========================================
+ Hits 1951 1957 +6
+ Misses 1056 1053 -3
Partials 221 221
Impacted Files
Coverage Δ
server/auth.go
66.99% <66.66%> (+1.99%)
:arrow_up:
content/content.go
77.77% <0.00%> (+0.82%)
:arrow_up:
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
:umbrella: View full report at Codecov.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
|
2025-04-01T06:37:29.307143
| 2023-11-01T09:52:56
|
1972048154
|
{
"authors": [
"codecov-commenter",
"kgaikwad",
"upadhyeammit"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2481",
"repo": "RedHatInsights/ros-backend",
"url": "https://github.com/RedHatInsights/ros-backend/pull/392"
}
|
gharchive/pull-request
|
Remove Pipfile.lock refs from codeql.yml
PR Title :boom:
Please title this PR with a summary of the change, along with the JIRA card number.
Suggested formats:
Fixes/Refs #RHIROS-XXX - Title
RHIROS-XXX Title
Feel free to remove this section from PR description once done.
Why do we need this change? :thought_balloon:
Please include the context of this change here.
Documentation update? :memo:
[ ] Yes
[ ] No
Security Checklist :lock:
Upon raising this PR please go through RedHatInsights/secure-coding-checklist
:guardsman: Checklist :dart:
[ ] Bugfix
[ ] New Feature
[ ] Refactor
[ ] Unittests Added
[ ] DRY code
[ ] Dependency Added
[ ] DB Migration Added
Additional :mega:
Feel free to add any other relevant details such as links, notes, screenshots, here.
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
:loudspeaker: Thoughts on this report? Let us know!.
Good catch! Thank you @upadhyeammit for pull-request.
|
2025-04-01T06:37:29.312465
| 2023-11-03T15:02:26
|
1976393890
|
{
"authors": [
"upadhyeammit"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2482",
"repo": "RedHatInsights/ros-backend",
"url": "https://github.com/RedHatInsights/ros-backend/pull/394"
}
|
gharchive/pull-request
|
RHTAP hermetic builds for pull and push
PR Title :boom:
Please title this PR with a summary of the change, along with the JIRA card number.
Suggested formats:
Fixes/Refs #RHIROS-XXX - Title
RHIROS-XXX Title
Feel free to remove this section from PR description once done.
Why do we need this change? :thought_balloon:
Please include the context of this change here.
Documentation update? :memo:
[ ] Yes
[ ] No
Security Checklist :lock:
Upon raising this PR please go through RedHatInsights/secure-coding-checklist
:guardsman: Checklist :dart:
[ ] Bugfix
[ ] New Feature
[ ] Refactor
[ ] Unittests Added
[ ] DRY code
[ ] Dependency Added
[ ] DB Migration Added
Additional :mega:
Feel free to add any other relevant details such as links, notes, screenshots, here.
We have RUN microdnf install --disableplugin=subscription-manager --nodocs -y python311 tar gzip gcc python3.11-devel as requirement hence hermetic mode won't work for us. Accordingly closing this for now!
|
2025-04-01T06:37:29.315622
| 2022-10-19T20:01:02
|
1415515550
|
{
"authors": [
"lindgrenj6",
"lpichler"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2483",
"repo": "RedHatInsights/sources-api-go",
"url": "https://github.com/RedHatInsights/sources-api-go/pull/615"
}
|
gharchive/pull-request
|
Reorganize various Authentication DAOs
Based on #607, so that one will need to go first.
This PR does a few things:
Renames authenticationDaoImpl -> authenticationDaoVaultImpl since that is what it is
Moves VaultClient itnerface (and creation) to dao/vault since it is a separate package
Creates dao/amazon for the secrets manager client
Initial empty secrets manager authentication dao impl
@lindgrenj6 this can be rebased
rebased!
/retest
/retest
Merging due to approval.
|
2025-04-01T06:37:29.326116
| 2022-12-19T21:57:14
|
1503664720
|
{
"authors": [
"Dannyb48",
"leelavg",
"waynesun09"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2484",
"repo": "RedHatQE/pylero",
"url": "https://github.com/RedHatQE/pylero/issues/115"
}
|
gharchive/issue
|
Support for pylero-cmd to be packaged as a docker container?
Was looking to leverage pylero-cmd within one of our test executor containers but rather than building the dependencies into my container was wondering if Pylero has any plans of having a containerized offering? that could be layered into other containers?
For base image we could consider the latest ubi9 images:
ubi9 base:
https://catalog.redhat.com/software/containers/ubi9/618326f8c0d15aff4912fe0b?container-tabs=gti
python 3.9 on ubi9:
https://catalog.redhat.com/software/containers/ubi9/python-39/61a61032bfd4a5234d59629e?container-tabs=gti
Ack, awaiting user info.
I see the requirement as "layered into other containers", so not sure whether they want pylero to be installed at a specific path and then later be copied during a multi-stage build or pylero standalone.
This is what I'm thinking, build the image with pylero installed in a venv and provide a way to copy the venv to any other container that already has python installed.
|
2025-04-01T06:37:29.328251
| 2024-05-15T20:21:29
|
2298771352
|
{
"authors": [
"calebevans",
"redhat-qe-bot1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2485",
"repo": "RedHatQE/qe-metrics",
"url": "https://github.com/RedHatQE/qe-metrics/pull/131"
}
|
gharchive/pull-request
|
[ #125 ] - Add retries to jira.connect and jira.search
Adds the use of the ignore_exceptions decorator to the connect and search functions in the jira lib
Closes #125
/verified
New container for quay.io/redhatqe/qe-metrics:latest published
|
2025-04-01T06:37:29.411863
| 2022-04-11T15:13:18
|
1200121871
|
{
"authors": [
"codecov-commenter",
"filipecosta90"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2486",
"repo": "RedisLabsModules/redisbench-admin",
"url": "https://github.com/RedisLabsModules/redisbench-admin/pull/340"
}
|
gharchive/pull-request
|
numpy.random fix BitGenerator attribute missing
Fixes #339 (based on https://github.com/aleju/imgaug/issues/537#issuecomment-1064647150 )
Codecov Report
Merging #340 (133d933) into master (b3b2d8c) will increase coverage by 0.06%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #340 +/- ##
==========================================
+ Coverage 54.26% 54.32% +0.06%
==========================================
Files 55 55
Lines 4412 4412
==========================================
+ Hits 2394 2397 +3
+ Misses 2018 2015 -3
Impacted Files
Coverage Δ
redisbench_admin/run/git.py
96.42% <0.00%> (ø)
redisbench_admin/utils/remote.py
68.26% <0.00%> (+0.75%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update b3b2d8c...133d933. Read the comment docs.
|
2025-04-01T06:37:29.418242
| 2024-10-06T17:19:16
|
2568776753
|
{
"authors": [
"StellaLovesTech",
"vktf1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2487",
"repo": "Redot-Engine/redot-engine",
"url": "https://github.com/Redot-Engine/redot-engine/issues/207"
}
|
gharchive/issue
| |
2025-04-01T06:37:29.496265
| 2022-10-26T14:21:59
|
1424125937
|
{
"authors": [
"Reithger",
"syoon2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2488",
"repo": "Reithger/Project-Diagram-Generator",
"url": "https://github.com/Reithger/Project-Diagram-Generator/issues/4"
}
|
gharchive/issue
|
Feature request: Attempt to Automatically Locate dot executable
It would be quite useful to automate (or at least attempt to automate) the process of finding the dot executable without asking user - possibly opening room for headless deployment in automated environments.
Might be able to do something like that, but I think I ran into issues of being able to freely explore the user's file system from an arbitrary position where the project is being run from. It should be 'do it once and forget', though I've been toying with embedding GraphViz in the project itself which I should get back to so we can avoid this entirely.
I think that (well, at least in *NIX environment) that attempting to find dot executable from the system's PATH environment variable would be a safe bet to make.
Since I intend to run this in an automated, noninteractive environment, e.g., GitHub Actions or Azure Pipelines, "do it once" part of "do it once and forget" is not really an option.
I've been toying with embedding GraphViz in the project itself which I should get back to so we can avoid this entirely.
Maybe you could consider replacing GraphViz.java with graphviz-java library? Seems like (from quickly glancing over README there... at least) it comes with automatic locating of dot executable.
|
2025-04-01T06:37:29.520825
| 2019-05-25T11:44:57
|
448456911
|
{
"authors": [
"iamrishishah",
"xiaochunjimmy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2489",
"repo": "Remix-Design/RemixIcon",
"url": "https://github.com/Remix-Design/RemixIcon/issues/42"
}
|
gharchive/issue
|
Shared and received folder
Icon Request
Icon name: shared folder
Tags of this icon: shared folder
Use case: Folder has been shared with team members
Screenshots of similar icons: A folder with right arrow
Icon name: received folder
Tags of this icon: received folder
Use case: Folder has been received by a team members
Screenshots of similar icons: A folder with left arrow
folder-shared-line
folder-shared-fill
folder-received-line
folder-received-fill
|
2025-04-01T06:37:29.525689
| 2019-02-28T02:27:29
|
415417487
|
{
"authors": [
"bwhite66"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2490",
"repo": "Remulus2006/Novelidea",
"url": "https://github.com/Remulus2006/Novelidea/pull/39"
}
|
gharchive/pull-request
|
Table of contents testing
trying to create better table of contents code
fixed table of contents
still need to stop audio on navigation bar pressed
|
2025-04-01T06:37:29.527822
| 2020-12-05T06:37:21
|
757565877
|
{
"authors": [
"ClaudioMerola",
"ryansmyth2001"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2491",
"repo": "RenatoGregio/AzureResourceInventory",
"url": "https://github.com/RenatoGregio/AzureResourceInventory/issues/3"
}
|
gharchive/issue
|
Testing the tool
Busy testing the tool and initially was failing, i needed the latest CLI version then started running..
Guess we do have largish environment - picked up on the error below.
{
"code": "ResponsePayloadTooLarge",
"message": "Response payload size is 19749321, and has exceeded the limit of 12582912. Please consider querying less data at a time and make paginated call if needed."
}
That is a very curious one, because we are already paging the results.
Can you download the new version 1.0.4 and try it again?
Also, can you run it using the “-Debug” parameter to tell us the last debug message you got before the error happen?
Sorry for delayed response - I will try again over the next few days and let you know..great tool!
|
2025-04-01T06:37:29.639205
| 2018-11-18T07:19:24
|
381935925
|
{
"authors": [
"Revan654",
"iAM261917"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2492",
"repo": "Revan654/staxrip",
"url": "https://github.com/Revan654/staxrip/issues/45"
}
|
gharchive/issue
|
A suggestion about releasing multiple-versions.
WinRAR has a feature called "Save identical files as references". It means that if two or more versions have a lot of identical files, you can just pack them in a same .rar package, and will not make the file size increase rapidly.
why am i suggest this? because i wanted to try both of <IP_ADDRESS> beta2 versions, sadly i had to download excess contents.
200 Megs is tiny, There are a few files that are different even if you don't notice them.
|
2025-04-01T06:37:29.657536
| 2024-01-21T09:17:32
|
2092496505
|
{
"authors": [
"Steen3S",
"rkalis"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2493",
"repo": "RevokeCash/revoke.cash",
"url": "https://github.com/RevokeCash/revoke.cash/issues/182"
}
|
gharchive/issue
|
Crypto.com Wallet Browser Extension causes connection issues
I've tested this scenario with no success.
Steps I took on Chrome (I've MetaMask installed)
install crypto.com extension
created a new wallet in crypto.com ext
The MetaMask connect button was still visible and operational. Also when switching the crypto.com ext to Ethereum main net nothing changed.
|
2025-04-01T06:37:29.670870
| 2021-10-12T07:22:01
|
1023458394
|
{
"authors": [
"rishiraj52"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2494",
"repo": "Reynadi531/a-patterns",
"url": "https://github.com/Reynadi531/a-patterns/pull/74"
}
|
gharchive/pull-request
|
Inverted Number Triangle added
Make numbers triangle
issue: #72
File changes made, can you review it now?
|
2025-04-01T06:37:29.674430
| 2022-09-08T07:32:19
|
1365689823
|
{
"authors": [
"KocsisGabor",
"Rezi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2495",
"repo": "Rezi/svelte-gestures",
"url": "https://github.com/Rezi/svelte-gestures/issues/14"
}
|
gharchive/issue
|
Possible limitation, or I just don’t understand something
This is a great library, thanks for sharing. However, in my use case I can't really use it, but maybe that's just because I don't understand something.
In my case, I have to watch for pan events on multiple elements. One option I have is to put the pan action on a container and manipulate the elements based on the target of the event, but this is not really a good solution for me because all the targets are complex, it would simply be too much work.
The other solution is to just place the action on each element individually. This works, but it doesn't do what I would do if I didn't use this library, because it only triggers the pointermove on the element itself.
What I would do is attach a pointerdown event to the element I want to manipulate, and when that event is triggered, I would attach a pointermove event to the window so I can move outside the element.
This library makes me attach the pointerdown and pointermove events to the same element, so I can't move outside of it.
Here is a repl which might help What I try to show in this repl is that the red box can not be moved outside of the grey box unless I attach the action only to the container in which case I need to deal with some unnecessary complexity.
I plan to fork your library to change the setPointerControls function so it accepts another node as a target for up, move and similar events but that is hard work for me since I need to understand everything that depends on this function. I thought I first should just ask you: do you think this change is easily possible or do you see problems that will appear after I made this change?
Sorry, but adding pan event on container is the right way to go. It is gesture library for a gesture recognition. You are probably looking rather for something like drag and drop. Or you simply need custom implementation.
As there is none else requesting this king of feature I am closing the issue in favour of keeping the library small with a single purpose
|
2025-04-01T06:37:29.702033
| 2018-09-11T18:11:15
|
359160607
|
{
"authors": [
"billyjanitsch",
"xtuc"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2496",
"repo": "Rich-Harris/packd",
"url": "https://github.com/Rich-Harris/packd/issues/60"
}
|
gharchive/issue
|
Server error for<EMAIL_ADDRESS>
https://bundle.run/<EMAIL_ADDRESS>yields the following error:
Cannot find module './common' from '/var/app/current/.tmp/3c70316a8727a02cb9b6de0d1d2c1ba749c33b98/package/node_modules/debug/dist'
I invalided the entry, could you please try again?
|
2025-04-01T06:37:29.737154
| 2022-10-21T19:17:41
|
1418760857
|
{
"authors": [
"Rickaym",
"kolibril13"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2497",
"repo": "Rickaym/plywood-gallery-for-vscode",
"url": "https://github.com/Rickaym/plywood-gallery-for-vscode/issues/2"
}
|
gharchive/issue
|
Including preview_image into gallery selection panel
It would be nice to use the social media preview image in gallery selection panel.
https://github.com/kolibril13/plywood-gallery-minimal-example/blob/e364cb73922018c9d3bbb09bf9fa9861befe3449/gallery_config.yaml#L5
Purposed design :
Current:
I like the proposal!
I think we should use this issue as part of the project from upgrading the TreeView into a WebView--which allows custom HTML rendering.
Currently, we're using the default implementation of VSCode's TreeDataProvider for our TreeViews: this does not allow custom HTML to be displayed. As part of a means to create a better user-friendly version for the TreeView, I've considered using a WebView rather than the data provider (pending implementation is at .../hub/hub.html containing the HTML code that should be rendered to display the TreeView)
Implementing a WebView means that we'll also have to factor in displaying the tree items (if the proposed implementation of the TreeView still wants to use it)
It would be awesome if you could contribute the HTML implementation for the TreeView--also do let me know if there's any confusion.
Ricky.
|
2025-04-01T06:37:29.749093
| 2020-08-26T09:01:02
|
686144432
|
{
"authors": [
"Pzixel",
"RicoSuter",
"mirinkinen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2498",
"repo": "RicoSuter/NJsonSchema",
"url": "https://github.com/RicoSuter/NJsonSchema/pull/1237"
}
|
gharchive/pull-request
|
Ignore case for discriminator
Currently Json.Net by default allows deserialize both "foo":1 and "Foo":1 into class Root { public int Foo {get;set;} }, but surprisingly JsonInheritanceConverter - doesn't.
Maybe it worth to add a constructor argument for specifying which StringComparison should it use but I think ignore case by default could be enough.
I dont think the OpenAPI spec allows case insensitive discriminators...
This is how Json.Net works by default. So different case actually work when real JSON get (de)serialized.
Being said, if it goes against the spec (which i'm not sure it does) then it could be at least configurable with spec-compatible setting by default.
I also have this problem, or a related problem at least.
I have all sorts of JSON in database that needs to be deserialized with custom discriminator fields and values. The same data needs to serialized for REST API clients, BUT the discriminator fields needs to be in different form than in the database. For example, I might have discriminator field of "Type" with value "text", which is first deserialized and then serialized again, but in this case, but the discriminator field name should be "type", all lower-case. So basically, I need different serialized forms for the same object, where the only difference is the casing of the discriminator field name.
This is not a huge problem of the API, but it looks weird when serialized objects have one field with capital letters, for example:
{ "id": 123, "age": "25", "Type": "person" }
I think the only way to do this, is to implement the whole JsonConverter abstract class, which I did. So I deserialized the discriminator like this:
var discriminator = jObject.GetValue(_discriminator, StringComparison.InvariantCultureIgnoreCase).Value<string>();
and serialized it for the REST API like this:
jObject[_discriminator.ToLower()] = JToken.FromObject(GetDiscriminatorValue(value.GetType()));
Not sure if this also could be tackled with Newtonsoft.Json.JsonSerializer settings or custom implementation of it.
Would you like to merge this if I add a configurable property (with default as it is now)?
I was wrong. If the class is not inherited from JsonInheritanceConverter, then the generated OpenAPI documentation will not show any discriminator mapping. It is mandatory to inherit from JsonInheritanceConverter and NOT from JsonConverter, if custom discriminator implementation is needed.
I'd say that Newtonsoft's behavior is just wrong, but as this is a Newtonsoft converter I'm ok with this change.
If there is an exact match it will win anyway so i dont see a problem with that...
The exact
// property name will be searched for first and if no matching property is found
// then the System.StringComparison will be used to match a property.
|
2025-04-01T06:37:29.829162
| 2021-05-19T18:21:58
|
895747699
|
{
"authors": [
"Cubicle13",
"RisingOrange"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2499",
"repo": "RisingOrange/anki-mindmap",
"url": "https://github.com/RisingOrange/anki-mindmap/issues/7"
}
|
gharchive/issue
|
Error (even with graphviz)
This add-on sounds amazing, but doesn't seem to be working. Every time I try to show or save a mindmap it gives me an error message (see debug info below). I've downloaded and installed version 2.47.1 of graphviz using the EXE installer for Windows 10 (64-bit).
How can I get this addon to work?
Error
An error occurred. Please start Anki while holding down the shift key, which will temporarily disable the add-ons you have installed.
If the issue only occurs when add-ons are enabled, please use the Tools > Add-ons menu item to disable some add-ons and restart Anki, repeating until you discover the add-on that is causing the problem.
When you've discovered the add-on that is causing the problem, please report the issue on the add-on support site.
Debug info:
Anki 2.1.22 (0ecc189a) Python 3.8.0 Qt 5.14.1 PyQt 5.14.1
Platform: Windows 10
Flags: frz=True ao=True sv=2
Add-ons, last update check: 2021-05-19 16:13:52
Caught exception:
Traceback (most recent call last):
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap_creator_dialog.py", line 46, in _on_show_button_click
self._save_mindmap_to_file(f.name)
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap_creator_dialog.py", line 89, in _save_mindmap_to_file
mindmap = TagMindmap(self.dialog.tag_prefix_lineedit.text())
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap.py", line 21, in init
self.notes_by_path = self._initialize_notes_by_path()
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap.py", line 53, in _initialize_notes_by_path
notes = get_notes(f'"{self.query_term}:{self.root_path}*"')
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\anki_util.py", line 41, in get_notes
for id in mw.col.find_notes(search_string)
AttributeError: '_Collection' object has no attribute 'find_notes'
I'm fine with updating, however the the avgEase addon nééds to work in order for my learning process. I saw you had an updated version of it. Does it function the same?
Yes, the only difference is that it starts to adjust the initial ease factors only once there are 50 mature cards in a deck.
Hm and I think the original add-on didn't adjust the ease factors at startup. Mine does.
Alright sounds good! This is going to be a problem. I think backwards compatibility (up to a certain point) is a must, because most other addons that aren't updated as frequently. What is the oldest version of Anki I can run this addon with?
I'm not sure what the oldest such version is, but it certainly works with 2.1.35.
2.1.24 works too.
Updated and got a new error when pressing show. Save feature works
Error
An error occurred. Please start Anki while holding down the shift key, which will temporarily disable the add-ons you have installed.
If the issue only occurs when add-ons are enabled, please use the Tools > Add-ons menu item to disable some add-ons and restart Anki, repeating until you discover the add-on that is causing the problem.
When you've discovered the add-on that is causing the problem, please report the issue on the add-on support site.
Debug info:
Anki 2.1.40 (cf446733) Python 3.8.6 Qt 5.14.2 PyQt 5.14.2
Platform: Windows 10
Flags: frz=True ao=True sv=2
Add-ons, last update check: 2021-05-19 16:13:52
Caught exception:
Traceback (most recent call last):
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap_creator_dialog.py", line 46, in _on_show_button_click
self._save_mindmap_to_file(f.name)
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap_creator_dialog.py", line 106, in _save_mindmap_to_file
raise e
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap_creator_dialog.py", line 91, in _save_mindmap_to_file
mindmap.save_as_img(
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap.py", line 41, in save_as_img
raise e
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890\mindmap.py", line 30, in save_as_img
create_mindmap_img(
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890_vendor\brain_dump\graphviz.py", line 112, in create_mindmap_img
pygraph.write_svg(output_file_path, prog=pydot_program, encoding='utf-8')
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890_vendor\pydot\pydot.py", line 1603, in new_method
self.write(
File "C:\Users\Toju\AppData\Roaming\Anki2\addons21\965278890_vendor\pydot\pydot.py", line 1686, in write
with io.open(path, mode='wb') as f:
PermissionError: [Errno 13] Permission denied: 'C:\Users\Toju\AppData\Local\Temp\tmpul32ua3u'
I uploaded a new version of the add-on to Ankiweb. It should fix the issue with the Show feature.
Does the Show button work for you now? Could you please test it? I am quite certain it does but I would like to know for sure.
Works perfectly now
Great, thanks!
|
2025-04-01T06:37:29.835590
| 2024-09-18T14:16:09
|
2533878015
|
{
"authors": [
"Ostrzyciel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2500",
"repo": "RiverBench/RiverBench",
"url": "https://github.com/RiverBench/RiverBench/issues/122"
}
|
gharchive/issue
|
Add a statistic about bytes per statement in NT/NQ encoding
In this paper we measured something called "bytes per triple": https://arxiv.org/abs/2406.16412
Basically, it's a measure of how many bytes we need to encode a triple or quad in the N-Triples/N-Quads format.
We can calculate this on the flat distribution serializer, by just examining the byte size of a stream element and dividing by the number of statements in that element. We'll need some changes to the stat counters to accomodate this.... maybe we'll need to merge in another stream with this extra information.
For this statistic, the "sum" and "unique" columns don't make sense, so skip them.
OK, it works! https://riverbench.github.io/v/dev/datasets/politiquices/#statistics-for-full-distributions
|
2025-04-01T06:37:29.872573
| 2016-07-09T10:01:19
|
164660402
|
{
"authors": [
"coveralls",
"gssiyankai",
"lemire",
"okrische"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2501",
"repo": "RoaringBitmap/RoaringBitmap",
"url": "https://github.com/RoaringBitmap/RoaringBitmap/pull/111"
}
|
gharchive/pull-request
|
Checkstyle fixes
Hey,
can someone please verify, if the current master branch can not be built because of checkstyle violations? Because here it does:
$ mvn clean package -DskipTests -Dgpg.skip=true
It does so on do/while blocks, when the "while" statement is put on the same line as the curly bracket.
If this is wanted, then i fixed those violations.
I also added checks for "run.sh" in the jmh folder, so that it fails, whenever one of the "mvn" commands fails. Right now it always ran through and jmh used old deployed versions of the jars (grrrr, took me some time to figure out, that it failed bcs of checkstyle)
Coverage remained the same at 84.027% when pulling f3617a045e90f4ad59c1722d3233e18f1e7834ee on okrische:feature/checkstyle-fixes into af363143c7566d78030e652d3371f34029cc6491 on RoaringBitmap:master.
Checkstyle also fails, when compiled/run with jdk7, hmm.
$ mvn clean checkstyle:checkstyle ... [ERROR] Failed to execute goal org.apache.maven.plugins:maven-checkstyle-plugin:2.17:checkstyle (default-cli) on project RoaringBitmap: Execution default-cli of goal org.apache.maven.plugins:maven-checkstyle-plugin:2.17:checkstyle failed: An API incompatibility was encountered while executing org.apache.maven.plugins:maven-checkstyle-plugin:2.17:checkstyle: java.lang.UnsupportedClassVersionError: com/puppycrawl/tools/checkstyle/api/AuditListener : Unsupported major.minor version 52.0 ...
Looks like, checkstyle already works with jdk8? Sigh.
Hi @okrische,
The build is failing because:
Roaring depends on the latest version of com.puppycrawl.tools:checkstyle
com.puppycrawl.tools:checkstyle:0.7 has been released 2 weeks ago
this version is not compatible with java 7
To fix the issue, just set explicitly the version to 6.19 in the root pom.
HTH
So, with 7.0 then, the RightCurly check will fail with the current master. Ok. Do you think, the RightCurly changes make sense nonetheless?
I am pretty used to this:
do { something(); } while(test());
But with 7.0 of the puppy checkstyle thingie, together with google style, it enforces:
do { something(); } while(test());
Coverage remained the same at 84.027% when pulling e369534939a52723aee3a86214ab45969c9d016b on okrische:feature/checkstyle-fixes into af363143c7566d78030e652d3371f34029cc6491 on RoaringBitmap:master.
@okrische
I am with you here... this style rule seems a tad odd. Nothing wrong with it per se, but forcing people to write this way is a bit much.
What are your thoughts @gssiyankai ?
Maybe the google style guide has not been updated yet on checkstyle side, yet.
https://google.github.io/styleguide/javaguide.html#s4.1-braces
Line break after the closing brace, only if that brace terminates a statement or terminates the body of a method, constructor, or named class. For example, there is no line break after the brace if it is followed by else or a comma.
And in a do/while the do {} is followed by "while". Hm.
@lemire,
No preference. I can go with both.
@okrische
I feel strongly that this is correct:
do {
System.out.println("Count is: " + count);
count++;
} while (count < 11);
while this is ugly...
do {
System.out.println("Count is: " + count);
count++;
}
while (count < 11);
If someone feels very differently, we can discuss it... but if not...
Maybe the google style guide has not been updated on checkstyle side, yet.
We do this work actually. Roaring has its own style which is a watered down version of the Google style as interpreted by checkstyle... (meaning, we allow more things) That is, we do not use the default "google style" from checkstyle, instead we provide our own. This gives us some flexibility.
Can you go to file style/roaring_google_checks.xml? Find <module name="RightCurly">... next look at the content:
https://github.com/RoaringBitmap/RoaringBitmap/blob/master/style/roaring_google_checks.xml#L60
I think that the problem is that we have LITERAL_DO in the list which forces you to make the changes from your commit...
https://github.com/RoaringBitmap/RoaringBitmap/pull/111/commits/f3617a045e90f4ad59c1722d3233e18f1e7834ee
I think that if you delete LITERAL_DO from the line 60 in style/roaring_google_checks.xml, you'll find that your "ugly" commit is unnecessary.
Can you check?
(Note: I do not mean to imply that you did ugly work. Quite the opposite. You brought forth an important issue.)
Coverage decreased (-0.1%) to 83.903% when pulling e378ae1c1b404a5a81748d5d3f10ed0a192ab477 on okrische:feature/checkstyle-fixes into af363143c7566d78030e652d3371f34029cc6491 on RoaringBitmap:master.
changed roaring_google_checks.xml,
set version of puppycrawl checkstyle tool to 6.19,
(also tested with puppycrawl checkstyle tool 7.0, so, when roaring goes jdk8, then it will work as well)
and added that fix to run.sh in jmh, so it will fail, when any build fails
Looks great to me. Since @gssiyankai went over some of this... let us merge.
|
2025-04-01T06:37:29.876358
| 2018-04-11T22:13:15
|
313506564
|
{
"authors": [
"coveralls",
"lemire",
"richardstartin"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2502",
"repo": "RoaringBitmap/RoaringBitmap",
"url": "https://github.com/RoaringBitmap/RoaringBitmap/pull/242"
}
|
gharchive/pull-request
|
Evaluate batch iteration
I have implemented a similar pattern to the go library for batch iteration. I will probably port this to the buffer package later this week. I would be interested in seeing benchmark results for this implementation.
Coverage increased (+0.06%) to 91.253% when pulling ff00f3e06d97589110ab4ba25e886c80fb8ed651 on richardstartin:iterators into 50515ad921d2c2c2232d7b8a5ffff7af36218696 on RoaringBitmap:master.
OK. I plan to port to the buffer version and have used the wrong email address on some commits which needs fixing (I will need to force push).
@richardstartin I think we want to do something meaningful like sum the values being iterated over. And compare the old-school iterators with the new iterators.
Is this what you have in mind?
@lemire The code I propose evaluating is on #243 - can we discuss there?
Of course.
I'm going to lock this conversation.
|
2025-04-01T06:37:29.888489
| 2022-01-14T20:25:42
|
1104094956
|
{
"authors": [
"blacelle",
"lemire",
"richardstartin"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2503",
"repo": "RoaringBitmap/RoaringBitmap",
"url": "https://github.com/RoaringBitmap/RoaringBitmap/pull/545"
}
|
gharchive/pull-request
|
Improving documentation.
Strictly speaking, our bitmaps violate the Java contract. It is possible for two bitmaps to be "equal" while having a different hash code.
The reason why that is is that...
We decided that equals was "set equality". So it is not sensitive to runOptimize(). It makes sense, I suppose. It can be computed quickly. We have fast algorithms.
For the hashCode computation, if we want to it to be insensitive to the runOptimize(), we effectively have to redo the run compression, it seems and I cannot find an efficient way to do it. Now, if we only had arrays and run containers, it would not be too hard. You could hash a run like this...
def fast(start, end):
K = end + 1 - start
return ((31**K-1)*(30*start+1)-30*K) // 900
and it would be equivalent to the array computation
def f(start, end):
h = 0
for i in range(start,end+1):
h = 31 * h + i
return h
But how do you deal with bitmap containers? It seems like you might have to iterate through every bit set... which is pretty terrible.
So if try to satisfy the contract, we will have slow hashes. Now hashes are typically used in a context where you want high speed.
So the solution I am proposing is that we selectively break the equal/hashCode contract by telling users that they either need to always do "run" compression, or never.
What is the impact? If you use bitmaps as keys in a hash table, you might try to use two bitmaps that differ in their storage but are otherwise equal as sets. So you'd get as a result that the 'key' is not found. But this is easily fixable by just calling "runCompress" before you use the bitmap if "runCompress" is used in the bitmaps that are hashed.
@richardstartin Can you have a look?
@Oppen Please have a look too!
In my opinion this should not be violating the spec. We should not implement hashCode or equals at all, and nobody should put a bitmap in a HashMap or HashSet and expect a good outcome. The equals method really tests set equality and not structural equality, give it another name like isSame or isEqual.
The other option is to implement the specification of equals and hashCode. If a user finds that’s slow, they are in the very small minority of people who care.
I’m guessing you don’t have time to actually fix this right now (nor do I) so the documentation is an adequate mitigation.
What is the impact? If you use bitmaps as keys in a hash table, you might try to use two bitmaps that differ in their storage but are otherwise equal as sets. So you'd get as a result that the 'key' is not found.
One thing quite worrying to me is you suggest OK to use RoaringBitmap as key of a hash structure. Any structure which hash may change (including mutable structures) is not a good candidate for a hash structure in my opinion (or at least, you need to know what you are doing). In the case of RoaringBitmap and discussed limitation, it may be too much to know. I feel we should recommend not using RoaringBitmaps as key in hash-based structure.
The other option is to implement the specification of equals and hashCode. If a user finds that’s slow, they are in the very small minority of people who care.
I feel current .equals implementation is OK, but current .hashCode implementation is dubious. Maybe we should drop .hashCode implementation, until implementing a slow one (as slow is better than broken (I feel current hashCode is broken, as very unsafe)).
I don't think we have a choice: there is still a hashCode. In Java, it cannot be helped.
My current view is that the current implementation is fast and simple. Improving the documentation probably helps.
@blacelle
We know that some people put bitmaps into sets which may effectively be implemented as a hash set:
https://github.com/RoaringBitmap/RoaringBitmap/issues/46
You are totally correct that this will not work if the instance is mutated. So I do not see how it can be safe.
I am actually not sure what purpose hashCode serves. My impression is that if it is used at all, it should be fast, but maybe I am wrong about it.
It seems that @Oppen wants to implement it in Go: https://github.com/RoaringBitmap/roaring/pull/347
I think we should discuss it.
It should be helpful to know and understand why people want to hash the bitmaps.
I don't think we have a choice: there is still a hashCode. In Java, it cannot be helped.
@lemire we could just not implement it actually, then it would default to Object semantics. I don't think this is a very important problem and the javadoc is sufficient.
I don't think we have a choice: there is still a hashCode. In Java, it cannot be helped.
We can either rely on default hashCode, which has a specific behaviour. We may also override, and throw UnsupportedOperationException, to prevent some usages. (May be a bad idea, still it is an option if we consider current .hashCode as broken.)
We may also limit .hashCode usage for Immutable implementations.
I am actually not sure what purpose hashCode serves.
?. You mean, out of hash-based structures?
We may also force .runOptimize on .hashCode calls.
We may also force .runOptimize on .hashCode calls.
I don't think it is quite satisfying.
we could just not implement it actually, then it would default to Object semantics.
It would still not meet the contract.
I'll merge. We can improve later.
|
2025-04-01T06:37:29.905461
| 2018-12-22T20:46:17
|
393702010
|
{
"authors": [
"RobThree",
"abatishchev"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2504",
"repo": "RobThree/IdGen",
"url": "https://github.com/RobThree/IdGen/pull/12"
}
|
gharchive/pull-request
|
Update README.md to use NuGet badge
Also fixing the casing of .NET
I see, you already have badges but in the bottom. Usually projects have them in the top. Would like to move them?
See f4c2425e43a5fc3f59ee7eeb18e7f61a9fd770e9
I prefer it below the title. Also, I prefer .Net over .NET 😉
Thanks anyway! 👍
You're welcome 🖖. Officially it's .NET, by the way ;)
|
2025-04-01T06:37:29.910290
| 2023-07-26T15:08:04
|
1822624861
|
{
"authors": [
"1220951120",
"RobTillaart"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2505",
"repo": "RobTillaart/DHT20",
"url": "https://github.com/RobTillaart/DHT20/issues/12"
}
|
gharchive/issue
|
DHT20::read() Need Time out
int DHT20::read()
{
// do not read to fast == more than once per second.
uint32_t temp ;
if (millis() - _lastRead < 1000)
{
return DHT20_ERROR_LASTREAD;
}
int status = requestData();
if (status < 0) return status;
// wait for measurement ready
temp = millis();
while (isMeasuring())
{
if(millis()-temp >=1000)
{
Serial.print("DHT20 Time Out\n");
return 0;
}
yield();
}
// read the measurement
status = readData();
if (status < 0) return status;
// convert it to meaningful data
return convert();
}
Thanks!
Thanks for the issue.
What is the problem exactly?
What platform(s) did it occur?
Note: Error code is already defined: DHT20_ERROR_READ_TIMEOUT in the DHT20.h file
Fixed in 0.2.3 release
|
2025-04-01T06:37:29.968737
| 2015-09-18T14:41:12
|
107215948
|
{
"authors": [
"RobinHerbots",
"goyney",
"socrim"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2506",
"repo": "RobinHerbots/Inputmask",
"url": "https://github.com/RobinHerbots/Inputmask/issues/1040"
}
|
gharchive/issue
|
Visualize regular expressions
Hi
We have evaluated a lot of plugins similar to yours but
jquery.inputmask ROCKS!!
Are there any plans to add visualization to your plugin?
Best
Mirco
@socrim ,
Thx ;-)
Do you mean visualization to the regex extension?
Hi Robin,
yes I mean visualizing the regex. E.g.: the mask would look like this, e.g. a
simple license code.
xxxx-xxxx-xxxx
A user would not have to type the dashes instead the cursor skips these and
forwards to the next position in the mask where user can enter something.
This would be a great extension to your plugin.
Best
Mirco
Robin Herbots<EMAIL_ADDRESS>hat am 18. September 2015 um 23:05
geschrieben:
@socrim https://github.com/socrim ,
Thx ;-)
Do you mean visualization to the regex extension?
—
Reply to this email directly or view it on GitHub
https://github.com/RobinHerbots/jquery.inputmask/issues/1040#issuecomment-141568694
.
@socrim ,
.... and why can't you use a mask instead of a regex?
in our software regex is used to define a mask.
the problem we have is visualizing this mask.
I thought that this would be an extension to your plugin.
But, sorry if I did not understand your question right.
Best
Mirco
Robin Herbots<EMAIL_ADDRESS>hat am 21. September 2015 um 11:52
geschrieben:
@socrim https://github.com/socrim ,
.... and why can't you use a mask instead of a regex?
—
Reply to this email directly or view it on GitHub
https://github.com/RobinHerbots/jquery.inputmask/issues/1040#issuecomment-141928548
.
@socrim ,
Don't get me wrong, I was just asking.
It is definitly a feature I want to add to the regex extension. I just have to figure out which way to go. Or translate the regex to a mask definition or analyse the regex and show a valid placeholder.
..
Hi Robin,
probably translating the regex to a mask definition would be step 1 and showing
a placeholder like
12345-abcde-a1b2c3
where the fixed chars are somehow styled differently maybe like this:
12345-abcde-a1b2c3
then you start typing and what you type then looks like this.
0815-abcde-a1b2c3
What do you think?
Best
Mirco
Robin Herbots<EMAIL_ADDRESS>hat am 21. September 2015 um 13:55
geschrieben:
@socrim https://github.com/socrim ,
Don't get me wrong, I was just asking.
It is definitly a feature I want to add to the regex extension. I just have
to figure out which way to go. Or translate the regex to a mask definition or
analyse the regex and show a valid placeholder.
..
—
Reply to this email directly or view it on GitHub
https://github.com/RobinHerbots/jquery.inputmask/issues/1040#issuecomment-141950994
.
@socrim ,
Can you provide the regex you want to use and I guess I dont see a styling difference?
I have attached the differences in style as attachment.
They propably did not make it through my email client.
Robin Herbots<EMAIL_ADDRESS>hat am 21. September 2015 um 14:36
geschrieben:
@socrim https://github.com/socrim ,
Can you provide the regex you want to use and I guess I dont see a styling
difference?
—
Reply to this email directly or view it on GitHub
https://github.com/RobinHerbots/jquery.inputmask/issues/1040#issuecomment-141959946
.
Would like to chime in on this feature request. We would greatly appreciate the addition of this functionality. All our validation rules are configured in RegEx and the visualization of the pattern (and the need to not type in things like hyphens, etc.) would be very helpful to our user's experience.
@socrim , @goyney ,
Finally there is some progress on this.
You can try with
Inputmask({ regex: "your regex" }).mask(selector);
There is no need for the extension anymore.
|
2025-04-01T06:37:29.987310
| 2018-02-04T16:39:51
|
294210682
|
{
"authors": [
"Robinson664",
"threadpool",
"vitya-maleyev"
],
"license": "MS-PL",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2507",
"repo": "Robinson664/bemu",
"url": "https://github.com/Robinson664/bemu/pull/1"
}
|
gharchive/pull-request
|
Linux/CMake support
I've managed the project to be built and run on Linux: tested on Centos 7 with CMake 2.8 and Arch Linux with CMake 3.10.1.
Windows build with CMake also works and seems faster as explores parallel build feature.
@Robinson664, could you please review?
Victor,
Sure! I'm not familiar with git hub and my git knowledge is weak so I'm
not sure what to do. I clicked "Merge Pull Request". I think that did it.
-Jordan
On Tue, Feb 6, 2018 at 3:49 PM, Victor Maleyev<EMAIL_ADDRESS>wrote:
@Robinson664 https://github.com/robinson664, could you please review?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/Robinson664/bemu/pull/1#issuecomment-363560362, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AeK_15ho52jQAB7EElezLaKC54q0wf1Gks5tSLrmgaJpZM4R4oT4
.
Many thanks!
Do you want to update the readme to include information about this?
On Feb 8, 2018, at 3:35 AM, Victor Maleyev<EMAIL_ADDRESS>wrote:
Many thanks!
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
Hi there, I have a few questions about the evolution the API. Are you still following this thread
I get emails when I'm mentioned in the thread. I may be able to help you out if you have questions. I don't actively work on this project anymore because I don't have the time or the Bloomberg account I used to have.
Hey there,
Thanks for getting back to me.
Was the initial code based on the de complied Bloomberg source ?
I’d be happy to clone and re write the java version and enhance it. Would you have an issue with that. I can publish it once it’s complete.
Cheers,
Oz
On 11 Nov 2018, at 24:41 , Robinson664<EMAIL_ADDRESS>wrote:
I get emails when I'm mentioned in the thread. I may be able to help you out if you have questions. I don't actively work on this project anymore because I don't have the time or the Bloomberg account I used to have.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub https://github.com/Robinson664/bemu/pull/1#issuecomment-437630667, or mute the thread https://github.com/notifications/unsubscribe-auth/ADK0DfmF9nQVn99vHe_t4_XYGiMARNNaks5ut2QVgaJpZM4R4oT4.
Can you drop me your email address please, I'll be easier to communicate. Thanks
I didn't decompile the Bloomberg source. For the. Net version anyways, I don't think the DLL is a .Net DLL. I'm not familiar with the tools available for decompiling non-.Net DLLs.
What you want to do with the Java version sounds cool. What did you have in mind?
My email is<EMAIL_ADDRESS>
I'm not sure if you saw my email. I’d be interested to know what your starting point for the project was. Was it the original Bloomberg source code. If so do you still have it and which language was it written in ?
|
2025-04-01T06:37:29.990039
| 2024-05-16T01:35:00
|
2299126434
|
{
"authors": [
"aetter",
"dogwarrior24"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2508",
"repo": "Roblox/creator-docs",
"url": "https://github.com/Roblox/creator-docs/pull/553"
}
|
gharchive/pull-request
|
Updates table.sort description to be clearer
Changes
Updates table.sort description to be clearer about how to create a valid comparison function. This question comes up often enough I think it is worth updating the documentation.
Checks
By submitting your pull request for review, you agree to the following:
[x] This contribution was created in whole or in part by me, and I have the right to submit it under the terms of this repository's open source licenses.
[x] I understand and agree that this contribution and a record of it are public, maintained indefinitely, and may be redistributed under the terms of this repository's open source licenses.
[x] To the best of my knowledge, all proposed changes are accurate.
Looks good to me, thanks a bunch @dogwarrior24!
|
2025-04-01T06:37:29.995774
| 2023-10-01T14:32:25
|
1920797877
|
{
"authors": [
"DimitarBogdanov",
"vegorov-rbx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2509",
"repo": "Roblox/luau",
"url": "https://github.com/Roblox/luau/issues/1057"
}
|
gharchive/issue
|
Feature request - Type assertion operator
Intro
Proposal: Create an operator to force-cast a value, skipping type-relation checks.
Proposed syntax - one of the following (read 2nd note at the bottom):
value ::! type
value as type
Purpose
Sometimes while programming, you come across situations in which you're certain a certain value is of a certain type. In Luau, it is currently too cumbersome to assert this type of functionality, requiring an assert call. While this is safer, often times it's unnecessary.
Example problem
--!strict
local player = game:GetService("Players")
local character = assert(player.Character)
local head = character:FindFirstChild("Head")
head.Color = Color3.fromRGB(120, 120, 120)
-- This is a problem: we can't assert that Head is a BasePart.
-- This is theoretically the case, but in practice, it's much cleaner to be able to assert obvious facts.
-- This does not change strong typing, and for codebases where this'd be a problem, the syntax can simply be discouraged.
Example use case (syntax 1):
local head = character:FindFirstChild("Head") ::! BasePart?
-- Even though FindFirstChild returns Instance, head is of type BasePart?, because it was forcecasted
Example use case (syntax 2):
local head = character:FindFirstChild("Head") as BasePart?
Current workaround
This is how you'd get around the error today, which is quite ugly:
local head = (character:FindFirstChild("Head") :: any) :: BasePart?
Notes
Cost to develop - I had a little look throughout the Luau codebase, and it seems that this shouldn't be too hard to implement. The logic of the :: operator can be duplicated, except for TypeChecker2, where visiting the AST node wouldn't actually perform type checks.
The ::! syntax is more expressive (and potentially less confusing for people coming over from TS).
This actually didn't use to be a problem - casting from Instance to any descendant of Instance was acceptable. A recent change in typechecking seems to have changed this behaviour, which prompts for this feature request.
This was discussed recently on the Roblox developer forums and I agree, this is a bug in the current operator that we intend to fix https://devforum.roblox.com/t/the-return-type-of-findfirstchild-is-incorrect/2599989/4
So this doesn't require a new language feature.
Sweet, thanks.
|
2025-04-01T06:37:29.999231
| 2023-11-17T10:50:25
|
1998821156
|
{
"authors": [
"Babanimus",
"EngyneMac",
"heliawins",
"othomson-roblox"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2510",
"repo": "Roblox/roblox-blender-plugin",
"url": "https://github.com/Roblox/roblox-blender-plugin/issues/33"
}
|
gharchive/issue
|
cant upload
if i upload my module something about error of python trace back (most resent call last):
file c:users/user/appdata/roaming/blender foundation/blender/4.0/scripts/addons/roblox-blender-plugin/lib/get_selected_objects.py",line 66, in get_selected_objects for selected_objects in context.selected_ids:
attributerror: 'context' object has no attribute 'selected_ids'. did you mean 'selected_bones'?
that's the error I'm getting pls help
I have the same issue
Solution:
https://github.com/Roblox/roblox-blender-plugin/issues/32#issuecomment-1816527799
Thanks for flagging! This plugin did not work with Blender 4.0 but this issue has been resolved in the latest release:
https://github.com/Roblox/roblox-blender-plugin/releases/tag/v1.0.2
|
2025-04-01T06:37:30.013032
| 2020-04-30T14:16:53
|
610051517
|
{
"authors": [
"RobotJesse",
"Yuhanun",
"rolfvdhulst",
"timovdk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2511",
"repo": "RoboTeamTwente/roboteam_ai",
"url": "https://github.com/RoboTeamTwente/roboteam_ai/pull/1064"
}
|
gharchive/pull-request
|
Stp/interface
Self-explanatory, updated the UI according to our new changes :)
I got a segfault with this stack trace after running your branch:
Branch is presumably failing because you committed an unresolved merge conflict xD
Resolved.
The plays tab does not segfault anymore :+1:, however, it is also currently not displaying any useful information :stuck_out_tongue_closed_eyes:. It should display the play name, score, and if possible also their invariants?
Why is help needed for this?
It segfaults in the invariants, probably due to some sort of data race for world/robots/field. I think most segfaults stem from robots that don't exist anymore
LGTM. Only wondering why in a previous PR getName() was determined to be char *. Why not use std::string or QString?
I don't want O(n^2) allocations every tick.
|
2025-04-01T06:37:30.015937
| 2023-12-21T01:09:21
|
2051510227
|
{
"authors": [
"EricCousineau-TRI",
"adityapande-1995",
"ahcorde"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2512",
"repo": "RobotLocomotion/drake-ros",
"url": "https://github.com/RobotLocomotion/drake-ros/pull/322"
}
|
gharchive/pull-request
|
[drake] Upgrade to v1.24.0
Resolves #321
This change is
I included this change in this other PR https://github.com/RobotLocomotion/drake-ros/pull/325 and now CI is green
already included and merged in this PR https://github.com/RobotLocomotion/drake-ros/pull/325
Closing this and the connected issue.
thanks!
|
2025-04-01T06:37:30.142213
| 2024-08-08T20:53:30
|
2456606269
|
{
"authors": [
"HaSistrunk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2513",
"repo": "RockefellerArchiveCenter/dimes",
"url": "https://github.com/RockefellerArchiveCenter/dimes/issues/731"
}
|
gharchive/issue
|
Fix aria-describedby in form help text and error messages
Describe the bug
aria-describedby is an attribute that should be used on the input element that points to an id on the help text element, not the other way around. In DIMES, this has been implemented backwards so that the help text has the aria-describedby attribute instead of the input.
To reproduce
Steps to reproduce the behavior:
Go to the request in reading room or request copies modal
Inspect the page
See that the help text below text inputs and text areas has an aria-describedby attribute, but the input itself does not.
Try submitting a form without filling it out
See that the error message below the inputs have an aria-describedby attribute, but the inputs do not.
Expected behavior
See the RAC style library text area with help text and with error for the correct implementation.
Impact on your work
When screen reader users navigate to a form input, they won't automatically hear the associated help text or error messages.
Update: this appears to currently only be an issue for the "Request in Reading Room" scheduled date input help text.
|
2025-04-01T06:37:30.147729
| 2024-04-04T12:46:34
|
2225404232
|
{
"authors": [
"binse03",
"jeanfbrito"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2514",
"repo": "RocketChat/Rocket.Chat.Electron",
"url": "https://github.com/RocketChat/Rocket.Chat.Electron/issues/2874"
}
|
gharchive/issue
|
Rocket.Chat crashes after opening a chat
Search before asking
[X] I had searched in the issues and found no similar issues.
Operating System
[ ] macOS
[X] Windows
[ ] Linux
Operating System Version
Windows 10 Enterprise
It happens on the web browser too?
No, it just happens on the Desktop app
Rocket.Chat Desktop App Version
3.9.14
Rocket.Chat Server Version
6.6.2
Describe the bug
A work colleague of mine opens a chat with a specific employee. The program then freezes and you can no longer do anything. The toolbar at the top still works, but there are fewer clicks than before. (For example, "Help > Reload" is then grayed out. The problem can only be solved with a new installation or with "Help > Reset app data". But the chat still cannot be opened. But then other Chats are working again. No files were shared in the chat, only plain text and basic images.
How to Reproduce
I don't know how he did it or how to reproduce it. The chat with the person still works for me on my PC.
Describe your Expected behavior
The Chat opens without a crash.
Anything else
When u need something (logs) please say me where i find it and i can post it for you.
Are you willing to submit a code contribution?
[ ] Yes, I am willing to submit a Pull Request!
Can you test our v4.0?
If it keeps crashing you should send us more information of whats going on?
Can you test our v4.0? If it keeps crashing you should send us more information of whats going on?
We will try.
|
2025-04-01T06:37:30.227115
| 2021-01-27T17:24:44
|
795296605
|
{
"authors": [
"AnandBaburajan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2515",
"repo": "RocketMeet/RocketMeet-client",
"url": "https://github.com/RocketMeet/RocketMeet-client/issues/38"
}
|
gharchive/issue
|
Send final meet time to votes via mailer
This sounds like a good feature. Currently only invites to vote can be sent.
Closing since #60 has been opened and assigned
Closing since #60 has been opened and assigned
|
2025-04-01T06:37:30.228852
| 2020-06-03T19:48:17
|
630278704
|
{
"authors": [
"AlbinoMeneguettiNeto",
"diego3g"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2516",
"repo": "Rocketseat/unform",
"url": "https://github.com/Rocketseat/unform/issues/246"
}
|
gharchive/issue
|
@unform/core exports Form
I'm using unform in a web app and I was using auto import from vscode and it auto imported from @unform/core.
The form from unform/core gave me the same errors as form from unform/web but it doesn't render.
Why is there a export Form from @unform/core? It gave me a little of trouble to figure out what was happening.
Fixed. Just update unform version using yarn add @unform/core @unform/web
|
2025-04-01T06:37:30.250055
| 2019-07-17T06:12:24
|
469005183
|
{
"authors": [
"RodainaMohamed",
"atologistkaushik",
"elega9t"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2517",
"repo": "RodainaMohamed/ionic-rating",
"url": "https://github.com/RodainaMohamed/ionic-rating/issues/4"
}
|
gharchive/issue
|
Change icon Color
rating {
ion-icon {
color: red;
&.filled {
color: #ffb400;
}
}
}
add above to my .css but it is not going to work, can anyone please tell me how to change star icon color for normal icon and filled icon?
rating { ion-icon { color: gray; &.filled { color: #fdbc2c !important; } } }
works for me
Please check the new version of the library that supports Ionic 4 and made it much easier to style the color of the stars using CSS Variables.
|
2025-04-01T06:37:30.255287
| 2017-03-10T06:28:29
|
213256253
|
{
"authors": [
"Roemer",
"jmaxxz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2518",
"repo": "Roemer/FlaUI",
"url": "https://github.com/Roemer/FlaUI/pull/44"
}
|
gharchive/pull-request
|
Refactor FlaUI.Core.Logging
Expose ability for consumers of FlaUI to customize
how FlaUI will log. Consumers can set both application
specific loggers, and the Default logger which will be
used by FlaUI.
Allow consumers to enable/disable each of the avaible
log levels.
Add multiple logger options
Addresses #43
I will add more unit test.
So, will you do a few adjustments so we can merge this PR? Or should we merge it and do all changes afterwards?
@Roemer, you can merge it. I will add more unit tests in a subsequent PR.
|
2025-04-01T06:37:30.262252
| 2016-12-23T07:59:11
|
197327352
|
{
"authors": [
"RogueException",
"Sentinent",
"advorange",
"khionu"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2519",
"repo": "RogueException/Discord.Net",
"url": "https://github.com/RogueException/Discord.Net/issues/437"
}
|
gharchive/issue
|
Unable to change position of a role.
Using the newest dev build. Bot has the top role in the server and has administrator. inputRole is somewhere near the middle of the role list.
IRole newRole = await Context.Guild.CreateRoleAsync(inputRole.Name, new GuildPermissions(0), inputRole.Color);
await newRole.ModifyAsync(x => x.Position = inputRole.Position);
I was able to change the name and permissions in a similar way, but position doesn't work.
I made a manual HTTP request with postman and the api seems to just ignore the position parameter. The way the client modifies positions is by using the modify guild roles positions endpoint.
Position is defined by Snowflake and Position, fyi
Is this still an issue? I'm unable to repro
|
2025-04-01T06:37:30.266264
| 2017-09-16T16:28:21
|
258242238
|
{
"authors": [
"BinkanSalaryman",
"Still34",
"khionu"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2520",
"repo": "RogueException/Discord.Net",
"url": "https://github.com/RogueException/Discord.Net/issues/816"
}
|
gharchive/issue
|
Inconsistency between Emoji and Emote
At the time of writing, to create an Emoji object, you'd have to pass the unicode into its public constructor; to create an Emote object, you'd have to do it through either the Parse or the TryParse method, and the object's constructor is internal only.
This feels extremely inconsistent and can sometimes be confusing.
Does Discord even make a distinction between Emoji and Emotes in the API?
Emote: To give expression to emotion
Emoji: Japanese for picture-character(s). (Yes, 絵文字 has no aspect of time like any jp. noun; "emojis" hurts to read. 🌵)
Mostly they can be used interchangeably in this case. But 🆎 is definitly not an emote. ;)
@BinkanSalaryman did you read the issue? The definition has nothing to do with the issue. The point of the issue was to point out the inconsistent implementation.
@Still34 I did, why so pricky? I hope I lifted your confusion that drove you to submit the naming as an issue!
Also, the terms used in this lib all follow the terms defined by Discord. Emote is specifically for a custom guild emote, while Emoji is for standard Unicode ones.
Discord only defines emoji at https://discordapp.com/developers/docs/resources/emoji...
Apologies if I sounded hostile; however, I still fail to understand why you brought up the definition when the issue itself was questioning the implementation between the current Emote and Emoji.
|
2025-04-01T06:37:30.269169
| 2022-09-26T16:57:43
|
1386421563
|
{
"authors": [
"RohanG326",
"nicm2"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2521",
"repo": "RohanG326/rohangfastpages",
"url": "https://github.com/RohanG326/rohangfastpages/issues/7"
}
|
gharchive/issue
|
Rohan Gaikwad Review Ticket #6
My Fastpages
Score: 2.7+/3 - based on crossover grade
Score: /3 - based on review
[x] Fibo Java
[x] JS data
[x] Project
Score 2.7/3 for good completion of individual work and team work
|
2025-04-01T06:37:30.270411
| 2022-10-26T11:07:00
|
1423842156
|
{
"authors": [
"Miloslav-RS"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2522",
"repo": "Rohde-Schwarz/RcIcPluginBugTracking",
"url": "https://github.com/Rohde-Schwarz/RcIcPluginBugTracking/issues/1"
}
|
gharchive/issue
|
SMA100A, SMA100B - FileBrowser does not work
The issue is related to the stricter file system access policy, it will be resolved in 1.2.0
Fixed in 1.3.0
|
2025-04-01T06:37:30.283254
| 2022-01-02T23:40:11
|
1092126907
|
{
"authors": [
"cjstoddard",
"jaemzfleming",
"nmbradley",
"roll20deploy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2523",
"repo": "Roll20/roll20-character-sheets",
"url": "https://github.com/Roll20/roll20-character-sheets/pull/9974"
}
|
gharchive/pull-request
|
Improvements to the [The Fantasy Trip] character sheet
This is a pretty big set of improvements to the sheet. These include:
Nicer looking Roll Templates for attacking and spells, including optional flavor text.
An Initiative button to add you to the Turn Order.
Tracking of unspent and spent experience points and IQ points to add in talent and attribute selection.
An array of attack buttons at various to-hit amounts, to allow single-click attacks at various to-hit amounts. You select your spell or weapon, then click the button you want. There are tooltips for common situations such as rear and flank attacks.
Tooltips on most captions and values.
A repeating section just for equipment, including current count, weight per item.
Description buttons to describe items/talents in the chat text.
Gear buttons to pop open repeating sections to edit detailed descriptions.
Weight tracking. every item has a weight, and a checkbox for whether it is currently carried or not. This is used to compute total weight and the Movement penalties, which are automatically applied.
Player data is preserved. I kept all the old names of attributes.
Layout is also nicer looking, using a grid layout and some formatting.
Roll20 Requests
Comments are very helpful for reviewing the code changes. Please answer the relevant questions below in your comment.
[ ] Does the pull request title have the sheet name(s)? Include each sheet name.
[ ] Is this a bug fix?
[ ] Does this add functional enhancements (new features or extending existing features) ?
[ ] Does this add or change functional aesthetics (such as layout or color scheme) ?
[ ] If changing or removing attributes, what steps have you taken, if any, to preserve player data ?
[ ] If this is a new sheet, did you follow Building Character Sheets standards ?
If you do not know English. Please leave a comment in your native language.
Character Sheet Info Roll20 Internal Use only.
Hello @nmbradley, thanks for the merge. Hey I just checked in-game and I'm not seeing any of my changes when I select "The Fantasy Trip" as my character sheet. How do I see the changes (I mean, without cutting and pasting into custom) Also, I wasn't expecting the pull to happen so quickly, and I actually added some more features today. Should I just make a new pull request?
Hi @jaemzfleming
I'll investigate why these updates are not showing up on the service and get back to you.
Yes, please create a new pull request for any further changes.
One further point: if you're not ready for a PR to be merged yet, you should be able to indicate that it's a draft.
Let me know if you have any problems doing this!
Thanks for looking into that. And I'll do the draft thing, that sounds smart.
Oh, @nmbradley it looks like the sheet.json had a trailing comma after the "instructions" field. This was failing validation. Maybe that's why it's not showing up? I've fixed it in my new pull request, which I've marked as draft. Maybe I'll un-mark it as draft then since it has that fix and do a new pull request next week for any other straggling features.
Hi @cjstoddard and @clevett, not sure if this is the right way to alert you, but I've made some improvements to The Fantasy Trip character sheet, they've been merged and you can grab them to see if you'd like. I'm new to GitHub, not sure exactly where/how discussions should take place.
I am glad to see others working on improvements to this character sheet. Honestly, what I did was pretty crude and my intention was just to get something onto Roll20 hoping others people with actual skills would improve on it.
Mission accomplished then! It provided a great launching point.
|
2025-04-01T06:37:30.287200
| 2024-05-12T09:35:49
|
2291253858
|
{
"authors": [
"RomainLanz",
"jarle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2524",
"repo": "RomainLanz/adonis-bull-queue",
"url": "https://github.com/RomainLanz/adonis-bull-queue/issues/37"
}
|
gharchive/issue
|
Pass payload to the rescue handler
In the @next branch, passing the payload to the rescue handler would be helpful for updating database entities with a failure state.
Released as 3.0.0-0 👍🏻
There are no breaking change, but 2.0.0 will be used for an AdonisJS 5 update.
|
2025-04-01T06:37:30.294469
| 2019-05-07T03:57:31
|
441014881
|
{
"authors": [
"JonTheNiceGuy",
"RomuloOliveira"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2525",
"repo": "RomuloOliveira/commit-messages-guide",
"url": "https://github.com/RomuloOliveira/commit-messages-guide/issues/39"
}
|
gharchive/issue
|
Include allcontributors.org to acknowledge the work of everyone
Reviewing translations is a valuable work. I'd like to be able to acknowledge and enhance this type of contribution.
https://allcontributors.org/
https://github.com/all-contributors/all-contributors
Hi! AllContributors is now available as a bot :)
https://allcontributors.org/docs/en/bot/installation
That said, it appears to only add entries to a single file, so you might struggle with adding acknowledgements to multiple language files.
|
2025-04-01T06:37:30.304787
| 2021-09-07T16:23:34
|
990151937
|
{
"authors": [
"MBeijer",
"RonenNess"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2526",
"repo": "RonenNess/sini",
"url": "https://github.com/RonenNess/sini/pull/1"
}
|
gharchive/pull-request
|
Support Dictionary<string,string> sections
Hi Ronen,
I have a legacy project where one section is used as a Dictionary<string,string> with filters where key can be anything just as value can be anything.
This solution solves this for me, but I am unsure if it would break anything else. I mean, the code will only execute if the instance passed is a Dictionary<string,string>.
I am open for suggestions to improve this, so it can be upstreamed and added to NuGet, and I don't have to keep maintaining my own fork.
Thanks for your work, saved me the hassle of doing the same on my own!
Best regards,
Marlon Beijer
Hello @MBeijer,
Thank you for the PR, I appreciate it!
It's a weird coincidence just today I worked with SINI and wanted to add sections getter, so your timing is good.
I decided to take a slightly different approach for your requirement, please see this new method:
https://github.com/RonenNess/sini/commit/7a2b0655657583c03da49e0cbaff6a7acc326059#diff-c661505f1c432d12a9a2a77ad3ce364d883cdaf110d6cf3c860845d49eded9b6R163
Which can be combined with this getter, if you want all keys from all sections:
https://github.com/RonenNess/sini/commit/7a2b0655657583c03da49e0cbaff6a7acc326059#diff-c661505f1c432d12a9a2a77ad3ce364d883cdaf110d6cf3c860845d49eded9b6R163
I think this way its a cleaner API, but ofc its a matter of personal preference :)
The new version was uploaded to Nuget and will probably be online soon.
@RonenNess In my case it was needed from ToObject as I am casting to an object that I use for both legacy .ini and new .json style config files.
I can implement this in ToObject using your new methods and send you a new PR, using this approach instead. :)
I initially tried to use CustomParsers and add a parser to the default configuration, but that fails because var asStr = ini.GetStr(section, key, null); returns null (of course) because it's a section and not a property-line.
Or am I missing something that makes this work already in ToObject? From your changes I didn't see anything that would do this
Maybe I didn't understand your use case.
Are you parsing an object containing a Dictionary<string, string> or are you trying to convert a whole file / section to Dictionary<string, string>?
@RonenNess
I have a Settings class that has a public Dictionary<string,string> Filter {get; set;}
So I am loading the settings.ini with IniFile.ToObject<Settings>("settings.ini");
public class Settings
{
public static async Task<Settings> CreateConfig(string config)
{
Settings settings = null;
if (config.EndsWith(".ini"))
{
settings = IniFile.ToObject<Settings>(config, IniFile.ParseObjectFlags.AllowMissingFields | IniFile.ParseObjectFlags.AllowAdditionalKeys | IniFile.ParseObjectFlags.SnakecaseKeysAndSections);
await Save(settings, config);
}
else if (config.EndsWith(".json"))
settings = await JsonFile.ToObject<Settings>(config);
if (settings != null)
settings.ConfigFile = config;
return settings;
}
private static async Task Save(Settings settings, string configFile)
{
string path = Path.GetDirectoryName(configFile);
await File.WriteAllTextAsync($"{path}{Path.DirectorySeparatorChar}configuration.json", settings?.ToJson(true));
}
private string ConfigFile { get; set; }
public GeneralModel General { get; set; }
public PogDesignModel Pogdesign { get; set; }
public RssModel Rss { get; set; }
public Dictionary<string,string> Filter { get; set; }
}
``
settings.ini:
[pogdesign]
username =
password =
[general]
loglevel = Info
storage = /home/user/Downloads
[filter]
key1 = value1
key2 = value2
key3 = value3
[rss]
username =
password =
passkey =
cookie_uid =
cookie_pass =
cookie_validation =
The filter section is entirely dynamic, which is why I need Dictionary<string,string>
Ah I understand now, so the whole section 'filters' is translated to a member that is a Dictionary<string,string>. Make sense to have that working.
It can now be better implemented with the new GetKeys() method. In the original PR what bothered me the most was the use of 'GetAllUnreadKeys', which can now be avoided.
I think I'll do it sometime tomorrow, that's a good requirement.
Thanks! :)
@RonenNess I tagged you in a comment on a commit i made, where I did a simple solution (but probably not a complete one) https://github.com/RonenNess/sini/commit/257a3b7fde322e7192ae6e3c9945073177fe14ba
Your solution is good, just added test + fixed AsDictionary() to update accessed keys, so it won't throw exception on unused keys.
https://github.com/RonenNess/sini/commit/9ccf9758b2ed07f476bfa1463b26a954385ed455
NuGet package should update shortly :)
@RonenNess Nice! Thank you! :)
|
2025-04-01T06:37:30.434360
| 2019-01-29T20:19:32
|
413261008
|
{
"authors": [
"doongmiin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2527",
"repo": "RoyaleAPI/family-config",
"url": "https://github.com/RoyaleAPI/family-config/issues/2"
}
|
gharchive/issue
|
[Family] Scunion eSports Family
Consult https://github.com/RoyaleAPI/cr-api-ux/wiki/Family-Config and fill in the following information.
name: Scunion eSports Family
key: scunion
color: black
emblem: Clover_02
info:
logo: https://drive.google.com/file/d/1JGSzp-eeErvvpIVuxv4GZgOwX9iqJSRX/view?usp=sharing
description: >
*scunion [skuhn-yuhn]: A term used in the late 1960s during the Vietnam War to signal the inflicting of distress, injury, or destruction // as in, "bring the scunion!" **Scunion eSports** is an eSports team competing in the International region of Clash Royale League. Ever since it began on September 11, 2017, our clans and members have been growing steadily. Our clan comprises members from many different places, different backgrounds, and different ages. We all share our love for the game, donate when possible, help in the Clan Wars when available, and help each other out. _Join us, and let's bring the scunion!_
social:
- facebook: https://www.facebook.com/scunion.esports.14
clans:
- name: Scunion eSports
tag: 82CVUPU2
- name: Scunion Elites
tag: 9Y2PUUYV9
I made a mistake, wrote the wrong facebook link, updated:
name: Scunion eSports Family
key: scunion
color: black
emblem: Clover_02
info:
logo: /static/img/brands/logo/scunion-esports.png
description: >
*scunion [skuhn-yuhn]: A term used in the late 1960s during the Vietnam War to signal the inflicting of distress, injury, or destruction // as in, "bring the scunion!" **Scunion eSports** is an eSports team competing in the International region of Clash Royale League. Ever since it began on September 11, 2017, our clans and members have been growing steadily. Our clan comprises members from many different places, different backgrounds, and different ages. We all share our love for the game, donate when possible, help in the Clan Wars when available, and help each other out. _Join us, and let's bring the scunion!_
social:
- facebook: https://www.facebook.com/groups/100478373784989/
clans:
- name: Scunion eSports
tag: 82CVUPU2
- name: Scunion Elites
tag: Y2PUUYV9
No worries! Thanks for your time and all you do! 👍 @smlbiobot
|
2025-04-01T06:37:30.443873
| 2023-09-12T23:15:41
|
1893462121
|
{
"authors": [
"ElGigi",
"andrewdalpino"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2529",
"repo": "RubixML/ML",
"url": "https://github.com/RubixML/ML/issues/307"
}
|
gharchive/issue
|
Softmax Classifier & partial training
Hi,
In the documentation it is stated that partial training can be used to reduce memory consumption.
I tried to train a Softmax classifier with several datasets and partial methods.
But only the first labels of the train() method are known. If new labels are present in the dataset given to the partial() method, they are not taken into account.
Can Dataset object retain set of all labels after Labeled::fold() method?
Regards.
Yes, the first training set defines all the possible labels for the model. If you want to fold your dataset such that each fold has samples that correspond to all possible classes in the master dataset then you can use the straftifiedFold() method.
$folds = $dataset->stratifiedFold(5);
https://docs.rubixml.com/2.0/datasets/labeled.html#stratification
|
2025-04-01T06:37:30.615921
| 2022-02-18T12:06:28
|
1142939685
|
{
"authors": [
"DurkelTheDonkey",
"coolreader18"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2530",
"repo": "RustPython/RustPython",
"url": "https://github.com/RustPython/RustPython/issues/3556"
}
|
gharchive/issue
|
rustpython VM storing defined functions across interpreter calls.
Hi, I'm using the rustpython VM and I'm trying to define some functions in one call and then at a later time I'd like to call those functions an arbitrary amount of times. The trouble I'm running into is saving the state across those calls to the interpreter and I end up trying to call functions that are not defined.
use rustpython_vm::compile::Mode;
use rustpython_vm::Interpreter;
use rustpython_vm::import::import_file;
fn main() {
Interpreter::default().enter(|vm| {
let imp = import_file(vm, "mylib", "mylib".to_owned(), MAIN.to_owned()).unwrap();
let scope = vm.new_scope_with_builtins();
let code_obj = vm
.compile(
r#"import * from mylib
init()"#,
Mode::Exec,
"<embedded>".to_owned(),
)
.unwrap();
vm.run_code_obj(code_obj, scope).unwrap();
});
println!("Hello, rust!");
}
const MAIN: &str = r#"print("PYTHON MODULE INIT")
def init():
print("Python says init")
"#;
This will print PYTHON MODULE INIT as it should when MAIN is compiled. However it will panic when I try to import the library or if I try to run init without importing the module.
How do I save the state (or scope) of the compiled MAIN module and pass it to the interpreter in later calls?
I'm not exactly sure what you mean by passing it to the interpreter, but I'd assume you're later calling Interpreter::default().enter() multiple times? Cause the solution for that is to just save the interpreter instead of recreating it every time, let interp = Interpreter::default(); for _ in _ { interp.enter(|vm| { .. }); }
Alternatively, you could just move your whole program into the enter() call.
Well in this you never have the from mylib import * bit, so there's not init() in scope. If you want to preserve that you should save scope as well and then it'll all happen in the same scope.
Thank you, that clears up a lot of confusion, sorry for the inconvenience :)
|
2025-04-01T06:37:30.618404
| 2023-01-03T07:46:02
|
1517035363
|
{
"authors": [
"youknowone",
"yt2b"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2531",
"repo": "RustPython/RustPython",
"url": "https://github.com/RustPython/RustPython/pull/4412"
}
|
gharchive/pull-request
|
Implement integer format validation
Integer Format validation is not implemented.
>>>>> f"{4096:,o}"
'1,0000' # Should be raised `ValueError`
>>>>> f"{123456:_n}"
'123_456' # Should be raised `ValueError`
I've implemented format validation.
With this fix, following tests now pass.
test_format_spec_errors in test_types.py
test_int__format__ in test_types.py
Thank you!
|
2025-04-01T06:37:30.636618
| 2019-02-23T04:34:34
|
413655802
|
{
"authors": [
"OddCoincidence",
"codecov-io",
"windelbouwman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2532",
"repo": "RustPython/RustPython",
"url": "https://github.com/RustPython/RustPython/pull/525"
}
|
gharchive/pull-request
|
[IDEA] Use "extractor" pattern for native functions
First off, this is far from being complete, but I wanted to share it early to see if this is a direction we actually want to go in before continuing.
The idea is to apply the "extractor pattern" used by Rust web frameworks such as actix to native/rust python functions, so that the function signature determines things such as type checking, conversions to and from python objects, etc.
For example, a function that adds two ints would currently look like this:
fn add(vm: &mut VirtualMachine, args: PyFuncArgs) -> PyResult {
arg_check!(
vm,
args,
required = [(zelf, Some(vm.ctx.int_type())), (other, Some(vm.ctx.int_type())]
);
Ok(vm.ctx.new_int(get_value(a) + get_value(a)))
}
This isn't bad, but it's a lot for such a simple function, and for more complex functions the boilerplate overwhelms everything else.
The new way would look like this:
fn add(a: PyInt, b: PyInt) -> PyInt {
a + b
}
Additionally, the old way still works for when you need it, because PyFuncArgs impls FromPyFuncArgs, PyResult impls IntoPyObject, etc.
Some other neat consequences:
A function could take a VarArgs<T> parameter that shifts the rest of the positional args into a Vec.
FromPyObject impls for primitives could raising overflow exceptions so they wouldn't have to be scattered around everywhere as they are now
For functions that take a lot of keyword args, one can imagine a derive(FromFuncArgs) proc-macro.
Codecov Report
Merging #525 into master will decrease coverage by 1.58%.
The diff coverage is 40%.
@@ Coverage Diff @@
## master #525 +/- ##
==========================================
- Coverage 50.08% 48.49% -1.59%
==========================================
Files 68 68
Lines 13924 14271 +347
Branches 3465 3590 +125
==========================================
- Hits 6974 6921 -53
- Misses 5096 5494 +398
- Partials 1854 1856 +2
Impacted Files
Coverage Δ
vm/src/obj/objint.rs
47% <0%> (-25.53%)
:arrow_down:
vm/src/obj/objbool.rs
40.74% <33.33%> (-22.42%)
:arrow_down:
vm/src/obj/objrange.rs
35.13% <40%> (-23.69%)
:arrow_down:
vm/src/pyobject.rs
69.68% <53.65%> (-15.1%)
:arrow_down:
vm/src/obj/objstr.rs
59.29% <0%> (ø)
:arrow_up:
vm/src/stdlib/pystruct.rs
24.5% <0%> (ø)
:arrow_up:
... and 6 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 2d19486...c442dee. Read the comment docs.
This sounds good to me, I'll have a better look later on. I think it is okay if we have multiple options of doing the same thing. We now have arg_check macro for type checking, and manual checking, this would be a nice third option. I would like to add this option, but also leave the existing way available.
Could you work out your PyInt example in the code?
I would like to add this option, but also leave the existing way available.
Completely agree, the existing ways are fully supported, via an identity impl of PyNativeFuncFactory. There will certainly be cases where it's preferred to work with PyFuncArgs directly.
Could you work out your PyInt example in the code?
I decided to leave int_add alone because I haven't worked out the nicest way to model the NotImplemented constant in this system. Instead, I converted str_endswith, so that should provide a good example.
I think this is ready to merge now. The type checking needs more work, but that's fine for now since almost nothing is using this. (Also, I discovered some edge cases not handled by arg_check!, so I'll spend some time fixing these up for both in a future PR.)
|
2025-04-01T06:37:30.645124
| 2016-12-05T10:56:12
|
193466516
|
{
"authors": [
"YanisSOTO",
"alexbredy"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2533",
"repo": "RxSwiftCommunity/RxDataSources",
"url": "https://github.com/RxSwiftCommunity/RxDataSources/issues/80"
}
|
gharchive/issue
|
RxAnimatedDataSource crash on performBatchUpdates (line 0)
Hello,
I've got a small amount of crashes on Fabric that seem to be triggered sometimes when calling the performBatchUpdates method.
The error displayed is the following:
UITableView internal bug: unable to generate a new section map with old section count: 2 and new section count: 0
I am initialising the View Controller with 2 sections (0 cells until the data is loaded from the server) and never remove them. I only replace them by an updated version in my handler method, which returns a new set of 2 sections.
Any ideas are much appreciated! Thanks.
The issue was related to the table view's data source not being set (RxCocoa.swift data source not set fatal error). Once I fixed that this crash disappeared.
Hello, i've the same problem and i can't fix it... This crash appears randomly when i load for the first time a collectionView. Any idea ? I think it's a problem of section.
|
2025-04-01T06:37:30.647491
| 2022-03-04T12:29:17
|
1159595621
|
{
"authors": [
"RyanNerd"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2534",
"repo": "RyanNerd/rxchart-web",
"url": "https://github.com/RyanNerd/rxchart-web/issues/328"
}
|
gharchive/issue
|
Bug where editing a client modal and the DOB fields are being transversed
This is getting called every time a DOB field gets focus (not a big deal because clients DOB is rarely edited) but still shouldn't be calling the API when focus moves about.
API Call: https://domain.com/v1/resident/search
Body:
{
"where": [
["FirstName", "=", "PersonsFirstName"],
["LastName", "=", "PersonsLastName"],
["DOB_YEAR", "=", 1976],
["DOB_MONTH", "=", 10],
["DOB_DAY", "=", "21"],
["Id", "<>", 1038]
],
"withTrashed": true
}
Also one of the fields is null and shouldn't be:
index.js:1 Warning: `value` prop on `input` should not be null. Consider using an empty string to clear the component or `undefined` for uncontrolled components.
at input
at http://localhost:3000/static/js/vendors~main.chunk.js:81498:23
at div
at http://localhost:3000/static/js/vendors~main.chunk.js:80149:23
at div
at http://localhost:3000/static/js/vendors~main.chunk.js:84260:23
at http://localhost:3000/static/js/vendors~main.chunk.js:81805:23
at form
at http://localhost:3000/static/js/vendors~main.chunk.js:81198:23
at div
at http://localhost:3000/static/js/vendors~main.chunk.js:85670:27
at div
at div
at http://localhost:3000/static/js/vendors~main.chunk.js:82735:23
at div
at Transition (http://localhost:3000/static/js/vendors~main.chunk.js:118425:30)
at http://localhost:3000/static/js/vendors~main.chunk.js:80989:24
at DialogTransition
at http://localhost:3000/static/js/vendors~main.chunk.js:116102:24
at http://localhost:3000/static/js/vendors~main.chunk.js:82431:23
at ClientEdit (http://localhost:3000/static/js/main.chunk.js:12019:32)
at ClientHeader (http://localhost:3000/static/js/main.chunk.js:528:87)
at Main (http://localhost:3000/static/js/main.chunk.js:958:81)
at App (http://localhost:3000/static/js/main.chunk.js:363:97)
|
2025-04-01T06:37:30.660561
| 2024-11-24T02:20:03
|
2686889418
|
{
"authors": [
"RyleiC",
"canbeyy1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2536",
"repo": "RyleiC/MagickaForgeV2",
"url": "https://github.com/RyleiC/MagickaForgeV2/issues/2"
}
|
gharchive/issue
|
Suggestion for new update
hi can you add edit spells I want create new summon spellls or edit vanilla spells
hi can you add edit spells I want create new summon spellls or edit vanilla spells
Like Magicks? Right now that would require source-code modding- although a patcher tool might be possible in the future
If you wish, a tool like https://github.com/dnSpy/dnSpy can be used to modify that kind of stuff in the meantime if you have knowledge in c#
I dont know how to using dnspy but thanks to answer my question
|
2025-04-01T06:37:30.734716
| 2022-10-15T08:23:46
|
1410092662
|
{
"authors": [
"VolkerDrees",
"gregorwolf"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2537",
"repo": "SAP-samples/abap-platform-refscen-flight",
"url": "https://github.com/SAP-samples/abap-platform-refscen-flight/issues/9"
}
|
gharchive/issue
|
ABAP BTP Branch pulled with error "No authorization to change authorization field &/DMO/CNTRY&." in Free Tier
Hello SAP ABAP Reference Scenario Team,
I'm the admin user of my SAP BTP Free Tier ABAP Environment. I've just tried to pull the updates from the https://github.com/SAP-samples/abap-platform-refscen-flight/tree/BTP-ABAP branch. But I get this error message:
/DMO/CNTRY | AUTH | E | No authorization to change authorization field &/DMO/CNTRY&.
For reference also this screenshot:
I've also activated the Authorization Trace for my user but I can't find any failed authorization checks there.
Best Regards
Gregor
Hello Gregor,
the error can be ignored. See: https://github.com/SAP-samples/abap-platform-refscen-flight/tree/BTP-ABAP#download point nr. 8.
Best Regards,
Volker
To bad that the GitHub search doesn't find the README when searching for /DMO/CNTRY. But it now finds this issue.
|
2025-04-01T06:37:30.736216
| 2019-03-15T13:05:08
|
421509252
|
{
"authors": [
"gladius-mtl"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2538",
"repo": "SAP/cloud-commerce-spartacus-storefront",
"url": "https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/1691"
}
|
gharchive/issue
|
DOC: Commit messages don't need scope anymore
Update the commit guidelines to indicate that scope is no longer necessary in the commit messages.
See https://sap-cx.slack.com/archives/GAC8PULNN/p1552634597090800?thread_ts=1552583300.089600&cid=GAC8PULNN for more info.
Fixed in the wiki version of this file, in one of the many commits from GH-1725.
|
2025-04-01T06:37:30.740732
| 2019-04-12T21:40:13
|
432755567
|
{
"authors": [
"developpeurweb",
"mikelMHybris"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2539",
"repo": "SAP/cloud-commerce-spartacus-storefront",
"url": "https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/2013"
}
|
gharchive/issue
|
Part 2/2: Move component styles to storefrontstyles lib
Based on the example set by Tobias on GH-1907 we'll now move the following components as well:
[x] Search Results - Category Listing / product-view component
[x] Store Finder / schedule component
[x] Store Finder / store-finder-grid component
[x] Store Finder / store-finder-header component
[x] Store Finder / store-finder-list-item component
[x] Store Finder / store-finder-map component
[x] Store Finder / store-finder-search component
[x] Store Finder / store-finder-list component
[x] Store Finder / store-finder-search-result component
[x] Store Finder / store-finder-store-description component
[x] Store Finder / store-finder-stores-count component
[x] pwa/components/add-to-home-screen-banner/add-to-home-screen-banner.component.scss
[ ] ui/layout/header/header.component.scss
[x] ui/components/form-components/item-counter/item-counter.component.scss
[x] ui/components/form-components/star-rating/star-rating.component.scss
[x] ui/components/spinner/spinner.component.scss
the header was already done.
|
2025-04-01T06:37:30.744690
| 2019-06-03T09:01:31
|
451368726
|
{
"authors": [
"Platonn"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2540",
"repo": "SAP/cloud-commerce-spartacus-storefront",
"url": "https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/2821"
}
|
gharchive/issue
|
es2015 build not working
Angular 8 allows for differential loading - production build outputs 2 bundles:
for browsers supporting only es5
for browser supporting es2015
You can enable it by setting target: "es2015" in tsconfig of the storefront app.
Unfortunately, when running the es2015 version, there displays only the top loader and blank page:
There seems to be a problem with ngrx + angular + es2015. Here is the PR of angular-cli that claims to fix it:
https://github.com/angular/angular-cli/pull/ 14585
The PR claims to fix all those issues:
https://github.com/ngrx/platform/issues/ 1888 (effects not dispatched with build --prod)
https://github.com/ng-packagr/ng-packagr/issues/ 1307 (ES2015 problem since Angular 8 )
https://github.com/angular/angular-cli/issues/ 14613#issuecomment-498960867 (App that has lazy loading modules is not rendering any pages in prod mode after update to angular 8. )
|
2025-04-01T06:37:30.753813
| 2019-07-15T16:30:19
|
468223101
|
{
"authors": [
"RadhepS",
"hackergil"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2541",
"repo": "SAP/cloud-commerce-spartacus-storefront",
"url": "https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/3674"
}
|
gharchive/issue
|
[Master] Fix e2e regression tests
The following e2e regression tests are failing and need to be fixed:
billing-address
my-account/payment-methods
reset-password
register
https://jkmaster.test.c3po.b2c.ydev.hybris.com/job/spartacus-regression-tests/234/console
Ticket for billing address: https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/3687
Billing Address: Bug fix ticket created for it #3687
Payment-Methods: I'm not able to fail this locally
Reset-Password: Requires dev17 to be updated
Register: I'm not able to fail this locally
Going to hold this ticket as I'll try to fail these tests. A solution may be to add waits() as that is the likely reason for these tests failing on Jenkins.
Closing because neither Payment-Methods or Register have failed on Jenkins.
Tests are still failing in the Jenkins job that we just ran. If different tests are failing, it is ok to change this ticket description and create follow up tickets but please do not close this one until regression tests are stable
New tests to fix:
regression/my-account/update-profile.e2e-spec.ts
regression/site-context/language/language-registration-page.e2e-spec.ts
regression/site-context/language/language-cart-page.e2e-spec.ts
mobile/register-mobile.e2e-spec.ts
All e2e tests failing on Jenkins have either been fixed or already have tickets created to fix them.
This ticket will remain open and I will continue to monitor Jenkins.
Re-ran Jenkins.
Added new tests that are failing that have not previously been mentioned.
mobile/product-search-store-flow-mobile.e2e-spec.ts - Fixed (https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/3705)
New failure(s) on Jenkins:
language-my-account-pages.e2e-spec.ts
New failure(s) on Jenkins:
mobile/register-mobile.e2e-spec.ts
New failure(s) on Jenkins:
regression/my-account/payment-methods.e2e-spec.ts
regression/register.e2e-spec.ts
mobile/cart-mobile.e2e-spec.ts
New failure(s) on Jenkins:
New locale failure(s):
regression/my-account/order-history-orders-flow.e2e-spec.ts - Fixed (https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/3807)
New failure(s) on Jenkins:
regression/my-account/payment-methods.e2e-spec.ts - Created (https://github.com/SAP/cloud-commerce-spartacus-storefront/issues/3970)
|
2025-04-01T06:37:30.757299
| 2023-12-21T09:16:31
|
2052024913
|
{
"authors": [
"newtork"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2542",
"repo": "SAP/cloud-sdk-java",
"url": "https://github.com/SAP/cloud-sdk-java/pull/201"
}
|
gharchive/pull-request
|
Fix compatibility v4 / v5 with CAP
Currently CAP has to make the following code change in their productive code (non-test) for migrating to SDK v5:
- httpDestination = new DefaultDestination(destination.getProperties()).asHttp();
+ httpDestination = DefaultHttpDestination.fromMap(destination.getProperties()).build();
Implication:
There's a chance for users of SDK v5 + CAP v2 to experience runtime exception about constructor visibility.
We could solve it by making the constructor public again.
Uses in current CAP here:
https://github.wdf.sap.corp/cds-java/cds-services/blob/main/cds-feature-remote-odata/src/main/java/com/sap/cds/services/impl/odata/RemoteODataClient.java#L88
https://github.wdf.sap.corp/cds-java/cds-services/blob/main/cds-feature-remote-hcql/src/main/java/com/sap/cds/services/impl/hcql/RemoteHcqlClient.java#L95
We're planning to solve it differently.
Closed in favor of this
https://github.wdf.sap.corp/cds-java/cds-services/pull/3723
|
2025-04-01T06:37:30.811852
| 2024-01-03T10:34:10
|
2063728766
|
{
"authors": [
"nahmed22"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2543",
"repo": "SAP/open-ux-tools",
"url": "https://github.com/SAP/open-ux-tools/pull/1573"
}
|
gharchive/pull-request
|
feat: move cds annotation converter module
Move cds annotation converter module to open-ux-tools repository
obsolete : created new PR: https://github.com/SAP/open-ux-tools/pull/1609
|
2025-04-01T06:37:30.813042
| 2024-03-21T15:39:59
|
2200526325
|
{
"authors": [
"tobiasqueck"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2544",
"repo": "SAP/open-ux-tools",
"url": "https://github.com/SAP/open-ux-tools/pull/1764"
}
|
gharchive/pull-request
|
Allow adding tiles pointing to remotely available apps in the preview-middleware
Implements: #1763
I tested it on a FE project and it worked nicely. @tobiasqueck Is it also supposed to work on ADP projects?
Yes. I haven't tested it explicitly but the logic is independent of whether it is an app project or an ADP
|
2025-04-01T06:37:30.820203
| 2017-03-24T20:38:27
|
216898733
|
{
"authors": [
"aborjinik",
"boghyon"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2545",
"repo": "SAP/openui5",
"url": "https://github.com/SAP/openui5/issues/1396"
}
|
gharchive/issue
|
sap.m.Table: Hidden elements shift adjacent cells when "minScreenWidth" is applied to all Columns
OpenUI5 version: Since 1.44.5 (affects also current nightly 1.47)
Browser/version (+device/version): All
URL (minimal example if possible): https://embed.plnkr.co/OepBn5/
Steps to reproduce the problem:
Open the Plunker example above
Shrink the preview panel horizontally till columns are shown as pop-in
What is the expected result?
Until 1.44.4
What happens instead?
Since 1.44.5
Any other information? (attach screenshot if possible)
This effect goes away as soon as minScreenWidth of one of the columns is not applied (e.g. change one minScreenWidth value to "Phone" or remove it).
Hi @boghyon if you configure all the columns to be shown as a popin then we cannot have a physical table anymore. There must be always at least one visible real column which is not configured for popin.
Such feature has been asked several times but if we allow this we know people start using table like a form. Thats why this limitation will stay. Please see this to make it work
https://plnkr.co/edit/Dy30RGa3cY00kW7vi6Ii?p=preview
Hey @aborjinik , thank you for the answer. I couldn't find any similar issue about this on GitHub. Sorry about that. Could we at least get a "warning" message placed in the API reference of sap.m.Column that it's not recommended to apply demandPopin to all Columns? Or that at least one column should have demandPopin disabled.
For other readers: since UI5 1.72, at least one column is always kept in the header:
Related commit: https://github.com/SAP/openui5/commit/307b09bc71d3cc2cd6f3fc3f83ffa8c824aa6ee1
|
2025-04-01T06:37:30.875670
| 2019-10-03T12:25:34
|
502033903
|
{
"authors": [
"LindsayBert",
"thecodester"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2546",
"repo": "SAPDocuments/Tutorials",
"url": "https://github.com/SAPDocuments/Tutorials/pull/4013"
}
|
gharchive/pull-request
|
Automatic commit: Move cp-portal-cloud-foundry-getting-started from Q…
…A to Production
You can commit multiple tutorials all at once :-)
|
2025-04-01T06:37:30.898575
| 2015-04-08T17:35:54
|
67180413
|
{
"authors": [
"brigb123"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2547",
"repo": "SCIInstitute/SCI-Solver_FEM",
"url": "https://github.com/SCIInstitute/SCI-Solver_FEM/issues/11"
}
|
gharchive/issue
|
Make examples more visual to show the program works.
Either a graph for convergence, or an image, or similar. Make the examples more demonstrative.
an output VTK is created. 4ded8863c6eb06f13616e288ac7539c79ceaca09
|
2025-04-01T06:37:30.915470
| 2017-12-02T12:40:51
|
278681125
|
{
"authors": [
"SCullman",
"eXistenZe"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2548",
"repo": "SCullman/DNN.FormAndList",
"url": "https://github.com/SCullman/DNN.FormAndList/issues/7"
}
|
gharchive/issue
|
Error when using calculated fields
When I added a calculated field based on [intfield] + 10, I tested it editing some rows and the field was calculated fine.
However, when I try to create a new record, the green confirmation did not show, no error was shown on screen, and the link had: /def/errormessage
Checked the admin logs and I had this:
AbsoluteURL:/Default.aspx
DefaultDataProvider:DotNetNuke.Data.SqlDataProvider, DotNetNuke
ExceptionGUID:b4bad84e-6b90-4255-84eb-e6694121bf63
AssemblyVersion:9.1.1
PortalId:0
UserId:1
TabId:219
RawUrl:/...
Referrer:https://...
UserAgent:Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.94 Safari/537.36
ExceptionHash:QKZuiSMY28G3MbdBenUr2XM7FWU=
Message:The Controls collection cannot be modified because the control contains code blocks (i.e. <% ... %>).
StackTrace:
InnerMessage:The Controls collection cannot be modified because the control contains code blocks (i.e. <% ... %>).
InnerStackTrace:
at System.Web.UI.ControlCollection.Add(Control child) at DotNetNuke.Services.Exceptions.Exceptions.ProcessModuleLoadException(String FriendlyMessage, Control ctrl, Exception exc, Boolean DisplayErrorMessage)
Source:
FileName:
FileLineNumber:0
FileColumnNumber:0
Method:
Server Name: ...
ModuleId:864
ModuleDefId:134
FriendlyName:Form and List
ModuleControlSource:DesktopModules/UserDefinedTable/Default.ascx
AbsoluteURL:/Default.aspx
DefaultDataProvider:DotNetNuke.Data.SqlDataProvider, DotNetNuke
ExceptionGUID:96179f9a-4a6a-47ac-b035-1b2d1af5d045
AssemblyVersion:9.1.1
PortalId:0
UserId:1
TabId:219
RawUrl:/...
Referrer:https://...
UserAgent:Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.94 Safari/537.36
ExceptionHash:EQDfGxtjEmUKwh4x8o9tEIVTw1U=
Message:Object reference not set to an instance of an object.
StackTrace:
InnerMessage:Object reference not set to an instance of an object.
InnerStackTrace:
at DotNetNuke.Modules.UserDefinedTable.DataTypes.EditString.get_Value() at DotNetNuke.Modules.UserDefinedTable.EditForm.cmdUpdate_Click(Object sender, EventArgs e)
Source:
FileName:
FileLineNumber:0
FileColumnNumber:0
Method:
Server Name: ...
Deleted the calculated field, and it worked fine again.
Any chance that the number was bigger than 999? Then it is the same issue as #6
Way bigger.
Not sure how it is related, but when using the int field only, it does not
give any error. Only when I add a calculated field based on that int.
On Dec 2, 2017 15:37, "Stefan Cullmann"<EMAIL_ADDRESS>wrote:
Any chance that the number was bigger than 999? Then it is the same issue
as #6 https://github.com/SCullman/DNN.FormAndList/issues/6
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/SCullman/DNN.FormAndList/issues/7#issuecomment-348699769,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AZSlXzse7_-A3C1kAUWKVZ4ZHTUoJlW8ks5s8W62gaJpZM4QzSuc
.
Should have been fixed in fnl.6.3.3
|
2025-04-01T06:37:30.922117
| 2020-06-19T20:01:19
|
642191997
|
{
"authors": [
"mwes"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2549",
"repo": "SD2E/experimental-intent-parser",
"url": "https://github.com/SD2E/experimental-intent-parser/issues/194"
}
|
gharchive/issue
|
Add additional golden file tests (7)
In GitLab by @mweston on Apr 28, 2020, 18:46
The following experiments have been processed and should be put under golden file testing.
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/Microbe-LiveDeadClassification.json
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/NovelChassis-OR-Circuit-Cycle1-ObstacleCourse.json
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/YeastSTATES-CRISPR-Short-Duration-Time-Series-20191213.json
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/y4d_cen_pk_inducible_crispr_characterization.json
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/y4d_crispr_dose_response.json
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/NovelChassis-OR-Circuit-Cycle0-8hour.json
https://gitlab.sd2e.org/sd2program/cp-request/raw/master/input/structured_requests/NovelChassis-OR-Circuit-Cycle0-24hour.json
In GitLab by @mweston on Apr 28, 2020, 18:46
changed the description
In GitLab by @mweston on Apr 28, 2020, 18:46
changed the description
In GitLab by @mweston on May 4, 2020, 13:02
changed the description
In GitLab by @mweston on May 18, 2020, 13:56
@jakebeal @tramyn can this be added to 2.5.1? It should be straightforward as these documents have revision_ids. Given that golden file testing is manual/opt-in, I'd encourage supporting coverage against a larger set of files so that we don't miss anything.
In GitLab by @tramyn on May 18, 2020, 14:06
@mweston There are no issue with this request. I can go ahead and add them now.
In GitLab by @mweston on May 18, 2020, 14:07
Great, thanks!
In GitLab by @tramyn on May 18, 2020, 14:17
created merge request !114 to address this issue
In GitLab by @tramyn on May 18, 2020, 14:17
mentioned in merge request !114
In GitLab by @tramyn on May 18, 2020, 15:40
closed via merge request !114
In GitLab by @tramyn on May 18, 2020, 15:40
mentioned in commit 21124d71583b1bba922c3262382ba7fd3dd07680
In GitLab by @tramyn on Jun 12, 2020, 18:58
mentioned in commit b657fd402c2007f4d495d5efc01a68b4586762ec
|
2025-04-01T06:37:30.953705
| 2022-08-09T16:39:29
|
1333515722
|
{
"authors": [
"BugMonkey",
"Jeyhey",
"RMehdid",
"Sri2611",
"Teglgaard",
"dreampiggy",
"frlefebvre",
"garrettrayj",
"guidev",
"lukeredpath",
"tichise",
"yoshirozay"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2550",
"repo": "SDWebImage/SDWebImageSwiftUI",
"url": "https://github.com/SDWebImage/SDWebImageSwiftUI/issues/222"
}
|
gharchive/issue
|
Xcode 14 - SwiftUI warning
If you try to run the sample app on Xcode 14 beta 5 you get the following warnings.
+1
At the same time our logs are spammed with
[SwiftUI] Publishing changes from within view updates is not allowed, this will cause undefined behavior.
Also seeing this - this is a genuine bug within the library because it is triggering a method from within the view body that has side effects, one of which is mutating some published state:
Specifically, WebImage does this:
public var body: some View {
// This solve the case when WebImage created with new URL, but `onAppear` not been called, for example, some transaction indeterminate state, SwiftUI :)
if imageManager.isFirstLoad {
imageManager.load()
}
...snip...
}
I have no context as to why this change was added but it is fundamentally wrong. Side effects should not be performed from within a SwiftUI's view body - this body method can potentially called many times (every time the view's state changes). In this case, calling .load() not only triggers a side effect, but it causes the @Published isLoading to be mutated in the middle of the view update cycle.
+1
also having the same issue, with the bug affecting my navigationStack and moves me back to previous page, needs a fix asap, at least a workaround for now...
I noticed that the picture flickered a bit in iOS16 beta 5
🥲
I encountered a similar message in my environment.
+1
Any chance this gets fixed before iOS 16's release?
Hi.
This current repo's code, need a refactory to make it works with latest SwiftUI. And drop the usage of that ObservedObject using StateObject.
If someone has expert coding experience, PR is welcomed.
Currently I'm working on LLVM/Swift toolchain in this year, so may not have enough time to maintain this changes. But I'll try to back to this repo once I have some extra time.
+1
Following @lukeredpath's hint, I made some changes that seem to fix things on iOS 16.
https://github.com/SDWebImage/SDWebImageSwiftUI/compare/master...garrettrayj:ios16-undefined-behavior
I haven't tested AppKit and I bet backwards compatibility is messed up, but the changes appear mainly to be removing old workarounds that are no longer needed. Story of my life with SwiftUI, so I'm feeling confident enough to roll with 'em for the sake of getting iOS 16 updates out the door.
If you don't need backwards compatibility another option would be to see if the built in AsyncImage does what you need and have one less third party dependency in your codebase. That's my plan.
@garrettrayj - Great work - Will you make a PR? Then it can be reviewed.
I had the same error with iOS16 RC version.
This warning appears with the latest SwiftUI update but the issue impacts all SwiftUI versions. For instance, I have observed that screens that have a SDWebImage as first element are not correctly rendered in a navigation animation (i.e. when they swipe in from right to left).
Please have a try with v2.1.0
Seems to work perfectly now. Thank you!!
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.