id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
284638128 | Unable to download from investigationdiscovery.com
Please follow the guide below
You will be asked some questions and requested to provide some information, please read them carefully and answer honestly
Put an x into all the boxes [ ] relevant to your issue (like this: [x])
Use the Preview tab to see what your issue will actually look like
Make sure you are using the latest version: run youtube-dl --version and ensure your version is 2017.12.23. If it's not, read this FAQ entry and update. Issues with outdated version will be rejected.
[ ] I've verified and I assure that I'm running youtube-dl 2017.12.23
Before submitting an issue make sure you have:
[ ] At least skimmed through the README, most notably the FAQ and BUGS sections
[ ] Searched the bugtracker for similar issues including closed ones
What is the purpose of your issue?
[ ] Bug report (encountered problems with youtube-dl)
[ ] Site support request (request for adding support for a new site)
[ ] Feature request (request for a new functionality)
[ ] Question
[ ] Other
The following sections concretize particular purposed issues, you can erase any section (the contents between triple ---) not applicable to your issue
If the purpose of this issue is a bug report, site support request or you are not completely sure provide the full verbose output as follows:
Add the -v flag to your command line you run youtube-dl with (youtube-dl -v <your command line>), copy the whole output and insert it here. It should look similar to one below (replace it with your log inserted between triple ```):
[debug] System config: []
[debug] User config: []
[debug] Command-line args: [u'-v', u'http://www.youtube.com/watch?v=BaW_jenozKcj']
[debug] Encodings: locale cp1251, fs mbcs, out cp866, pref cp1251
[debug] youtube-dl version 2017.12.23
[debug] Python version 2.7.11 - Windows-2003Server-5.2.3790-SP2
[debug] exe versions: ffmpeg N-75573-g1d0487f, ffprobe N-75573-g1d0487f, rtmpdump 2.4
[debug] Proxy map: {}
...
<end of log>
If the purpose of this issue is a site support request please provide all kinds of example URLs support for which should be included (replace following example URLs by yours):
Single video: https://www.youtube.com/watch?v=BaW_jenozKc
Single video: https://youtu.be/BaW_jenozKc
Playlist: https://www.youtube.com/playlist?list=PL4lCao7KL_QFVb7Iudeipvc2BCavECqzc
Note that youtube-dl does not support sites dedicated to copyright infringement. In order for site support request to be accepted all provided example URLs should not violate any copyrights.
Description of your issue, suggested solution and other information
Explanation of your issue in arbitrary form goes here. Please make sure the description is worded well enough to be understood. Provide as much context and examples as possible.
If work on your issue requires account credentials please provide them or explain how one can obtain them.
./youtube-dl --cookies cookies.txt -F https://www.investigationdiscovery.com/tv-shows/web-of-lies/
[generic] web-of-lies: Requesting header
WARNING: Falling back on generic information extractor.
[generic] web-of-lies: Downloading webpage
[generic] web-of-lies: Extracting information
ERROR: Unsupported URL: https://www.investigationdiscovery.com/tv-shows/web-of-lies/
./youtube-dl -U
youtube-dl is up-to-date (2017.12.23)
./youtube-dl --cookies cookies.txt -F https://www.investigationdiscovery.com/tv-shows/web-of-lies/
[generic] web-of-lies: Requesting header
WARNING: Falling back on generic information extractor.
[generic] web-of-lies: Downloading webpage
[generic] web-of-lies: Extracting information
ERROR: Unsupported URL: https://www.investigationdiscovery.com/tv-shows/web-of-lies/
I'm thinking its not working because it is not investigationdiscoverygo.com? Youtube-dl worked before through the "GO" page. I think they've gone away with "GO"? Anyway to use Youtube-dl on the main website?
https://github.com/rg3/youtube-dl/issues/14954
| gharchive/issue | 2017-12-27T02:52:19 | 2025-04-01T06:45:37.908326 | {
"authors": [
"StevenDTX",
"captbanana"
],
"repo": "rg3/youtube-dl",
"url": "https://github.com/rg3/youtube-dl/issues/15082",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
336449774 | http 500 error on embedded vimeo video that was working a few hours ago
Please follow the guide below
You will be asked some questions and requested to provide some information, please read them carefully and answer honestly
Put an x into all the boxes [ ] relevant to your issue (like this: [x])
Use the Preview tab to see what your issue will actually look like
Make sure you are using the latest version: run youtube-dl --version and ensure your version is 2018.06.25. If it's not, read this FAQ entry and update. Issues with outdated version will be rejected.
[x] I've verified and I assure that I'm running youtube-dl 2018.06.25
Before submitting an issue make sure you have:
[x] At least skimmed through the README, most notably the FAQ and BUGS sections
[x] Searched the bugtracker for similar issues including closed ones
[x] Checked that provided video/audio/playlist URLs (if any) are alive and playable in a browser
What is the purpose of your issue?
[x] Bug report (encountered problems with youtube-dl)
[ ] Site support request (request for adding support for a new site)
[ ] Feature request (request for a new functionality)
[ ] Question
[ ] Other
youtube-dl --verbose --referer https://app.jumpcut.com/course/growth-hacking-facebook/video/raising-500k-in-1-month -i -f 'bestvideo[height<=720]+bestaudio/best[height<=720]' -o "Raising $500k in 1 Month.%(ext)s" https://app.jumpcut.com/course/growth-hacking-facebook/video/raising-500k-in-1-month
[debug] System config: []
[debug] User config: []
[debug] Custom config: []
[debug] Command-line args: [u'--verbose', u'--referer', u'https://app.jumpcut.com/course/growth-hacking-facebook/video/raising-500k-in-1-month', u'-i', u'-f', u'bestvideo[height<=720]+bestaudio/best[height<=720]', u'-o', u'Raising k in 1 Month.%(ext)s', u'https://app.jumpcut.com/course/growth-hacking-facebook/video/raising-500k-in-1-month']
[debug] Encodings: locale UTF-8, fs utf-8, out UTF-8, pref UTF-8
[debug] youtube-dl version 2018.06.25
[debug] Python version 2.7.10 (CPython) - Darwin-17.6.0-x86_64-i386-64bit
[debug] exe versions: ffmpeg 4.0, ffprobe 4.0
[debug] Proxy map: {}
[generic] raising-500k-in-1-month: Requesting header
WARNING: Could not send HEAD request to https://app.jumpcut.com/course/growth-hacking-facebook/video/raising-500k-in-1-month: HTTP Error 500: Internal Server Error
[generic] raising-500k-in-1-month: Downloading webpage
ERROR: Unable to download webpage: HTTP Error 500: Internal Server Error (caused by HTTPError()); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
File "/usr/local/bin/youtube-dl/youtube_dl/extractor/common.py", line 598, in _request_webpage
return self._downloader.urlopen(url_or_request)
File "/usr/local/bin/youtube-dl/youtube_dl/YoutubeDL.py", line 2211, in urlopen
return self._opener.open(req, timeout=self._socket_timeout)
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/urllib2.py", line 437, in open
response = meth(req, response)
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/urllib2.py", line 550, in http_response
'http', request, response, code, msg, hdrs)
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/urllib2.py", line 475, in error
return self._call_chain(*args)
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/urllib2.py", line 409, in _call_chain
result = func(*args)
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/urllib2.py", line 558, in http_error_default
raise HTTPError(req.get_full_url(), code, msg, hdrs, fp)
...
<end of log>
Earlier today I was running YT-DL for other videos and would have to run the same command 6-18 times depending on the video size it would "timeout" using a wired connection with NAS speeds of up to 700mbps I was only getting 6-11Mbps (presumably server side limitations) so I would have to repeated run the same command multiple times to get one video to download completely.
now roughly 12 hours later, when I run the command I get this Http 500 Error. does this mean YT-DL error or server I am trying to download from has an error? Is this a simple wait and try again later or something more complicated? thanks in advance.
Please close/delete, I realized my issue now, I used the referer URL as the source URL.
| gharchive/issue | 2018-06-28T02:22:49 | 2025-04-01T06:45:37.917361 | {
"authors": [
"elgallo"
],
"repo": "rg3/youtube-dl",
"url": "https://github.com/rg3/youtube-dl/issues/16839",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
43527582 | Request to download all playlists of a channel/user account
Hi together,
As a follow up on this [1] discussion I want to propose a new feature which allows the user to download all playlists of a provided channel/user account.
Unfortunately I cannot provide you with a pull request, but I am happy to think with you about the design of the implementation. :)
Looking forward to any comments. :+1:
[1] https://github.com/rg3/youtube-dl/issues/3659
This feature will be supported in the next version.
Dear @dstftw,
This is fantastic news. Thank you very much for sharing this. :)
~ Robert
| gharchive/issue | 2014-09-22T18:10:32 | 2025-04-01T06:45:37.919996 | {
"authors": [
"dstftw",
"orschiro"
],
"repo": "rg3/youtube-dl",
"url": "https://github.com/rg3/youtube-dl/issues/3817",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
55850799 | site support campus.codeschool
Please support http://campus.codeschool.com/courses/shaping-up-with-angular-js/
+1 for this.
From #10208: https://www.codeschool.com/courses/surviving-apis-with-rails/videos should be supported, too. This example requires login.
So what's the update , any current workaround?
Apparently nobody is working on it.
alrighty then. Guess I have to work on it.
So is Code School currently unsupported?
Any update on this one @picwellwisher12pk?
| gharchive/issue | 2015-01-29T03:13:08 | 2025-04-01T06:45:37.922920 | {
"authors": [
"InfiniteStyles",
"MichaelDeBoey",
"collegeimprovements",
"muhasturk",
"picwellwisher12pk",
"yan12125"
],
"repo": "rg3/youtube-dl",
"url": "https://github.com/rg3/youtube-dl/issues/4810",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
130742322 | [Errno 9] Bad file descriptor in Unreal Engine 4
I tried to include youtube-dl in Unreal Engine as a link parser, but it seems like it may have some conflicts with the engine. Any idea why this error may occur?
Read this and provide all required information.
| gharchive/issue | 2016-02-02T16:47:29 | 2025-04-01T06:45:37.924418 | {
"authors": [
"dstftw",
"h27han"
],
"repo": "rg3/youtube-dl",
"url": "https://github.com/rg3/youtube-dl/issues/8406",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
151924416 | Liveleak video cannot get download url.
[roo@local ~]$ youtube-dl -g http://www.liveleak.com/view?i=dc0_1461466199
ERROR: Unable to extract embed URL; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
[roo@local ~]$ youtube-dl -g http://www.liveleak.com/view?i=dc0_1461466199 --verbose
[debug] System config: []
[debug] User config: []
[debug] Command-line args: [u'-g', u'http://www.liveleak.com/view?i=dc0_1461466199', u'--verbose']
[debug] Encodings: locale ANSI_X3.4-1968, fs ANSI_X3.4-1968, out ANSI_X3.4-1968, pref ANSI_X3.4-1968
[debug] youtube-dl version 2016.04.24
[debug] Python version 2.7.5 - Linux-2.6.32-042stab104.1-x86_64-with-centos-7.2.1511-Core
[debug] exe versions: avconv 10.1, avprobe 10.1
[debug] Proxy map: {}
ERROR: Unable to extract embed URL; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
Traceback (most recent call last):
File "/usr/local/bin/youtube-dl/youtube_dl/YoutubeDL.py", line 673, in extract_info
ie_result = ie.extract(url)
File "/usr/local/bin/youtube-dl/youtube_dl/extractor/common.py", line 341, in extract
return self._real_extract(url)
File "/usr/local/bin/youtube-dl/youtube_dl/extractor/liveleak.py", line 87, in _real_extract
webpage, 'embed URL')
File "/usr/local/bin/youtube-dl/youtube_dl/extractor/common.py", line 644, in _search_regex
raise RegexNotFoundError('Unable to extract %s' % _name)
RegexNotFoundError: Unable to extract embed URL; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
I checked the link and it seems like this isn't a standard liveleak video but instead an embedded Facebook video.
I copied the link to the embedded FB video and it downloaded nicely with youtube-dl.
Long story short, to solve this problem we need to add support for embedded Facebook videos.
| gharchive/issue | 2016-04-29T17:11:19 | 2025-04-01T06:45:37.932000 | {
"authors": [
"EnginePod",
"dspstudio"
],
"repo": "rg3/youtube-dl",
"url": "https://github.com/rg3/youtube-dl/issues/9350",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
197088648 | Config Path by Setting
I am using Atom to run NodeJs in a portable environment. In the portable environment, I cannot change the environment path variable.
Is there any config files or settings that I can change the path variable? Or How could I set the node.exe path.
Hi.
Script - > Run with options -> fill first field as you like
@nixel2007 Is it under Menu, Package -> Script
I cannot find this option in my Atom.
My Atom version: 1.11.2
Script version: 3.13.0
I finally set the environment in the Init Script.
process.env.PATH = [, process.env.PATH].join(";");
@imben1109 , so I'm not too sure what I've done here and why your suggestion worked.
I am running everything portable (Atom and python) via a USB drive on windows.
Obviously if I'm running portable then python won't be in my path, so I wanted to tell Script where to look for the python.exe.
Without knowing what I was going I copied the following into my init.coffee script in Atom.
process.env.PATH = ["\\program_files\\python_3.6.0.1\\python-3.6.0.amd64"] which is the path to my python.exe
And now Script will run .py files fine. Great!
But how does this actually work?
How do I add more paths, for example to python2.7? (which incidentally is \\program_files\\python_2.7.10.3\\python-2.7.10.amd64
And how do I get Script to run python2 over python3 in this scenario?
@jackjameshoward
process.env.PATH = ["\program_files\python_3.6.0.1\python-3.6.0.amd64"] is similar to set the System Path. Therefore, the python command can be executed in atom environment.
| gharchive/issue | 2016-12-22T03:46:51 | 2025-04-01T06:45:37.938983 | {
"authors": [
"imben1109",
"jackjameshoward",
"nixel2007"
],
"repo": "rgbkrk/atom-script",
"url": "https://github.com/rgbkrk/atom-script/issues/1189",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
74952470 | Info that shows info after running script appears twice.
I tried in an html document, which I knew wasn't support, a javascript and a blank document. All three produced the same result - two dialogues. Also, when you close one dialogue, the other closes with it.
I updated to Script 2.22.0 not too long ago and this started to happen. Beforehand, only one instance of the info appeared.
HTML document:
JavaScript document:
Blank document:
Windows 7
Atom 0.196.0
Script 2.22.0
In my experience this only happens when running something after updating script without "restarting" atom. If you close atom, open it again and try to run something, everything should work as expected.
Just did that now and it's fixed itself... Now don't I feel stupid. Thanks, and sorry for my inablility to think about actually closing and reopening.
Thanks, and sorry
No problem, thanks for filing an issue with details. The first time I ran into that I was super confused.
| gharchive/issue | 2015-05-10T16:28:13 | 2025-04-01T06:45:37.943243 | {
"authors": [
"JustCameron",
"rgbkrk",
"rodrigomb81"
],
"repo": "rgbkrk/atom-script",
"url": "https://github.com/rgbkrk/atom-script/issues/440",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2533619174 | Automated project now working
Updated to pipelines v2:
Pipeline server
Elyra secret
Elyra rolebinding
DS connection to use route
Note: Currently using wrong notebook image, needs to update after the imagepuller has been fixed
#113
lgtm
| gharchive/pull-request | 2024-09-18T12:33:31 | 2025-04-01T06:45:37.945103 | {
"authors": [
"RHRolun",
"guimou"
],
"repo": "rh-aiservices-bu/parasol-insurance",
"url": "https://github.com/rh-aiservices-bu/parasol-insurance/pull/127",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
858181976 | Unable to load in
Looks like a database error, which could be likely given how out of date it is. Feel free to create a fix and pull request, and I'll push it to the main branch.
Yeah I dont code.. I wanted to try the mode to see if I would be able to set it up as intended and then get some devs on it to uupdate and keep it maintained
| gharchive/issue | 2021-04-14T19:01:33 | 2025-04-01T06:45:37.947641 | {
"authors": [
"dodim5",
"rhapidfyre"
],
"repo": "rhapidfyre/Cops-n-Robbers",
"url": "https://github.com/rhapidfyre/Cops-n-Robbers/issues/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1751171036 | Example VFS locking can have name collisions across filesystems
The example IndexedDB and OPFS VFS classes use WebLocks to implement locking. The current implementation locks a database file by acquiring a WebLock with the filename. This naming scheme is not specific enough because there may be database files with the same name across different filesystems in the same origin.
For example, there can be two IndexedDB databases, "idb1" and "idb2", with each containing an SQLite database named "foo.db". These SQLite databases will use the same named WebLock(s). This could cause unnecessary degraded performance.
The fix needs to made in each IndexedDB and OPFS VFS that uses WebLocks (so not AccessHandlePoolVFS, which doesn't need locks because it doesn't support concurrency) to make the lock name unique across filesystems and filenames. A workaround is to use unique SQLite database names if your application uses multiple filesystems.
There has been no engagement on this bug and it's easy to work around if necessary. I'm marking it won't fix.
| gharchive/issue | 2023-06-11T00:33:35 | 2025-04-01T06:45:37.949633 | {
"authors": [
"rhashimoto"
],
"repo": "rhashimoto/wa-sqlite",
"url": "https://github.com/rhashimoto/wa-sqlite/issues/93",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
738676926 | Feature request: Circular List
It would be nice to see an option to enable a circular picker.
As an example, if I use numbers 0 to 59, presently, the scroll picker will start at 0 and end at 59. Instead, can it have the last element(59) before the first element(0), such that the picker will be able to scroll infinitely?
@rheng001 is this issue solved?
| gharchive/issue | 2020-11-09T05:00:04 | 2025-04-01T06:45:37.950960 | {
"authors": [
"JairajJangleSLIOT"
],
"repo": "rheng001/react-native-wheel-scrollview-picker",
"url": "https://github.com/rheng001/react-native-wheel-scrollview-picker/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
265522665 | Bump to Elm 0.18
Bump elm-test
Add descriptions to test to make elm-test work (descriptions are mandatory now)
Thanks!
| gharchive/pull-request | 2017-10-14T20:49:10 | 2025-04-01T06:45:37.970711 | {
"authors": [
"rhofour",
"tryzniak"
],
"repo": "rhofour/elm-pairing-heap",
"url": "https://github.com/rhofour/elm-pairing-heap/pull/7",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2583849136 | 🛑 BOS2 is down
In 69475e2, BOS2 ($URL_BOS2) was down:
HTTP code: 0
Response time: 0 ms
Resolved: BOS2 is back up in 6ca59b6 after 30 minutes.
| gharchive/issue | 2024-10-13T10:11:23 | 2025-04-01T06:45:37.972925 | {
"authors": [
"rholak"
],
"repo": "rholak/upptime",
"url": "https://github.com/rholak/upptime/issues/129",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2686858023 | 🛑 BOS2 is down
In bbf0b6b, BOS2 ($URL_BOS2) was down:
HTTP code: 0
Response time: 0 ms
Resolved: BOS2 is back up in 6e8fd4f after 2 hours, 25 minutes.
| gharchive/issue | 2024-11-24T01:01:23 | 2025-04-01T06:45:37.975071 | {
"authors": [
"rholak"
],
"repo": "rholak/upptime",
"url": "https://github.com/rholak/upptime/issues/1535",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
260753708 | Update Jenkins and Nexus To Display As Applications in OCP
Like Sonar does
This used to work in OCP 3.5 and I am not sure why it is not showing up that way anymore. Suggestions?
I found the answer. I need to add some metadata to the templates:
labels:
app: openshift-sonarqube
appName: sonarqube
name: openshift-sonarqube
| gharchive/issue | 2017-09-26T20:10:32 | 2025-04-01T06:45:37.976949 | {
"authors": [
"InfoSec812",
"sherl0cks"
],
"repo": "rht-labs/labs-ci-cd",
"url": "https://github.com/rht-labs/labs-ci-cd/issues/10",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
145424571 | Refactor the display annotations method
The NSSet methods doesn't work. I try array methods. Now the toAdd- and toRemove-Sets are correct. Unfortunately it isn't possible to use the Equatable protocol for FBAnnotation and FBAnnotationCluster.
Thanks @AFcgi! I merged in this PR manually, after resolving a conflict.
| gharchive/pull-request | 2016-04-02T22:17:29 | 2025-04-01T06:45:37.982897 | {
"authors": [
"AFcgi",
"chenr2"
],
"repo": "ribl/FBAnnotationClusteringSwift",
"url": "https://github.com/ribl/FBAnnotationClusteringSwift/pull/28",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1548860793 | DirectoryNotFoundException: Could not find a part of the path
Hi
I always get this error
DirectoryNotFoundException: Could not find a part of the path "/Users/xxx/Desktop/Unity Projects/xxx /my/explicit/savefile/pat/my-save-file.data".
The code is this (config)
// build your config
var config = new FBPPConfig()
{
SaveFileName = "my-save-file.data",
AutoSaveData = false,
ScrambleSaveData = true,
EncryptionSecret = "my-secret",
SaveFilePath = "my/explicit/savefile/pat"
};
// pass it to FBPP
FBPP.Start(config);
any news ?????
Reading at the code you posted it seems you just copied the explanatory one?
For example, this is what i used. I added an empty gameobject in the Main menu with this script it in.
using System.Collections;
using System.Collections.Generic;
using UnityEngine;
public class SaveData : MonoBehaviour
{
void Awake()
{
var config = new FBPPConfig();
{
config.SaveFileName = "saveData.txt";
config.AutoSaveData = true;
config.ScrambleSaveData = false;
config.SaveFilePath = Application.persistentDataPath;
}
FBPP.Start(config);
}
}
You can change the settings to your liking, each setting is in the initialization part of the documentation.
| gharchive/issue | 2023-01-19T10:12:33 | 2025-04-01T06:45:37.999579 | {
"authors": [
"aminos1941",
"rubenoriginal"
],
"repo": "richardelms/FileBasedPlayerPrefs",
"url": "https://github.com/richardelms/FileBasedPlayerPrefs/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
139388678 | Added release.sh to simplify releases
This PR was originally opened as #124 but it was pointing to the wrong branch
Now that we are publishing the PyPI, the release process is becoming more complex. At this stage, I typically like to automate things to make sure I don't miss a step and have to re-release all over.
In this PR:
Added release.sh
Takes care of bumping __version__, creating a git commit, creating a git tag, pushing new commit and tag, and publishing .gz.tar and .zip to PyPI
Thanks :)
| gharchive/pull-request | 2016-03-08T20:45:24 | 2025-04-01T06:45:38.144079 | {
"authors": [
"richrd",
"twolfson"
],
"repo": "richrd/suplemon",
"url": "https://github.com/richrd/suplemon/pull/131",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
638620380 | -e SQLPAD_TIMEOUT_SECONDS=1000 useless
docker timeoutSeconds useless:
about 50s show Server responded not ok
Please do not create duplicate issues. I see you already commented in https://github.com/rickbergfalk/sqlpad/issues/483. Will continue the conversation there.
| gharchive/issue | 2020-06-15T07:41:41 | 2025-04-01T06:45:38.145479 | {
"authors": [
"kpioneer123",
"rickbergfalk"
],
"repo": "rickbergfalk/sqlpad",
"url": "https://github.com/rickbergfalk/sqlpad/issues/688",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2348625548 | No reduction in VRAM usage
I tried running the following code, with just having the ```ridger/MMfreeLM-1.3B```` model initialized:
root@r4-0:~/matmulfreellm# python
>>> import os
>>> os.environ["TOKENIZERS_PARALLELISM"] = "false"
>>> import mmfreelm
>>> from transformers import AutoModelForCausalLM, AutoTokenizer
>>> # Change here to our open-sourced model
>>> name = "ridger/MMfreeLM-1.3B"
>>> tokenizer = AutoTokenizer.from_pretrained(name)
>>> model = AutoModelForCausalLM.from_pretrained(name).cuda().half()
Having another terminal opened with 'watch rocm-smi', showing 68% VRAM usage meaning about 5.5GB
Every 2.0s: rocm-smi r4-0: Wed Jun 12 12:16:17 2024
======================================== ROCm System Management Interface ========================================
================================================== Concise Info ==================================================
Device [Model : Revision] Temp Power Partitions SCLK MCLK Fan Perf PwrCap VRAM% GPU%
Name (20 chars) (Edge) (Socket) (Mem, Compute)
==================================================================================================================
0 [RX Vega64 : 0xc1] 30.0°C 11.0W N/A, N/A 852Mhz 167Mhz 9.41% auto 220.0W 68% 0%
Vega 10 XL/XT [Radeo
==================================================================================================================
============================================== End of ROCm SMI Log ===============================================
Contradicting what was said in the paper?
Hi,
We have highlighted in the paper that we use BitBLAS for conducting those experiments. However, BitBLAS can be challenging to install and is only compatible with NVIDIA GPUs. In fact, we even had to recompile it during our installation process. We haven't merged it into this repo yet. Additionally, due to the different ways FuseBitLinear stores weights, there is still some compatibility work that needs to be completed.
We are also working on merging MatmulFreeLLM into BitBLAS examples. In the meantime, you can try Bitnet's example to achieve a similar level of VRM reduction, which should be comparable to our model.
I see, so we would still have to wait for the repo to be fully functionally working with BitBLAS until that we can not experience the results from the paper nor do training, right?
For training it is okay, since we have integrated triton in our current repo, so you can still enjoy the accelerated training, for inference maybe not…
Wait, so you could still train a model and get faster training + vram reduction? It just doesn't work on inference? I might be wrong here but how would we evaluate the model during and after training for the losses, ouputs?
A little bit of context, I'm wanting to train a video generative model
You can refer to an and b, these two figures show that how our fused bilinear help to reduce the memory and training speed. (in pure MLP situation)
Hi,
I tested FusedBitLinear vs nn.Linear using a small MLP and I don't see any time speedup for training, in fact it is slower. Here is my model and training time curves for a batch size of 32 run for 10 epochs.
On testing FusedBitLinear vs vanilla BitLinear the time is similar to what you showed.
If you are using BitBLAS for inference, how is it matmul free? Does that not just use mixed precision multiplication?
Hi,
The fused BitLinear in small MLPs will not be significantly accelerated with Triton. You can verify this by testing cases where in_features/out_features > 2048.
Regarding Matmul-free operations: As mentioned in our paper, modern GPUs actually don't benefit from Matmul-free approaches, which is why we developed our own FPGA hardware implementation. That's why we still maintain Matmul operations in our code. Our solution is compatible with both Matmul-free and Matmul approaches. While Matmul-free can provide benefits on custom hardware implementations, retaining Matmul operations often yields better performance on general-purpose GPUs. Therefore, we use the fused version to leverage GPU training speed, while utilizing custom hardware to fully benefit from Matmul-free operations.
| gharchive/issue | 2024-06-12T12:16:34 | 2025-04-01T06:45:38.160698 | {
"authors": [
"pranav-asthana",
"radna0",
"ridgerchu"
],
"repo": "ridgerchu/matmulfreellm",
"url": "https://github.com/ridgerchu/matmulfreellm/issues/17",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2045478899 | Jittery Mouse
from humancursor import SystemCursor import random sleep_time = random.uniform(0.03, 0.1) cursor = SystemCursor() cursor.click_on([2800,540],click_duration=sleep_time,steady=True)
Very simple code, I tried adjusting the duration to close to 0, and the mouse jitters very slowly to destination, takes around 15 seconds to move 300 down and 200 right (roughly). Tried disabling any mouse related software - still same issue persists.
Any ideas?
The click_duration parameter has nothing to do with the duration of the movement, the time it takes from point A to B. It is just the time it holds down the left button of the mouse when clicking, it's not really your fault you got confused because the duration parameter should have been there. (I will add it soon).
For the slow and jittery movement, it is from steady=True, which tries to mimic human doing the movement in a straight line, which tends to be slower to do and unstable, jittery.
The only problem I see is, it is taking it 15 seconds to move 300 by 200 pixels, which with even those parameters (steady=True which should only longen it a bit), is very long and abnormal, I will check it out and hopefully fix it soon.
The click_duration parameter has nothing to do with the duration of the movement, the time it takes from point A to B. It is just the time it holds down the left button of the mouse when clicking, it's not really your fault you got confused because the duration parameter should have been there. (I will add it soon).
For the slow and jittery movement, it is from steady=True, which tries to mimic human doing the movement in a straight line, which tends to be slower to do and unstable, jittery.
The only problem I see is, it is taking it 15 seconds to move 300 by 200 pixels, which with even those parameters (steady=True which should only longen it a bit), is very long and abnormal, I will check it out and hopefully fix it soon.
I was editing the lib itself with duration, as system cursor doesn't seem to have a duration variable you can set (unless I messed up big somewhere). I'll try again without steady but I'm pretty sure it was still very slow
Yeah, you can just momentarily do that if you dont want to wait me, just add parameter duration=None on the click_on() function, and add duration=duration inside this line self.move_to(point, steady=steady). I will find what could be wrong with the steady option.
https://i.imgur.com/ZeyUqIc.gif
No matter how little the duration is set to, this is the result, if I set it absurdly high, it will take roughly the time I set it as.
No matter how short the duration is set to, this is the result.
If i set it to say 120, it does take a lot longer.
No matter how short the duration is set to, this is the result. If i set it to say 120, it does take a lot longer.
I have been testing almost every scenario, also your exact code, but can't seem to replicate this error you are encountering. Maybe it is something local on your pc, because I can't seem to find out what's wrong on my side. My last thought is the computing power, maybe try changing the target_points of human_curve = HumanizeMouseTrajectory(...) on the move_to function of SystemCursor class, to target_points=int(target_points / 2) or even int(target_points / 3). This may help fasten and smoothen the movement if your computer is slow. I don't have any other idea i'm sorry, if you find the solution let me know.
Tried disabling my secondary monitor, lowering the resolution on my ultrawide, disabling logitech mouse software, nothing helped.
I tried another HMM lib, the mouse has no issue moving with that one (it however has issues with either 1ms mouse clicks or a minimum of 100ms, whilst realistic would be ~30-80)
Update
Completely wiped python from my pc, reinstalled, and now it seems to be working fine. Very odd since before this I tried forcing upgrade on all dependencies and everything.
Don't worry, hopefully you solved it.
Hey, had the same issue on macos. What fixed it for me was setting pyautogui.DARWIN_CATCH_UP_TIME = 0. Not sure if this a good idea or not since the pyautogui documentation makes it sound like it's necessary:
There is also a pyautogui.DARWIN_CATCH_UP_TIME setting which adds an additional delay on macOS after keyboard and mouse events, since the operating system appears to need a delay after PyAutoGUI issues these events. It is set to 0.01 by default, adding an additional hundredth-second delay.
But it seems to be working for me, so going to run with this for now.
| gharchive/issue | 2023-12-18T00:59:00 | 2025-04-01T06:45:38.197295 | {
"authors": [
"Stivijs97",
"ekerik220",
"riflosnake"
],
"repo": "riflosnake/HumanCursor",
"url": "https://github.com/riflosnake/HumanCursor/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
753764100 | Updating docker file
This updates the dockerfile to allow for the base packages to be updated before they get released into the build. Adds an additional layer of security by keeping up to date packages
Accidentally was submitting a pull request to the wrong branch, apologies
| gharchive/pull-request | 2020-11-30T20:04:02 | 2025-04-01T06:45:38.209000 | {
"authors": [
"jacoknapp"
],
"repo": "riking/AutoDelete",
"url": "https://github.com/riking/AutoDelete/pull/22",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2122379761 | Unable to control cursor
Describe the bug
Is this to be expected?
Configuration
swipe direction="s" fingers=2 end="mkdir -p /tmp/gg/swipe_start"
pinch direction="in" fingers=2 end="mkdir -p /tmp/gg/pinch_start"
hold fingers=2 action="mkdir -p /tmp/gg/hold"
I'm not quite sure what the problem is, could you explain what is happening? One problem could be using 2 fingers for swipe gestures, the desktop environment and apps use this for scrolling.
could you explain what is happening
What I mean is that when I use one finger to move on the touchpad, the mouse cursor in the screen does not follow my finger!
One problem could be using 2 fingers for swipe gestures, the desktop environment and apps use this for scrolling.
But I've turned off Xorg's Libinput's capture of touchpad events as per your doc.
/usr/share/X11/xorg.conf.d/99-ignore-device.conf
But I've turned off Xorg's Libinput's capture of touchpad events as per your doc.
I read the code for gestures, is it because gestures doesn't handle input::event::pointer::PointerEvent? I'm not sure.
https://github.com/riley-martin/gestures/blob/a6447d422b8fd6ca45b0991af49c33c6b7c13871/src/event_handler.rs#L94
the desktop environment and apps use this for scrolling.
Also I don't use any other desktop environments besides xorg and awesome wm
What I mean is that when I use one finger to move on the touchpad, the mouse cursor in the screen does not follow my finger.
That is because you told x11 to ignore your device. Libinput recognizes different types of events; moving the cursor and scrolling are different from multi-touch gestures. To make this tool work properly, you should disable multi-touch events but not totally ignore the device.
I'm not familiar with your setup so I can't tell you how to do this, but the setting is probably in the awesomewm configuration rather that in the x11 configuration.
What I mean is that when I use one finger to move on the touchpad, the mouse cursor in the screen does not follow my finger.
That is because you told x11 to ignore your device. Libinput recognizes different types of events; moving the cursor and scrolling are different from multi-touch gestures. To make this tool work properly, you should disable multi-touch events but not totally ignore the device.
I'm not familiar with your setup so I can't tell you how to do this, but the setting is probably in the awesomewm configuration rather that in the x11 configuration.
I was able to get it working again today, I'm guessing it might have been because I added the user to the input group
What I mean is that when I use one finger to move on the touchpad, the mouse cursor in the screen does not follow my finger.
That is because you told x11 to ignore your device. Libinput recognizes different types of events; moving the cursor and scrolling are different from multi-touch gestures. To make this tool work properly, you should disable multi-touch events but not totally ignore the device.
I'm not familiar with your setup so I can't tell you how to do this, but the setting is probably in the awesomewm configuration rather that in the x11 configuration.
Tankyou for help
| gharchive/issue | 2024-02-07T07:39:11 | 2025-04-01T06:45:38.218758 | {
"authors": [
"riley-martin",
"yyy33"
],
"repo": "riley-martin/gestures",
"url": "https://github.com/riley-martin/gestures/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2621332792 | Task`1[RingCentral.RestClient]' is not assignable to service 'RingCentral.RestClient'.
When trying to register a RestClient, I get this exception
System.ArgumentException: The type 'System.Threading.Tasks.Task1[RingCentral.RestClient]' is not assignable to service 'RingCentral.RestClient'. at Autofac.Builder.RegistrationBuilder.CreateRegistration(Guid id, RegistrationData data, IInstanceActivator activator, IResolvePipelineBuilder pipelineBuilder, Service[] services, IComponentRegistration target) at Autofac.Builder.RegistrationBuilder.CreateRegistration[TLimit,TActivatorData,TSingleRegistrationStyle](IRegistrationBuilder3 builder)
at Autofac.Builder.RegistrationBuilder.RegisterSingleComponent[TLimit,TActivatorData,TSingleRegistrationStyle](IComponentRegistryBuilder cr, IRegistrationBuilder3 builder) at Autofac.RegistrationExtensions.<>c__DisplayClass42_01.b__0(IComponentRegistryBuilder cr)
at Autofac.ContainerBuilder.Build(IComponentRegistryBuilder componentRegistry, Boolean excludeDefaultModules)
at Autofac.ContainerBuilder.UpdateRegistry(IComponentRegistryBuilder componentRegistry)
at Autofac.Module.Configure(IComponentRegistryBuilder componentRegistry)
at Autofac.Core.Registration.ModuleRegistrar.<.ctor>b__1_0(IComponentRegistryBuilder reg)
at Autofac.ContainerBuilder.Build(IComponentRegistryBuilder componentRegistry, Boolean excludeDefaultModules)
at Autofac.ContainerBuilder.Build(ContainerBuildOptions options)
at Autofac.Extensions.DependencyInjection.AutofacServiceProviderFactory.CreateServiceProvider(ContainerBuilder containerBuilder)
at Microsoft.Extensions.Hosting.HostApplicationBuilder.Build()
at Microsoft.AspNetCore.Builder.WebApplicationBuilder.Build()
// using autofac
builder.Register(async c =>
{
var restClient = new RestClient(Settings.ClientId, Settings.ClientSecret, Settings.Server);
await restClient.Authorize(Settings.JwtToken); // if I remove the authorize, it seams to be ok
return restClient;
}).As<RestClient>().SingleInstance();
Yes indeed... my bad, I took it for granted, but since it's async, it returns a Tas
| gharchive/issue | 2024-10-29T13:51:41 | 2025-04-01T06:45:38.234109 | {
"authors": [
"fp-jcorriveau"
],
"repo": "ringcentral/RingCentral.Net",
"url": "https://github.com/ringcentral/RingCentral.Net/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
387479645 | Use material-colors instead of google-material-color
fixes #186
I think this change makes sense because material-colors seems more maintained than google-material-color (last commit was ~4 months ago vs ~2 years ago), and (as far as I know) they provide basically the same features.
Seems like a good fix.
@riophae Can this be merged?
| gharchive/pull-request | 2018-12-04T21:06:21 | 2025-04-01T06:45:38.248278 | {
"authors": [
"SkoricIT",
"jaimesanz",
"m-zuro"
],
"repo": "riophae/vue-treeselect",
"url": "https://github.com/riophae/vue-treeselect/pull/198",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1405103149 | Add LIFE RIPARIAS project page to dataset
LIFE RIPARIAS project description is ready:
https://docs.google.com/document/d/1SyZmiLcpFJ2AcDaDFZfnqrndwQFaUcxr3eKHy3fJKQY/edit#
To be added in metadata together with project ID: LIFE RIPARIAS.
Done. Thanks @LienReyserhove
| gharchive/issue | 2022-10-11T19:05:30 | 2025-04-01T06:45:38.254471 | {
"authors": [
"damianooldoni"
],
"repo": "riparias/anb-crayfishes-occurrences",
"url": "https://github.com/riparias/anb-crayfishes-occurrences/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
189502574 | Add 12904 annotation
@mshuler to review
+1
| gharchive/pull-request | 2016-11-15T20:48:51 | 2025-04-01T06:45:38.267522 | {
"authors": [
"mshuler",
"ptnapoleon"
],
"repo": "riptano/cassandra-dtest",
"url": "https://github.com/riptano/cassandra-dtest/pull/1387",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2186759037 | Feedback on RISC-V N-Trace Specification 1.0.0 rc2.0 - Chapter 2
Please update the terminology used in Chapter 2 (and the rest of the
specification) as follows:
itype Description
0 No special type
1 Exception
2 Interrupt
3 Trap return
4 Not-taken branch
5 Taken branch
6 Indirect jump (with or without linkage)
7 Reserved
8 Indirect call
9 Direct call
10 Indirect jump (without linkage)
11 Direct jump (without linkage)
12 Co-routine swap
13 Function return
14 Other indirect jump (with linkage)
15 Other direct jump (with linkage)
Both ECALL and EBREAK cause synchronous exceptions, and are not considered
to retire. The epc reported by both is the pc of the instruction itself and
not the address of the following instruction. In table 2 - "ecall is
reported after retirement" is wrong as ecall does not "retire".
For 4-bit itype encodings, The rd != link should be classified further
into two classes - 15 (direct jump with linkage) when rd != x0 and 11
(direct jump without linkage) when rd = x0.
c.j and cm.jt are misclassified. They should be classified as type 11
(direct jump without linkage) and not 15.
"If Pop returns the same address as PC at next valid ingress
port cycle, emit Indirect Branch message with B-TYPE=0.". This needs to be
corrected as the Indirect Branch messages hould be generated if the PC is
not the same as the address Poped from the return address stack.
Table 3 - "Add 0 as least significant bit to HIST field.". Consider updating
to just "Update HIST field" since HIST is not introduced yet.
"Encoder must handle call stack action as described in the Implicit Return
Optimization chapter." Update to clarify that the encoder must handle call
stack actions only if the Implicit return Optimization is implemented.
The last non-normative note - "If optional trTeInstEnAllJumps..." should be
a normative statement.
Editorial and typos:
9. "Table below provides.." -> "The table below provides.."
11."detailed mapping of of causes" -> remove repeated of. This paragraph can
also use rewriting for clarity. Use "operands" instead of "arguements".
All instructions should use uppercase. For instance, "mret" -> "MRET".
13."N-trace is using the..." -> "N-trace uses the same..."
14."detailed mapping of of" -> "detailed mapping of"
15."When ingress port is implemented as 4-bit" -> "When the itype input of
the ingress port is 4-bit wide"
16.Table 4 - Reseved -> Reserved
17."other instruction" -> "Other instructions"
"not listed in this table" - "All other instructions that are not listed in
this table".
"Implicit x1" -> The "Expands to jal x1, offset"
"N-trace encoder does not require cause ..." to
"The N-Trace encoder does not require 'cause' and 'tval' ingress port
signals, which are valid only for exceptions and interrupts, as these
details are not reported in N-Trace messages. Instead, N-Trace solely
provides the address of the exception or interrupt handler."
"As almost every ingress port..." to
"Since almost every ingress port cycle updates I-CNT, there is a possibility
of overflow. For more information, see the I-CNT Details chapter regarding
I-CNT management and overflow handling."
"If the optiona trTeInstEnAllJumps..." to
"If the optional trTeInstEnAllJumps bit is set, the trace ingress port is
required to report itype=5 (Taken branch) for all direct unconditional
jumps, which are normally reported as itype = 0 or itype = 15". See also
comment 8.
First point (unification of 'itype' is still TODO).
Closing this issue as these new names (with linkage) are now done (in some earlier commit).
| gharchive/issue | 2024-03-14T16:13:51 | 2025-04-01T06:45:38.284887 | {
"authors": [
"mipsrobert",
"ved-rivos"
],
"repo": "riscv-non-isa/tg-nexus-trace",
"url": "https://github.com/riscv-non-isa/tg-nexus-trace/issues/45",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2265433781 | Extension names don't follow the official convention
Very minor but we have Zcheripurecap, Zcheripte, Zcherilegacy, Zstid,
From the unprivileged spec:
The first letter following the "Z" conventionally indicates the most closely related alphabetical extension category, IMAFDQCVH. For the "Zam" extension for misaligned atomics, for example, the letter "a" indicates the extension is related to the "A" standard extension.
So I think they should be Zicheripurecap, Zistid, etc.
We know that c is wrong - but we're hoping to get an official letter allocated.
Whatever we call it will be changed again so I don't see a strong reason to change it now.
| gharchive/issue | 2024-04-26T10:01:13 | 2025-04-01T06:45:38.297335 | {
"authors": [
"Timmmm",
"tariqkurd-repo"
],
"repo": "riscv/riscv-cheri",
"url": "https://github.com/riscv/riscv-cheri/issues/220",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1008584832 | pull request for issue #96 m/s/ucliccfg
#96 – different priv modes may want different level/priority control. mcliccfg, scliccfg, ucliccfg instead of single cliccfg. Place at different byte offsets in case implementation wants to overlap m/s/u regions and control priv access in a method not using separate address ranges.
changed references to cliccfg to xcliccfg
Discussed some changes to this in the TG meeting 2021/9/28:
Current design
0x0000 xcliccfg
0x1000 per-interrupt bytes
One aperture per priv mode.
S-mode PMP NAPOT 0x0000-0x1FFF
Memory map for overlapping CLIC apertures, where assumption is bus
transactions carries privilege mode and CLIC interface modifies
behavior based on bus transaction privilege mode.
0x0000 M-mode aperture
0x0000 mcliccfg
... (reserved, future M-mode config registers)
0x0400 HS-mode aperture
0x0400 hscliccfg
...
0x0800 S-mode aperture
0x0800 scliccfg
...
0x0C00 U-mode aperture
0x0C00 ucliccfg
...
0x1000 per-interrupt bytes (all modes)
...
S-mode PMP range 0x0800-0x1FFF (but intended for use with priv mode
on bus transactions)
Using shifted per-mode cfg with separate apertures:
0x0000 mcliccfg
0x1000 per-interrupt bytes (M-mode)
...
0x0400 HS-mode
0x1000 per-interrupt bytes (HS-mode)
...
0x0000-0x07FF reserved
0x0800 scliccfg
0x1000 per-interrupt bytes (S-mode)
...
0x0C00 ucliccfg
0x1000 per-interrupt bytes (U-mode)
Discussion was that allowing mode-specific difference in functionality at same address would be problematic in general, and even in this specific case. For example, an M-mode routine trying to access S-mode register would have to know to use mstatus.mprv feature to change effective mode on access. For more general case in other specifications, we've adopted policy of not having mode-specific functional differences on allowed accesses to the same address.
closing this. created pull #183 instead.
| gharchive/pull-request | 2021-09-27T20:41:26 | 2025-04-01T06:45:38.305258 | {
"authors": [
"dansmathers",
"kasanovic"
],
"repo": "riscv/riscv-fast-interrupt",
"url": "https://github.com/riscv/riscv-fast-interrupt/pull/172",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
388046860 | Hypervisor guest time delta
It has been requested that we add htimedelta[h] CSRs so that hosts can lie to guests about the current time, without requiring trapping and emulating. There is no intent to add equivalent CSRs for cycle/instret.
https://groups.google.com/a/groups.riscv.org/forum/#!searchin/isa-dev/Time$20offset|sort:date/isa-dev/9F_-XcUd2kU/bhxHgb62BQAJ
cc @jhauser-us
Please do include the same CSRs for cycle, since the SBI set timer callback has absolute cycles as the argument. I agree that instret and performance counters can be handled via trap and emulate.
sbi_set_timer is specified in terms of the time counter, not the cycle counter. We're covered, as far as timer interrupts go.
| gharchive/issue | 2018-12-06T03:58:35 | 2025-04-01T06:45:38.307534 | {
"authors": [
"aswaterman",
"bonzini"
],
"repo": "riscv/riscv-isa-manual",
"url": "https://github.com/riscv/riscv-isa-manual/issues/298",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2117560880 | Background ddl tests failed with: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
Describe the bug
It occurs in our recovery tests:
https://buildkite.com/risingwavelabs/main-cron/builds/1778#018d74dd-0b1e-4e29-8578-39dc395dd5ee
Error message/log
2022-07-18T10:40:24.314185Z DEBUG risingwave_simulation::slt: retry count iteration=5
2022-07-18T10:40:25.325206Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:25.325206Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:28.924928Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:28.924928Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:32.702053Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:32.702053Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401009 is in state INITIAL
2022-07-18T10:40:37.381402Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:37.381402Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401009 is in state INITIAL
2022-07-18T10:40:39.165781Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:39.165781Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:40.456610Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:40.456610Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401007 is in state INITIAL
2022-07-18T10:40:42.244699Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:42.244699Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:44.470323Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:44.470323Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401007 is in state INITIAL
2022-07-18T10:40:46.658468Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:46.658468Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:47.006909Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:47.006909Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:47.073633Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:47.073633Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401007 is in state INITIAL
2022-07-18T10:40:49.112516Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:49.112516Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401008 is in state INITIAL
2022-07-18T10:40:52.450178Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:52.450178Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401009 is in state INITIAL
2022-07-18T10:40:55.598938Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:40:55.598938Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401009 is in state INITIAL
2022-07-18T10:40:56.332080Z DEBUG risingwave_simulation::slt: Record Statement { loc: Location { file: "e2e_test/streaming/rate_limit/snapshot_amplification.slt", line: 20, upper: None }, conditions: [], connection: Default, expected_error: None, sql: "CREATE SINK sink AS\n SELECT x.i1 as i1 FROM table x\n JOIN table s1 ON x.i1 = s1.i1\n JOIN table s2 ON x.i1 = s2.i1\n WITH (connector = 'blackhole');", expected_count: None } finished in 17.894836ms
thread '<unnamed>' panicked at /risingwave/src/tests/simulation/src/slt.rs:384:33:
failed to run test after retry 5 times: statement failed: db error: ERROR: Failed to run the query
Caused by these errors (recent errors listed first):
1: gRPC request to meta service failed: Internal error
2: The cluster is recovering
[SQL] CREATE SINK sink AS
SELECT x.i1 as i1 FROM table x
JOIN table s1 ON x.i1 = s1.i1
JOIN table s2 ON x.i1 = s2.i1
WITH (connector = 'blackhole');
at e2e_test/streaming/rate_limit/snapshot_amplification.slt:20
stack backtrace:
0: rust_begin_unwind
1: core::panicking::panic_fmt
2: risingwave_simulation::slt::run_slt_task::{{closure}}
at ./src/tests/simulation/src/slt.rs:384:33
at ./src/tests/simulation/src/main.rs:268:84
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
context: node=14 "client", task=681 (spawned at /risingwave/src/tests/simulation/src/cluster.rs:590:21)
note: run with `MADSIM_TEST_SEED=12` environment variable to reproduce this error
To Reproduce
No response
Expected behavior
No response
How did you deploy RisingWave?
No response
The version of RisingWave
No response
Additional context
No response
Not sure if it's a blocking issue for 1.7 release.
2022-07-18T10:39:53.645088Z DEBUG risingwave_meta::rpc::ddl_controller: starting stream job id=402010 definition="CREATE SINK sink AS SELECT x.i1 AS i1 FROM table AS x JOIN table AS s1 ON x.i1 = s1.i1 JOIN table AS s2 ON x.i1 = s2.i1 WITH (connector = 'blackhole')"
2022-07-18T10:39:53.645088Z DEBUG risingwave_meta::rpc::ddl_controller: preparing stream job id=402010
2022-07-18T10:39:53.645088Z DEBUG risingwave_meta::rpc::ddl_controller: building stream job id=402010
2022-07-18T10:39:53.786265Z INFO risingwave_simulation::cluster: kill compute-3
2022-07-18T10:39:53.904617Z INFO risingwave_simulation::cluster: kill meta-1
2022-07-18T10:39:53.909685Z INFO risingwave_meta::barrier::recovery: recovery start!
2022-07-18T10:39:53.916741Z INFO risingwave_simulation::cluster: restart compute-3
2022-07-18T10:39:53.937698Z DEBUG risingwave_meta::rpc::ddl_controller: creating stream job id=402010
2022-07-18T10:39:53.937698Z DEBUG risingwave_meta::manager::catalog: creating_tables ids: []
2022-07-18T10:39:53.947170Z DEBUG risingwave_meta::barrier::recovery: clean dirty table fragments: {}
2022-07-18T10:39:53.952062Z INFO risingwave_meta::barrier::recovery: recovering mview progress
2022-07-18T10:39:53.961197Z INFO risingwave_meta::barrier::recovery: recovered mview progress
2022-07-18T10:39:53.961197Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:39:53.961197Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401007 is in state INITIAL
2022-07-18T10:39:53.969957Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:39:53.969957Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401007 is in state INITIAL
2022-07-18T10:39:54.149030Z DEBUG recovery_attempt: risingwave_meta::barrier::recovery: start resetting actors distribution
2022-07-18T10:39:54.149030Z WARN recovery_attempt: risingwave_meta::barrier::recovery: scale actors failed error=the materialized view of fragment 401007 is in state INITIAL
2022-07-18T10:39:54.166391Z INFO risingwave_simulation::cluster: kill compute-1
The timeline is as follows: during the process of creating a sink, CN was killed and recovery started. During the clean dirty phase, the table fragment of the sink was retained. Before auto scaling was enabled, the sink still had the possibility of success or injecting and collecting barriers causing another recovery to be cleaned up by the thread that created it. Now that auto scaling has been enabled, it fails during assertion, leading to continuous recovery.
Since it cannot be determined whether the DDL thread writes into table fragments during cleaning, allowing auto scaling to operate together with initializing or creating table fragments should currently be the only solution.Cc @shanicky
| gharchive/issue | 2024-02-05T02:50:09 | 2025-04-01T06:45:38.316161 | {
"authors": [
"kwannoel",
"yezizp2012"
],
"repo": "risingwavelabs/risingwave",
"url": "https://github.com/risingwavelabs/risingwave/issues/14992",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2482920040 | Protocol error: FORMAT DEBEZIUM forbids additional columns, but got ["_rw_kafka_timestamp"]
Describe the bug
CREATE SOURCE IF NOT EXISTS cdc_hr_waf_core_entity
WITH (
connector = 'kafka',
--'debezium-avro-confluent.schema-registry.url' = 'http://kafka-stream-registry-cp-schema-registry.kafka-cluster:8081',
--'format' = 'debezium-avro-confluent',
properties.bootstrap.server = 'kafka-stream-kafka-bootstrap.kafka-cluster:9092',
--'properties.group.id' = 'flink-cdc',
scan.startup.mode = 'earliest',
topic = 'HR.HROUTPUT.WAF_CORE_ENTITY'
)
FORMAT debezium ENCODE AVRO (
schema.registry = 'http://kafka-stream-registry-cp-schema-registry.kafka-cluster:8081'
)
Error message/log
org.jkiss.dbeaver.model.sql.DBSQLException: SQL 错误 [XX000]: ERROR: Failed to execute the statement
Caused by:
Protocol error: FORMAT DEBEZIUM forbids additional columns, but got ["_rw_kafka_timestamp"]
at org.jkiss.dbeaver.model.impl.jdbc.exec.JDBCStatementImpl.executeStatement(JDBCStatementImpl.java:133)
at org.jkiss.dbeaver.ui.editors.sql.execute.SQLQueryJob.executeStatement(SQLQueryJob.java:614)
at org.jkiss.dbeaver.ui.editors.sql.execute.SQLQueryJob.lambda$2(SQLQueryJob.java:505)
at org.jkiss.dbeaver.model.exec.DBExecUtils.tryExecuteRecover(DBExecUtils.java:194)
at org.jkiss.dbeaver.ui.editors.sql.execute.SQLQueryJob.executeSingleQuery(SQLQueryJob.java:524)
at org.jkiss.dbeaver.ui.editors.sql.execute.SQLQueryJob.extractData(SQLQueryJob.java:976)
at org.jkiss.dbeaver.ui.editors.sql.SQLEditor$QueryResultsContainer.readData(SQLEditor.java:4155)
at org.jkiss.dbeaver.ui.controls.resultset.ResultSetJobDataRead.lambda$0(ResultSetJobDataRead.java:123)
at org.jkiss.dbeaver.model.exec.DBExecUtils.tryExecuteRecover(DBExecUtils.java:194)
at org.jkiss.dbeaver.ui.controls.resultset.ResultSetJobDataRead.run(ResultSetJobDataRead.java:121)
at org.jkiss.dbeaver.ui.controls.resultset.ResultSetViewer$ResultSetDataPumpJob.run(ResultSetViewer.java:5152)
at org.jkiss.dbeaver.model.runtime.AbstractJob.run(AbstractJob.java:115)
at org.eclipse.core.internal.jobs.Worker.run(Worker.java:63)
Caused by: org.postgresql.util.PSQLException: ERROR: Failed to execute the statement
Caused by:
Protocol error: FORMAT DEBEZIUM forbids additional columns, but got ["_rw_kafka_timestamp"]
at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2725)
at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2412)
at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:371)
at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:502)
at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:419)
at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:341)
at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:326)
at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:302)
at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:297)
at org.jkiss.dbeaver.model.impl.jdbc.exec.JDBCStatementImpl.execute(JDBCStatementImpl.java:330)
at org.jkiss.dbeaver.model.impl.jdbc.exec.JDBCStatementImpl.executeStatement(JDBCStatementImpl.java:131)
... 12 more
To Reproduce
No response
Expected behavior
No response
How did you deploy RisingWave?
No response
The version of RisingWave
No response
Additional context
No response
use "CREATE SOURCE" just fine.
| gharchive/issue | 2024-08-23T11:18:27 | 2025-04-01T06:45:38.323059 | {
"authors": [
"gaozuo"
],
"repo": "risingwavelabs/risingwave",
"url": "https://github.com/risingwavelabs/risingwave/issues/18222",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1675501739 | refactor(stream-agg): remove Emitter
I hereby agree to the terms of the RisingWave Labs, Inc. Contributor License Agreement.
What's changed and what's your intention?
abstraction not very useful
Checklist For Contributors
[x] I have written necessary rustdoc comments
[x] I have added necessary unit tests and integration tests
[x] I have added fuzzing tests or opened an issue to track them. (Optional, recommended for new SQL features #7934).
[x] I have demonstrated that backward compatibility is not broken by breaking changes and created issues to track deprecated features to be removed in the future. (Please refer to the issue)
[x] All checks passed in ./risedev check (or alias, ./risedev c)
Checklist For Reviewers
[ ] I have requested macro/micro-benchmarks as this PR can affect performance substantially, and the results are shown.
Documentation
[x] My PR DOES NOT contain user-facing changes.
Click here for Documentation
Types of user-facing changes
Please keep the types that apply to your changes, and remove the others.
Installation and deployment
Connector (sources & sinks)
SQL commands, functions, and operators
RisingWave cluster configuration changes
Other (please specify in the release note below)
Release note
RPITIT here :p
https://github.com/risingwavelabs/risingwave/blob/650b42102f29db027d6949eea34a1c5d2bad96d1/src/stream/src/common/cache/state_cache/mod.rs#L51
| gharchive/pull-request | 2023-04-19T19:32:46 | 2025-04-01T06:45:38.329709 | {
"authors": [
"xxchan"
],
"repo": "risingwavelabs/risingwave",
"url": "https://github.com/risingwavelabs/risingwave/pull/9286",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1275086161 | Completed activity
Checklist
Required changes:
[x] Center the description (Line 17 in HTML file) using a single change in CSS
[x] Fix JS so that when the search is executed, the search engine is not present in query
[x] Add Google Drive, Google Slides and Google Sheets as a provider
Optional changes:
[ ] Add other search providers (GitHub, Spotify, etc)
[ ] Additional CSS styling (as you like)
Site link:
https://aphievel.github.io/multisearch/
Good work, was looking for flexbox specific change in requirement but anything that works is good.
| gharchive/pull-request | 2022-06-17T14:16:31 | 2025-04-01T06:45:38.337083 | {
"authors": [
"Aphievel",
"riskycase"
],
"repo": "riskycase/multisearch",
"url": "https://github.com/riskycase/multisearch/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2231409142 | Platform UI; 2 Fast 2 Furious
Roadmap
We've decided to drop Sass entirely from this build to take advantage of the progress/evolution of CSS.
Customization
Colors
Naming conventions
Color Space
More Colors
CSS Custom Properties
Refactor components to use more generic custom properties
Design Tokens
JSON file for configuring custom properties
Spacing
Logical Properties
Fluid Spacing Scale
Refactor spacers to prevent overrides and clashing
Theming
Support Custom Themes
Typography
Fluid Typographical Scale
Font Families
default font family: system-ui
Line height and other spacing
Components
Existing Components
Accordions
Support non-JavaScript approach using <details> and <summary>
Drop the header background color on hover and active
Block Quotes
Need examples in the Launch Pad
Buttons
Naming conventions
Revisit default stylings
Hover styles
Border radius
Cards
Remove extraneous classes: card__header and card__content
Refactor towards pattern used by Steph Eckles in workshop
Carousel
Lives in the Platform 2.0 Branch
Drawers
Remove extraneous classes: drawer__inner and drawer__header
Forms
Spacing within forms
Remove unnecessary modifiers: form--fancy, form--bordered, form--rounded, and form--light - maybe handled via configuration file
Replace Checkboxes & Radios with Shop & Enroll style inputs
Revisit form states: error, success
Required/Optional Field stylings
Toggle styles: Needs tightened up, determine default stylings
Revisit form stylings
Figures
Refactor to be simpler
Image Round
Revisit need, may be able to replace with a circle class
Loaders
Revisit default stylings
Menus
Add more examples to Launch Pad
Site Menu
Refactor to make simpler
Fixed Header
Should not depend on the + adjacent sibling selector
Alternate approach could be adding a class to body
Icon Menu
Review if necessary, might be recreated using utility classes
Messages
Review data-header attribute
Modals
Implement as dialog element
Pills
Might not need Pill Circle as it could use circle as noted in Image Round notes
Tables
Make more customizable, such as background color on table rows
Tabs
Refactor away from radio buttons for tabs due to issues related to z-index and responsiveness
Tooltips
Candidate for popover for non-JavaScript approach
New Components
Disclosure Component
Collapsing Table Component ?
Hero Component
Icon Component
Image Gallery Component
Media Component
Off-Canvas Navigation Component
Methodology
Discuss BEM methodology
Figure out how we want to approach this
CSS Class naming conventions
Eliminate double-dash, double-underscore classnames like: component__header
Utility classes, translate to single dashes as well.
Components should still have top-level class names, but can style based on semantic HTML structure
example: accordion will have .accordion class, but could use semantics
Utilities
Usage of !important
Backgrounds, other places
Remove !import in favor of CSS Layers
Existing Utilities
Borders
Discuss adding rounded (border radius utility) to borders
Cursors
We only have one, should we add more?
Height and Width
More granular height options
Layouts
Block layout - revisit box model
Remove direction up in responsive block utility names
laptop-up-3 => laptop-3
Use gap property to provide gutter to block utilities
Review modifiers: block--shrink, block--fill, and block--full
Max Container
Maybe max-container should be rolled into block-container
possibly just a container class
Screen Readers
Re-name sr-only to visually-hidden
Spacers
Use CSS Logical Properties
What does that mean for naming? padding-top does not necessarily equate to padding-block-start
Maybe: pxs pys pxe pye
Refactor p- and m- to prevent overriding individually set sides
New Utilities
CSS Grid Utilities
gap utilities
Flow utility
"Exception" utilities
:not(.fullbleed)
more research / examples
Platform UI 2.0 Discussion - 4-9-2024
Notes
CSS Class naming conventions
Eliminate double-dash, double-underscore classnames like: component__header
Utility classes, translate to single dashes as well.
Components should still have top-level class names, but can style based on semantic HTML structure
example: accordion will have .accordion class, but could use <details> <summary> semantics
Spacers
Fluid spacing scale that adjusts with browser viewport by default
https://utopia.fyi/
Default spacings can be overridden by config / design tokens
Independent Packages
Utilities
Utilize Sass for loops
Need way to communicate defaults / configuration between packages to import CSS custom properties
Icon Font
Find a path towards SVG Sprites
https://ryantrimble.com/blog/what-the-heck-is-an-svg-sprite-sheet/
Remove !import in favor of CSS Layers
Layouts
Grid
Grid utilities - https://codepen.io/mrtrimble/pen/MWZWaja/342f35e920083a42522e0dd859ee0ee1
Page Layouts - Pre-configured layouts to scaffold pages
Improving blocks to use gap instead of padding/margin
Typography
Fluid Typographical Scale
https://utopia.fyi/
Default font family: system-ui
Solid defaults:
Line height and spacing
Headings, paragraphs, etc...
Reimagine Components
Buttons
Loaders
Messages
Tabs
Infobox - maybe progressively enhance with popover / JavaScript
Menus
Icon Menu - offload to Platform, remove from Platform UI
Questions:
Monorepo
What problems does this present?
Infrastructure / Github Actions need updated to support
What problems does this solve?
Better Developer Experience
Easier to work between multiple projects at once
Packages will stay up-to-date
closing. not moving forward at this time.
| gharchive/issue | 2024-04-08T14:42:36 | 2025-04-01T06:45:38.384012 | {
"authors": [
"hougasian",
"mackenziegates",
"mrtrimble"
],
"repo": "ritterim/platform-ui",
"url": "https://github.com/ritterim/platform-ui/issues/838",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2324280858 | live server is not working
live server is not working...i can't open chrom, it just notifies me that the server is open, but it's not starting.
does anyone else have this problem?
Go to “Edit system environment variables”
Click on “Environment Variables”
In the System tab, look for “PATH” or “Path”
Click on “Edit”
Add C:\Windows\System32
Click Ok and close everything.
If it doesn't work:
Uninstall the extension
Restart your computer.
Reinstall the extension
Go to the live server extension's settings
Click in "Edit in settings.json"
Remove the line “liveServer.settings.NoBrowser”: true,"
And manually add your preferred browser. Example:
“liveServer.settings.CustomBrowser”:“chrome:PrivateMode”
IF this help you, please follow me! 🤘
| gharchive/issue | 2024-05-29T21:30:50 | 2025-04-01T06:45:38.390721 | {
"authors": [
"isabelacarone",
"pedro-pelicioni"
],
"repo": "ritwickdey/vscode-live-server",
"url": "https://github.com/ritwickdey/vscode-live-server/issues/2946",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1384055936 | 🛑 Fontsss is down
In 9763bf2, Fontsss (https://fontsss.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Fontsss is back up in da9c1a3.
| gharchive/issue | 2022-09-23T16:45:56 | 2025-04-01T06:45:38.414827 | {
"authors": [
"rizalrenaldi"
],
"repo": "rizalrenaldi/ruptime",
"url": "https://github.com/rizalrenaldi/ruptime/issues/55",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2098997279 | WIP: handle large chunks that get split by http lib
Closes #44 by looking for HTTP chunk delimiters in the HTTP stream coming back from Ollama.
WIP because it needs a bit more testing.
Here's an example of a stream captured with tcpdump:
HTTP/1.1 200 OK
Content-Type: application/x-ndjson
Date: Wed, 24 Jan 2024 17:14:00 GMT
Connection: close
Transfer-Encoding: chunked
69
{"model":"codellama:13b-code","created_at":"2024-01-24T17:14:00.708054246Z","response":"(","done":false}
6e
{"model":"codellama:13b-code","created_at":"2024-01-24T17:14:00.719238491Z","response":"module","done":false}
69
{"model":"codellama:13b-code","created_at":"2024-01-24T17:14:00.730212903Z","response":".","done":false}
The server is sending two newlines at the end of each chunk. This PR updates streamResponse to concatenate all the data it receives from the HTTP stream until it gets a piece of data ending with \n, then it send all the data (the whole HTTP chunk) to the caller.
Tested with HTTP chunks that did get split coming back from the server as well as normal chunks.
@rjmacarthy
Thanks for the PR see comment https://github.com/rjmacarthy/twinny/issues/44#issuecomment-1908784692 which includes a fix similar already to this. I did try to notify you before you submitted, thanks.
Thanks for the PR see comment #44 (comment) which includes a fix similar already to this. I did try to notify you before you submitted, thanks.
Yeah, np, I just got into it and didn't see your message until after.
Reopening as may be better solution
| gharchive/pull-request | 2024-01-24T20:05:24 | 2025-04-01T06:45:38.533308 | {
"authors": [
"rcgtnick",
"rjmacarthy"
],
"repo": "rjmacarthy/twinny",
"url": "https://github.com/rjmacarthy/twinny/pull/46",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1344861671 | Bogus diagnostics in macro modules?
The Fennel manual describes using macro modules to export commonly re-used macros. In macro modules, macros are defined not using (macro) but rather as plain (fn) functions. This leads to hotpot mistakenly treating macro code as "run-time" (rather than "compile-time") code, reporting errors such as
"unknown-identifier: assert-compile"
"symbols may only be used at compile time"
etc.
Is there some way to configure hotpot so that it knows how to distinguish between regular (run-time) modules vs. macro (compile-time) modules?
There is https://github.com/rktjmp/hotpot.nvim/blob/b942e8760ea26f6ff3782f675a8d6c1323f3e7d4/fnl/hotpot/api/diagnostics.fnl#L118-L126 (under :h hotpot.api.diagnostics.set-options but the docs are lacking, PR's welcome).
You should be able to call that with macro compiler options and it will call the handler internally. I haven't tested this rigorously, but that's the idea from memory.
I would either create a bind, or auto-cmd that calls that for files you consider "macro files".
See also :h hotpot-setup for the default compiler options to pass for macros.
The options will override what's set here:
https://github.com/rktjmp/hotpot.nvim/blob/b942e8760ea26f6ff3782f675a8d6c1323f3e7d4/fnl/hotpot/api/diagnostics.fnl#L71-L72
So you may want to pass in a filename and allowed globals in addition to any env options.
I don't think there is a really safe way to detect "macro modules", besides perhaps if they are called init-macros.fnl which is why we don't set anything like this automatically currently.
Thanks! This set-options thing seems useful for sidestepping the "unknown identifier: assert-compile" error, but it seems to have no effect on the "symbols may only be used at compile time" issue (presumably, since that's not an unresolved global variable issue).
For reference, here's what I ran (as a vim command, via lua):
lua require 'hotpot.api.diagnostics'['set-options']({ env = '_COMPILER' })
Which, per the Fennel API docs, "will cause the code to be run/compiled in a context which has all compiler-scoped values available. This can be useful for macro modules or compiler plugins."
Upon running that, the "unknown identifier: assert-compile" error did indeed go away, but "lists/symbols may only be used at compile time" errors remained.
Perhaps this is a Fennel-side issue that I should take up with them?
Hmm, it's because macros are kind of always eval-at-compile-time but not technically compiled.
When you compile a regular module that uses a macro, the macro module is evaluated during compile and the functions are basically added to the Fennel runtime, there is never any resulting lua code for the macro module.
To get the diagnostics for normal modules we give the code to the compiler and show any errors, but for macros that wont work.
Swapping compile-buffer to eval-buffer (with the env alteration) will let it work but:
it can only detect errors in unquoted code because the quoted code just compiles out to a list, it's never evaluated, so (bad-let ,(bad call) will show an error for (bad call) but not (bad-let.
there's the potential for harmful side effects as eval'ing the code will actually run it, which might accidentally run code that you don't want to run!
you could get around this probably by passing in an env that has compiler symbols and dead stubs for stuff like fopen etc.
AFAIK Fennel upstream doesn't provide any way to "check a macro compiles" as we are seeing. Unsure how hard that would be to implement. I know that Andrey Listopadov wrote some "eval-safely" stuff for his fennel-doc project.
I can add an additional function, perhaps (set-mode buf :compile|:eval) which would solve the issue but it wont ever be turned on by default due to the potentially unsafe nature.
I've added an extra argument to set-options, (set-options buf opts :compile|:eval|nil) which effects how code is run, as above, be very careful you are not writing potentially damaging code and evaluating it unexpectedly.
Probably there is room in the API for diagnostics.check-once where you can explicitly force a diagnostic run on a buffer when you know its safe, if you didn't want it checked after each insert leave.
Changes: https://github.com/rktjmp/hotpot.nvim/commit/0c192b64addebfdc97d92803ce1e73d8600a2b03
Should be fixed, requires macros to be in a file that ends in macro.fnl or macros.fnl, doc: https://github.com/rktjmp/hotpot.nvim/blob/e77fc345e3f7b8bdcc18f7da22edb4cde01613f4/doc/hotpot-api.txt#L79-L80
| gharchive/issue | 2022-08-19T20:18:06 | 2025-04-01T06:45:38.561249 | {
"authors": [
"kwshi",
"rktjmp"
],
"repo": "rktjmp/hotpot.nvim",
"url": "https://github.com/rktjmp/hotpot.nvim/issues/73",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2471800186 | [BUG] Restore fail using along with luukvbaal/statuscol.nvim
Describe the bug
Restore fails with last empty buffer (no files opened) while using along with https://github.com/luukvbaal/statuscol.nvim.
To Reproduce
Steps to reproduce the behavior:
Open nvim
Open files
Close all files. NvChad will create the empty buffer.
Quit nvim
Open nvim
Error appear
Expected behavior
Nvim should start normally without error.
Screenshots
Baseline (please complete the following information):
Result of set sessionoptions?: sessionoptions=blank,buffers,curdir,folds,help,tabpages,winsize,winpos,terminal,localoptions
OS. e.g uname -a: Linux kevostro-arch 6.10.2-arch1-2 #1 SMP PREEMPT_DYNAMIC Sat, 03 Aug 2024 17:56:17 +0000 x86_64 GNU/Linux
Neovim version nvim --version: v0.10.1 LuaJIT 2.1.1720049189
URL to your current config (if public): https://github.com/KevinNitroG/Alexis12119-nvim
Can you share the contents of the session file it's loading (should be in ~/.local/share/nvim/sessions/)?
I suspect the issue is that the luukvbaal/statuscol.nvim isn't getting loaded because the FilePost user event in nvchad doesn't fire when the buftype is nofile:
https://github.com/NvChad/NvChad/blob/fea5d02d2d07b148243c058d1b8b775f9cccac5e/lua/nvchad/autocmds.lua#L14-L15
And I bet the session file has something is expecting statuscol to be loaded.
I spent more time looking into this. I think this is an issue with statuscol putting things in the local options that aren't lazy loading resilient, as described here in the link above. I submitted a PR but if they don't merge, you can fix the error by requiring statuscol before restoring a session by adding this to your auto-session config:
pre_save_cmds = {
require('statuscol'),
},
@tan-wei workaround for statuscol issue is above (assuming they don't merge my PR)
Thank you guys!
statuscol was updated to make it resilient to lazy loading.
There is small detail, tho, people with saved sessions that have the original setlocal statuscolumn=%!v:lua.StatusCol() will get an error loading those sessions (because StatusCol no longer exists) until they save a new session with the new value. Not a huge deal but worth documenting here in case ppl search for this issue in the future.
If you get a statuscol error after loading your session, just run :SessionSave to save a session with the new value (must have at least one buffer open that uses statuscol) to write out session with the updated statuscolumn value.
| gharchive/issue | 2024-08-18T05:46:45 | 2025-04-01T06:45:38.591447 | {
"authors": [
"KevinNitroG",
"cameronr"
],
"repo": "rmagatti/auto-session",
"url": "https://github.com/rmagatti/auto-session/issues/353",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2673189681 | 🛑 Statuslog is down
In 2555522, Statuslog (https://r.mateu.me) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Statuslog is back up in 2347067 after 4 minutes.
| gharchive/issue | 2024-11-19T18:49:37 | 2025-04-01T06:45:38.594717 | {
"authors": [
"rmateu"
],
"repo": "rmateu/statuspage",
"url": "https://github.com/rmateu/statuspage/issues/602",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2609072838 | Retrieve data by Query
Example: Query my SQL Server DB to retrieve the part description and use it in the Title property.
I would add a settings panel where insert the Server connection string.
TBD Somewhere store the common queries. Structure them to use input parameters same as propertyTask
A query could retrieve more than a value then decide how to store and use them.
You're outta my league in the SQL department, but I can help with the save/retrieve part.
If you have time and add a panel with two texboxes as a start, one for the connection string and one for the query.
Provide an example on how to use the text of those two text boxes would be enough as a start.
Can do.
Also, I need to know how to retrieve their values to be used in the code, have you named some properties?
The query will contain input parameter in the usual format %{System.Title}
Added public properties ServerConnectionString and ServerQuery. They should be saved and restored correctly now.
Not sure how the returned values from the server would be stored/identified/used. I was hoping that would be your problem. :smiley:
Best would be to let users use a formula like %{server1.query1.result1} that will produce a result value
Ideally we should be able to store more than one query, and perhaps even more than one server
I will work on the query side next week
I have done a very simple implementation of the query to start coding; it's triggered by a console.writeline after a property edit on the ListViewFiles here: https://github.com/rmcanany/SolidEdgeHousekeeper/commit/4a05535da2623f01b523be9c6ae2ffa7f1a254ed#diff-a640283a9313a42d0466bb813b6d68afb8fefb020c89d3c471ae11533b5fe122R3069
The query that contains a property retrieve text needs to be implemented, I have just inserted a replacement text for testing purposes here: https://github.com/rmcanany/SolidEdgeHousekeeper/commit/4a05535da2623f01b523be9c6ae2ffa7f1a254ed#diff-a640283a9313a42d0466bb813b6d68afb8fefb020c89d3c471ae11533b5fe122R3256
Last but not least, I have added an example access db in the project folder, for some reason it creates a My Project folder when in the binary; perhaps you know a way to provide this example file in a better folder, fell free to move it as you think is better to distribute it.
To open the MDB you need the MS OLE Components, its something to install, I know it's not ideal, but let me know what do you think. I will add a SQL connection example, maybe with an online example DB
Sounds good. Honestly, it's a bit over my head. Let me know if you need anything from me.
OK, I'll take a look at that.
I ran it with the example you provided. It didn't complain and it remembered a property edit I made by double-clicking a cell in the ListView.
I noticed it added the following lines to Housekeeper.vbproj. It looks like a machine-specific location, but it calls it a <HintPath>, so maybe it's OK.
<Reference Include="System">
<HintPath>..\..\..\..\Program Files (x86)\Reference Assemblies\Microsoft\Framework\.NETFramework\v4.7.2\System.dll</HintPath>
</Reference>
I have implemented the Query in SubstitutePropertyFormula; the syntax to use it is %{System.Query1} and it returns the first value of the query execution. The query itself may contain properties that are similarly correctly converted.
Here is an example where I retrieved the Subject value from the Query and wrote in the property
Do you think an online testing SQL database would be useful or the access DB is enough?
I found this free service, never used it.
To me the Access DB is enough, but I'll let you decide.
What do you think about adding a new PropertySetName Server? Wouldn't that make it easier to parse the property formula?
To me the Access DB is enough, but I'll let you decide.
What do you think about adding a new PropertySetName Server? Wouldn't that make it easier to parse the property formula?
Yes it could be useful to differentiate what is a query and what not.
I'm working on turning TemplatePropertiesDict into a class. It's a lot cleaner for sure with better control. I've got Enums and everything! I'll add Server to PropertySetNameConstants.
If the query were a Task, wouldn't that mess up the property filter and listview properties? Those happen before processing, right?
Let's leave the query thing separate from tasks by now, I will think about it another time
Syntax improved with #171
I have implemented %{Server.Query|Rx}
I like that. If you don't get to it, I'll add Server to PropertySetNameConstants when I get back.
| gharchive/issue | 2024-10-23T15:44:43 | 2025-04-01T06:45:38.607272 | {
"authors": [
"farfilli",
"rmcanany"
],
"repo": "rmcanany/SolidEdgeHousekeeper",
"url": "https://github.com/rmcanany/SolidEdgeHousekeeper/issues/161",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
105583615 | Минификатор срабатывает каждый раз при аякс запросе
Минификатор срабатывает каждый раз при аякс запросе. Получается что каждый запрос генерируется js файл в папке minify.
в контроллере используется $this->renderPartial
бандл подключается только в шапке
папка web/assets очищена
upd. если используется POS_HEAD, то проблемы не возникает
Accepted.
Как временное решенеие изменил 122 строчку в файле View.php:
self::PH_BODY_END => ($ajaxMode) ? "" : $this->renderBodyEndHtml($ajaxMode)
Внутри render partial подключаются какие-то assets?
Нет, только в хедере. Но после того как запрос ушел генерируется JS с другим (меньшим) размером.
На каждый запрос генерируется? Или только один раз?
Особенность этого расширения в том, что оно минифицирует любой набод assets. То есть если в каком-то разделе помимо условного AppAssetBundle подключаются какие-то ещё assets, то он минифицирует это в новый пак.
Например, на всех страницах подключается только AppAssetBundle, расширения просто минифицирует его.
Но в одном из разделов во вьюхе подключается доп скрипт, например
$this->registerJsFile('/js/user/edit.js');
В этом случае расширения создаст новый минифицированный пак AppAssetBundle + /js/user/edit.js
В Вашем случае возможно в renderPartial подключается какой-то виджет, который тянет свой AssetBundle, и расширение именно его пытается минифицировать. Теоретически, он должен минифицироваться только один раз, пока нет кэша в директории minify
Только один раз. Но получается два инициализированных бутстрапа, которые не дают кликнуть по чекбоксу (один включает по клику, второй сразу выключает)
В yii2 кажется есть механизм, который предотвращает повторное добавление скрипта в шапку, но не в конец страницы
Рекомендую просто внутри афкс вызова не подключать assets. Это плохая практика. Скорее всего Вы используете pjax?
Да, используется pAjax. Внутри подключаются бутстрап-карусели, которые и тянут за собой бутсрап.
Кстати после запроса JS скрипт добавляется в header, вместо body.
Значит всё верно. Вам требуется не подключать assets внутри ajax действия. Тогда pjax не будет подключать assets повторно. Вместе подключения внутри ajax, Вам нужно подключить бандл виджета заранее. А внутри ajax сделать только инициализацию скрипта виджета.
У меня похожая проблема, есть большой список скриптов в AppAsset, при ajax запросе, если даже включить следующие бандлы yii\web\JqueryAsset, yii\validators\ValidationAsset, yii\widgets\ActiveFormAsset в AppAsset, то при использование в ajax запросе ActiveForm, они всё равно будут добавлены, minify в этом случае добавляет просто весь (полный) набор скриптов (в одном минифицированном файле), я решил эту проблему расширением rmrevin\yii\minify\View скопировал один метод renderBodyEndHtml() и добавил условие (первый строки метода)
if (!$ajaxMode && !empty($this->jsFiles[self::POS_END])) { $lines[] = implode("\n", $this->jsFiles[self::POS_END]); }
То, что предлагает RsD0p9BK, подойдет не всем, т.к. в этом случае будут удалены в том числе обычные скрипты, а не только дублирующие бандлы.
Ошибка работы получается в том, что при ajax запросе нужно формировать новый файл для нового небольшого набора подключенных скриптом, а не добавлять в ответ полный уже созданный файл.
@whale001 Добрый
Сама проблема в общем в том, что не все библиотеки "терпят" повторное подключение.
@whale001 это Вы сегодня письмо написали, ещё не успел протестировать с семантиком. Скорее всего нужно использовать уже минифицированную версию. Либо текеущий минификатор ломает скрипт (такое бывает, когда разработчики плагинов не ставят ;, например).
Да, я писал, потом уже нашел эту тему. Дело именно в повторном подключении через ajax, а не в том, что скрипты не работают в минифицированном виде. В письме, я , пожалуй, не правильно описал проблему.
Могу собрать простой пример для воспроизведение этой особенности. Сделать?
@whale001 я думаю просто не нужно в аякс обработчике подключать бандл. Зачем это делать?
Он подключается автоматически, это заложено в поедении Yii2
Ошибка в том, что минификатор в ответ на это подключает полный файл, а не новый небольшой
@whale001
В yii не заложено автоматическое подключение бандла. Вы подключаете его где-то самостоятельно. Как раз в ajax запросах он не должен подключаться.
В yii заложена возможность понимать скрипты, которые приходят вместе с ответом сервера. А какие Вы там подключаете скрипты, это уже дело Ваших рук.
Минификатор минифицирует всё что попадает в view->jsFiles view->cssFiles.
vendor\yiisoft\yii2\widgets\ActiveForm.php:200
Это подключение бандла виджета. Он не подключает Ваш основной бандл.
Иными словами Вы предлагаете не использовать или переопределять поведение ActiveForm
Основной подключается один раз в layout
Ещё раз сформулируйте, в чём проблема? В аякс ответе где нет лейаута минификатор подключает основной бандл? Покажите тест.
Сейчас сделаю
Оказалось, это связано с другой особенностью - повторное добавление библиотеки jquery, при использовании minify и добавлении всех файлов одним - ошибка есть, без использования minify , когда из ajax запроса файлы добавляются по одному - ошибки нет. Буду разбираться.
Разобрался с minify не срабатывает
Заменить строку vendor\rmrevin\yii2-minify-view\View.php:188
на
if (true === $this->enableMinify && !$ajaxMode) {
создал pull request https://github.com/rmrevin/yii2-minify-view/pull/31
| gharchive/issue | 2015-09-09T12:11:11 | 2025-04-01T06:45:38.657120 | {
"authors": [
"RsD0p9BK",
"rmrevin",
"whale001"
],
"repo": "rmrevin/yii2-minify-view",
"url": "https://github.com/rmrevin/yii2-minify-view/issues/22",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
120595236 | Start CaptureActivity from external app
Can I start CaptureActivity from an external app?
Like zxing barcode:
Intent i = new Intent("com.google.zxing.client.android.SCAN");
startActivityForResult(i, REQ_CODE);
and get the result in data.getStringExtra("SCAN_RESULT")
You can do this, but you'd need to make the modifications to this app's code yourself to add an intent filter and handle the intent.
http://developer.android.com/guide/components/intents-filters.html
| gharchive/issue | 2015-12-05T23:23:20 | 2025-04-01T06:45:38.676282 | {
"authors": [
"poseidonjm",
"rmtheis"
],
"repo": "rmtheis/android-ocr",
"url": "https://github.com/rmtheis/android-ocr/issues/38",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
238398152 | Swipe notifications to dismiss
Use SwipeCellKit
Send notification POST request
Optimistically dismiss
Re-show if request fails
Just as an FYI: the SwipeCellKit behavior (aka the Mail.app behavior) is a public API in UIKit in iOS 11.
@basthomas on UICollectionView?
Looks like the new swipe actions are still table-only
Oh, right. 😩
Swipe view setup, just needs wiring
| gharchive/issue | 2017-06-25T18:46:57 | 2025-04-01T06:45:38.689596 | {
"authors": [
"BasThomas",
"rnystrom"
],
"repo": "rnystrom/Freetime",
"url": "https://github.com/rnystrom/Freetime/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1932344505 | Adds summary metric type
Add summary metric type
$metrics->declare(
'registered_users',
Collector::summary()->withHelp('Total registered users counter.')->withLabels('type')
);
issue https://github.com/roadrunner-php/issues/issues/21
Codecov Report
All modified lines are covered by tests :white_check_mark:
Comparison is base (15e89b2) 73.55% compared to head (dc3e944) 73.98%.
:exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality.
Additional details and impacted files
@@ Coverage Diff @@
## 3.x #11 +/- ##
============================================
+ Coverage 73.55% 73.98% +0.43%
- Complexity 49 50 +1
============================================
Files 6 6
Lines 121 123 +2
============================================
+ Hits 89 91 +2
Misses 32 32
Files
Coverage Δ
src/Collector.php
100.00% <100.00%> (ø)
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
| gharchive/pull-request | 2023-10-09T05:45:11 | 2025-04-01T06:45:38.696315 | {
"authors": [
"butschster",
"codecov-commenter"
],
"repo": "roadrunner-php/metrics",
"url": "https://github.com/roadrunner-php/metrics/pull/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2506123991 | chore: move google/protobuf requirement to dev
Hello.
Trying to do the same as in the repository https://github.com/roadrunner-php/goridge - google/protobuf is removed from require_dev and added to suggest ext-protobuf as well. The purpose is not to install google/protobuf when ext-protobuf already installed.
However, I can not add an exception throw same as here: https://github.com/roadrunner-php/goridge/blob/8fe68b0c871862f9ec852659de1cc7c1a68b783f/src/RPC/Codec/ProtobufCodec.php#L26
All class are autogenerated, trying to find out where to make an user-friendly exception if nothing is installed. Need assistance c:
cc @Kaspiman
Summary by CodeRabbit
New Features
Added support for Protobuf codec in the development environment.
Documentation
Improved clarity and organization of dependency requirements in the project configuration.
@msmakouz could you please take a look?
| gharchive/pull-request | 2024-09-04T19:31:17 | 2025-04-01T06:45:38.700777 | {
"authors": [
"AtCliffUnderline",
"rustatian"
],
"repo": "roadrunner-php/roadrunner-api-dto",
"url": "https://github.com/roadrunner-php/roadrunner-api-dto/pull/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1105119268 | [💡FEATURE REQUEST]: Scheduled jobs
Add separate simple queue (use jsondb for example) with scheduled jobs, which will send job to broker at scheduled time. Like https://github.com/mindblowup/taskq
Done via: https://github.com/spiral-packages/scheduler
| gharchive/issue | 2019-07-31T16:23:40 | 2025-04-01T06:45:38.702407 | {
"authors": [
"myavchik",
"rustatian"
],
"repo": "roadrunner-server/roadrunner",
"url": "https://github.com/roadrunner-server/roadrunner/issues/924",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
140555508 | 1.0.0 Release
No major bugs since 0.12.0 was released, so I'm happy to call this 1.0.0 with the minor fix for #43 in place.
cc @tbabej
Tests seem to be happy about this, let's merge then.
| gharchive/pull-request | 2016-03-14T02:02:17 | 2025-04-01T06:45:38.776414 | {
"authors": [
"robgolding63",
"tbabej"
],
"repo": "robgolding63/tasklib",
"url": "https://github.com/robgolding63/tasklib/pull/44",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
314277251 | Upload Hang with only a couple bytes left
FTP OS: Windows Server 2016
FTP Server: FileCOPA FTP
Computer OS: Windows Server 2016
In multiple cases, we are seeing the Upload() with a byte array hang. Its not consistent and sometimes it works and sometimes it doesn't. I have reproduced this on the server running as a service and debug in Visual Studio on my Win10 workstation.
Was using 19.1.2, upgraded to 19.1.3 and it is behaving the same.
I trace the hang to FtpClient_HighLevel.cs, Line 1169:
while (upStream.Position < upStream.Length) {
}
In my case,
upStream.Position = 27739
upStream.Length = 27740
upStream.Position = 253149
upStream.Length = 253151
and no matter how long I let it sit, Position does not increment. In my test, this usually happened after the exact same code runs 3-4 times w/out any issues. It almost seems like the last write/flush didn't update Position.
Can something be put in the while to handle a time/limit and move forward?
Logs :
# Connect()
Status: Connecting to x.x.x.x:990
Status: FTPS Authentication Successful
Status: Time to activate encryption: 0h 0m 0s. Total Seconds: 0.2805905.
Response: 220 InterVations FileCOPA FTP Server Version 10.01 12th September 2016
Command: USER test
Response: 331 Password required for test
Command: PASS ***
Response: 230 User test logged in.
Command: PBSZ 0
Response: 200 PBSZ=0
Command: PROT P
Response: 200 PROT P OK, data channel secure.
Command: FEAT
Response: 211 End
Response: 211-Features:
Response: MDTM
Response: REST STREAM
Response: SIZE
Response: PBSZ
Response: PROT
Response: SITE PSWD
Response: MODE Z
Response: XMD5 "filename" SP EP
Response: XSHA1 "filename" SP EP
Response: XCRC "filename" SP EP
Response: RMDA pathname
Response: DSIZ pathname
Response: AVBL pathname
Response: THMB PNG|BMP|GIF|JPEG|TIFF max_width max_height pathname
Response: MFMT
Response: MFCT
Response: EPRT
Response: EPSV
Response: MLST Type*;Size*;Modify*;Perm*
Status: Text encoding: System.Text.ASCIIEncoding
Command: SYST
Response: 215 Windows_NT
Status: Auto-detected Windows listing parser
# Upload("webforms.xlsx", NoCheck, False)
# OpenWrite("webforms.xlsx", Binary)
Command: TYPE I
Response: 200 Type set to I
# GetFileSize("webforms.xlsx")
Command: SIZE webforms.xlsx
Response: 213 27741
# OpenPassiveDataStream(PASV, "STOR webforms.xlsx", 0)
Command: PASV
Response: 227 Entering Passive Mode (x,x,x,x,15,161).
Status: Connecting to x.x.x.x:4001
Command: STOR webforms.xlsx
Response: 150 Opening BINARY mode data connection for webforms.xlsx
Status: FTPS Authentication Successful
Status: Time to activate encryption: 0h 0m 0s. Total Seconds: 0.1270272.
2018-04-13 22:18:34 y.y.y.y "test" 0 0 USER 331 "test" " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 PASS 230 "********" " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 PBSZ 200 "0" " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 PROT 200 "P" " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 FEAT 211 - " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 SYST 215 - " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 TYPE 200 "I" " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 SIZE 213 "webforms.xlsx" " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 PASV 227 - " "
2018-04-13 22:18:34 y.y.y.y "test" 0 0 STOR 150 "webforms.xlsx" " "
As a follow up to this issue, I have pinpointed what is actually happening. This scenario occurs when uploading a file using the ExistsMode of NoCheck and the existing file is BIGGER than the file being uploaded.
OpenWrite for checkIfFileExists comment says: "Only set this to false if you are SURE that the file does not exist. If true, it reads the file size and saves it into the stream length."
So if the file does exist, the stream length is set to the existing file size, and the while loop will never complete. Would a good solution be in UploadFileInternal changing upstream.Length to fileData.Length (see BUG FIX below)?
// loop till entire file uploaded
long len = fileData.Length;
byte[] buffer = new byte[TransferChunkSize];
// BUG FIX?
if (len < upStream.Length)
upStream.SetLength(len);
if (UploadRateLimit == 0) {
while (offset < len) {
Good work debugging it. Maybe the "bug fix" can only be run in the case of ExistsMode.NoCheck so it does not affect the other cases.
Please reply if you still have this issue.
I am having this exact issue on version 24.0.0 - did this fix get reverted somehow?
We have a lot of contributors and to keep this project fluid and alive I
don't restrict people from changing stuff. Unfortunately our test cases
don't cover everything so a regression is definitely possible. Could you
browse the code and see if you can spot if the bug fix was reverted? Thanks.
On Thu, Jun 13, 2019, 11:09 PM CW Andrews notifications@github.com wrote:
I am having this exact issue on version 24.0.0 - did this fix get reverted
somehow?
—
You are receiving this because you modified the open/close state.
Reply to this email directly, view it on GitHub
https://github.com/robinrodricks/FluentFTP/issues/288?email_source=notifications&email_token=ABT3UKQFISGUHRPU2T7WRVLP2KA4JA5CNFSM4E2TJUY2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXUPK3Y#issuecomment-501806447,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABT3UKRFXZNBQTCIMPLFYGTP2KA4JANCNFSM4E2TJUYQ
.
We have a lot of contributors and to keep this project fluid and alive I don't restrict people from changing stuff. Unfortunately our test cases don't cover everything so a regression is definitely possible. Could you browse the code and see if you can spot if the bug fix was reverted? Thanks.
…
On Thu, Jun 13, 2019, 11:09 PM CW Andrews @.***> wrote: I am having this exact issue on version 24.0.0 - did this fix get reverted somehow? — You are receiving this because you modified the open/close state. Reply to this email directly, view it on GitHub <#288?email_source=notifications&email_token=ABT3UKQFISGUHRPU2T7WRVLP2KA4JA5CNFSM4E2TJUY2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXUPK3Y#issuecomment-501806447>, or mute the thread https://github.com/notifications/unsubscribe-auth/ABT3UKRFXZNBQTCIMPLFYGTP2KA4JANCNFSM4E2TJUYQ .
Yea, I already looked and didn't see it but am not super familiar with the code base. However, I am going to open a pull request to get the fix back in place.
I didn't see any information on contribution guidelines - do people usually clone the repo or just open a new branch (i.e. #411)?
@robinrodricks, You may have already been notified about it but I opened pull request #412 to add the bug fix back in.
Awesome thanks. I'll check it out as soon as I have a moment.
On Fri, Jun 14, 2019, 9:07 PM CW Andrews notifications@github.com wrote:
@robinrodricks https://github.com/robinrodricks, You may have already
been notified about it but I opened pull request #412
https://github.com/robinrodricks/FluentFTP/pull/412 to add the bug fix
back in.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/robinrodricks/FluentFTP/issues/288?email_source=notifications&email_token=ABT3UKU747JUP5V3EL32KLDP2O3M3A5CNFSM4E2TJUY2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXXFALA#issuecomment-502157356,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABT3UKXPDQOLUTO2IWYCIV3P2O3M3ANCNFSM4E2TJUYQ
.
| gharchive/issue | 2018-04-13T23:31:23 | 2025-04-01T06:45:38.798328 | {
"authors": [
"cw-andrews",
"robinrodricks",
"spokanedj"
],
"repo": "robinrodricks/FluentFTP",
"url": "https://github.com/robinrodricks/FluentFTP/issues/288",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
634549709 | Correct toTimestamp
Issue https://github.com/robinvdvleuten/php-ulid/issues/23
Thanks for your PR! Could you make sure that the toTimestamp() method returns an integer instead of a float? Also keep the test ULIDs the same.
Sorry, maybe I don`t understand what need todo.
https://github.com/robinvdvleuten/php-ulid/pull/24/commits/69ede1f810f16e25e96e5cff556133c2d4db7023
@DenisGavrilov no worries. What I tried to say is that there are already two testcases for the toTimestamp() method;
public function testConvertsToTimestamp(): void
{
$this->assertEquals(1561622862, Ulid::fromString('0001EH8YAEP8CXP4AMWCHHDBHJ')->toTimestamp());
$this->assertEquals(1561622862, Ulid::fromString('0001eh8yaep8cxp4amwchhdbhj', true)->toTimestamp());
}
So as I understand your issue and PR correctly, you're saying that 1561622862 is not a correct timestamp and should be divided by a thousand.
But when I convert the integer 1561622862 to a date instance, I'll get 06-27-2019 @ 8:07am (UTC) which looks valid to me.
Could you comment on your PR why you want to divide it and have a float instead of an integer?
I devide use Ulid in my project because "All included" :) Unique Id, sortable and CreatedTimestamp. But, when I testing, get incorrect date.
Look result this code
$ulid = Ulid::generate(); var_dump($ulid->toTimestamp()); $dt = new \DateTime; var_dump($dt->setTimestamp($ulid->toTimestamp())); var_dump($dt->setTimestamp($ulid->toTimestamp()/1000));
OK, so I checked it myself and get the following results;
<?php
$ulid = Ulid::fromString('0001EH8YAEP8CXP4AMWCHHDBHJ');
$dt = new \DateTime();
var_dump($dt->setTimestamp($ulid->toTimestamp())->format('c'));
var_dump($dt->setTimestamp($ulid->toTimestamp() / 1000)->format('c'));
Which gives me the following output;
ulid.php:6: string(25) "2019-06-27T08:07:42+00:00"
ulid.php:7: string(25) "1970-01-19T01:47:02+00:00"
I still come to the conclusion that the current timestamp is definitely correct. And this PR can be closed, thanks for the effort.
| gharchive/pull-request | 2020-06-08T12:05:02 | 2025-04-01T06:45:38.804253 | {
"authors": [
"DenisGavrilov",
"robinvdvleuten"
],
"repo": "robinvdvleuten/php-ulid",
"url": "https://github.com/robinvdvleuten/php-ulid/pull/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
50928943 | How to Register Custom Adapter
I see in the docs the code snippet for registering custom adapters.
http://docs.phinx.org/en/latest/configuration.html#supported-adapters
Question is, where in the world do I put that code?
I have this fully working now (without docs), but it currently depends on #428. If you want a preview of the changes, compare it on my fork.
I have this fully working now (without docs), but it currently depends on #428. If you want a preview of the changes, compare it on my fork.
#439.
| gharchive/issue | 2014-12-04T05:03:35 | 2025-04-01T06:45:38.826618 | {
"authors": [
"shadowhand",
"twilsonxpert"
],
"repo": "robmorgan/phinx",
"url": "https://github.com/robmorgan/phinx/issues/388",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
749908577 | Added possibility to reset the cameras at startup.
With this option, the module tries to reset the cameras at startup. It can be disabled from configuration files.
Thank you @S-Dafarra for the PR. I had some comments and requests which I have mentioned them with the review.
Thank you @S-Dafarra for this PR, confirmed and merged.
| gharchive/pull-request | 2020-11-24T17:35:14 | 2025-04-01T06:45:38.910576 | {
"authors": [
"S-Dafarra",
"kouroshD"
],
"repo": "robotology/walking-teleoperation",
"url": "https://github.com/robotology/walking-teleoperation/pull/45",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1581043283 | Driver Assist
This needs to be tested on the bots. Also needs the substation locations as alignment points. Alignment points will also need to be tuned.
Known issue: paths are being generated with heading, not rotation at the moment. As seen in https://github.com/mjansen4857/pathplanner/wiki/PathPlannerLib:-Java-Usage, the second parameter of PathPoint is heading, not rotation.
FYI for those working on this, this still needs to be able to mirror the path correctly when the blue alliance is used.
Oh, so it turns out PathPlanner doesn't mirror for us?
It does, but then it mirrors your starting position too
Drafted until vision is in a good state
| gharchive/pull-request | 2023-02-11T22:53:22 | 2025-04-01T06:45:38.915846 | {
"authors": [
"KangarooKoala",
"jbko6"
],
"repo": "robototes/ChargedUp2023",
"url": "https://github.com/robototes/ChargedUp2023/pull/39",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
52132601 | Recursive Grammars
So I was messing around with Madness and tried implementing a very rudimentary regex (as in "real" regex) parser using Madness:
<expression> ::= <union> | <term>
<union> ::= <term> '|' <expression>
<term> ::= { <factor> }
<factor> ::= <base> { '*' }
<group> ::= '(' <expression> ')'
<base> ::= <char> | '\' <char> | <group>
A crude translation to Madness would look like this I guess:
let expression = union | term
let union = term ++ ignore("|") ++ expression
let term = factor *
let factor = base ++ ignore("*") *
let group = ignore("(") ++ expression ++ ignore(")")
let base = any | ignore("\\") ++ any | group
The obvious first question that comes to mind (apart from hunting down left-recursions, etc) is:
How does one define recursive grammars in Madness like the one seen above?
Obviously the compiler complained about unresolved identifiers. Just as I'd expect it to do in such a situation.
Would you mind to elaborate on this very briefly (or even add an example to the playground)?
Going to close this out now.
| gharchive/issue | 2014-12-16T16:16:14 | 2025-04-01T06:45:38.920006 | {
"authors": [
"regexident",
"robrix"
],
"repo": "robrix/Madness",
"url": "https://github.com/robrix/Madness/issues/40",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
672274796 | __construct(string $projectUrl, ?string $separator = ":")
Boa tarde robson, joia?
Está correto aqui o method?
public function __construct(string $projectUrl, ?string $separator = ":")
Ao subir em determinador acusa erro no ?string
Está correto sim. Qual versão do seu PHP?
| gharchive/issue | 2020-08-03T18:52:54 | 2025-04-01T06:45:38.925892 | {
"authors": [
"eurairapelli",
"robsonvleite"
],
"repo": "robsonvleite/router",
"url": "https://github.com/robsonvleite/router/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
139914215 | React Generator
Think it would be a good idea to get React up in here, up in here.
@raksonibs Hey you don't have an issue queue on your repo generator-jekyllrb-react -- but I ran into an error and seems likely its related to this pull request.
Background
I'm fairly new to yeoman, so this could totally be PEBCAK or RTFM.
I tried to research a bit, found this #154 -- which seems similar. However some manual fixes equivalent to this fix #155 solves. So I'm at a loss, though I do recall a lodash depreciation when I installed something related to yeoman last night, I think: generator-jekyllrb
My Setup
All that said, I'm using ubuntu 15.10 with global installs set to my local user path per NPM Docs. Except for some outdated generators, my setup seems to be functioning ok ... but it did take me awhile to figure out to install various dependencies ... which might be related here ... or might be a simple lodash update that I'm unsure how to do.
Error
Note: I modified output path to include $HOME
events.js:154
throw er; // Unhandled 'error' event
^
ReferenceError: yeoman is not defined
at eval (eval at template ($HOME/.node/npm-global/lib/node_modules/generator-jekyllrb-react/node_modules/yeoman-generator/node_modules/lodash/dist/lodash.js:6306:22), <anonymous>:8:11)
at underscore ($HOME/.node/npm-global/lib/node_modules/generator-jekyllrb-react/node_modules/yeoman-generator/lib/util/engines.js:23:45)
at Generator.actions.engine ($HOME/.node/npm-global/lib/node_modules/generator-jekyllrb-react/node_modules/yeoman-generator/lib/actions/actions.js:216:32)
at Generator.template ($HOME/.node/npm-global/lib/node_modules/generator-jekyllrb-react/node_modules/yeoman-generator/lib/actions/actions.js:191:19)
at Generator.gruntfile ($HOME/.node/npm-global/lib/node_modules/generator-jekyllrb-react/app/index.js:370:8)
at $HOME/.node/npm-global/lib/node_modules/generator-jekyllrb-react/node_modules/yeoman-generator/lib/base.js:409:16
at processImmediate [as _immediateCallback] (timers.js:516:17)
Hey @mdrmike, I ended up pivoting this branch into another opensource repo: https://github.com/raksonibs/generator-jekyll-react-es6. Let me know if you have problems there and I can sort through them. I think the development will separate quite away from this great start by @robwierzbowski!
| gharchive/pull-request | 2016-03-10T15:04:40 | 2025-04-01T06:45:38.946965 | {
"authors": [
"mdrmike",
"raksonibs"
],
"repo": "robwierzbowski/generator-jekyllrb",
"url": "https://github.com/robwierzbowski/generator-jekyllrb/pull/156",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1352779515 | refactor store code
Fixes https://github.com/roc-lang/rbt/issues/39 to be a little more performant and (hopefully) a lot more understandable. It also moves the responsibility for output path sanitization to Job, which means we will cut off problematic builds much sooner.
Based on #37; review/merge that first!
Been a bit busier than expected, will hopefully be able to review a few of these PRs tomorrow. Just thought I would give an update.
All good. Thank you!
| gharchive/pull-request | 2022-08-26T22:13:30 | 2025-04-01T06:45:38.948944 | {
"authors": [
"BrianHicks",
"bhansconnect"
],
"repo": "roc-lang/rbt",
"url": "https://github.com/roc-lang/rbt/pull/40",
"license": "UPL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2020697476 | Fixed wrapping for long words to avoid horizontal scroll
Here's an example of the issue (in Num#Dec docs):
And the result of applied fix:
I've signed my commit, so it can be merged successfully.
Looks like this got stuck on CI machines. Just kicked it along.
| gharchive/pull-request | 2023-12-01T11:32:55 | 2025-04-01T06:45:38.951374 | {
"authors": [
"andreystepanov",
"lukewilliamboswell"
],
"repo": "roc-lang/roc",
"url": "https://github.com/roc-lang/roc/pull/6131",
"license": "UPL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
316527785 | The lib works only with min android 8 !!
Hi, can you help me pls because the lib works only with min android 8 and I need to work with min android 6
if I use this:
implementation ('com.github.rockaport:alice:0.8.0'){
transitive = false;
}
then the apk build is OK but when I run it to a device with sdk version 24 I get this error at encryption:
ava.security.NoSuchAlgorithmException: PBKDF2WithHmacSHA512 SecretKeyFactory not available
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at javax.crypto.SecretKeyFactory.(SecretKeyFactory.java:249)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at javax.crypto.SecretKeyFactory.getInstance(SecretKeyFactory.java:287)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at com.rockaport.alice.Alice.derivePbkdfKeyBytes(Alice.java:621)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at com.rockaport.alice.Alice.deriveKey(Alice.java:589)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at com.rockaport.alice.Alice.encrypt(Alice.java:178)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at app.r13.com.r13.Home.encrypt(Home.java:95)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at app.r13.com.r13.Home.access$000(Home.java:43)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at app.r13.com.r13.Home$1.onClick(Home.java:148)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at android.view.View.performClick(View.java:5637)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at android.view.View$PerformClick.run(View.java:22429)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at android.os.Handler.handleCallback(Handler.java:751)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at android.os.Handler.dispatchMessage(Handler.java:95)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at android.os.Looper.loop(Looper.java:154)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at android.app.ActivityThread.main(ActivityThread.java:6119)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at java.lang.reflect.Method.invoke(Native Method)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:886)
04-21 21:52:00.360 3725-3725/app.r13.com.r13 W/System.err: at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:776)
Okkk... I fixed by adding to alicecontext
.setPbkdf(AliceContext.Pbkdf.PBKDF_2_WITH_HMAC_SHA_1)
I've removed the okio dependency in v0.9.0
| gharchive/issue | 2018-04-21T21:41:57 | 2025-04-01T06:45:38.968708 | {
"authors": [
"florindumitru",
"rockaport"
],
"repo": "rockaport/alice",
"url": "https://github.com/rockaport/alice/issues/8",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
829749165 | adding support for "defense"
Are you taking any ideas / pull requests?
Going through 7DRL 2021 right now and using this toolkit. I have some ideas around providing some support for a "defense" (eg. armour) rating to work in conjunction with what you have for "weapon".
referenced here - https://github.com/rocket-boots/rote/pull/9
| gharchive/issue | 2021-03-12T04:35:24 | 2025-04-01T06:45:38.987671 | {
"authors": [
"erikyuzwa"
],
"repo": "rocket-boots/rote",
"url": "https://github.com/rocket-boots/rote/issues/8",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
913633122 | Trigger Button
Click on Enter Key Press in search bar
Please assign @rockingrohit9639
Click on Enter Key Press in search bar
Please assign @rockingrohit9639
Ok @SamikshaSankar I will assign you as soon as the program starts.
Hey, @SamikshaSankar you can start working on this issue now. Make sure you complete this issue in 5 days.
Hey @SamikshaSankar are you working on it?
Yes...but search bar is not working..I guess bcz whenever I search something on it..not show anything
Yes...but search bar is not working..I guess bcz whenever I search something on it..not show anything
TypeError: projects is undefined shows like this
Have you pulled all the current code?
@SamikshaSankar hi, is there any problem you are not able to make PR?
Yes... Please assigned next issue...I will be worked later on this issue
@SamikshaSankar are you workong on it?
| gharchive/issue | 2021-06-07T14:49:02 | 2025-04-01T06:45:38.994889 | {
"authors": [
"SamikshaSankar",
"rockingrohit9639"
],
"repo": "rockingrohit9639/project-zone",
"url": "https://github.com/rockingrohit9639/project-zone/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1118347829 | Information about resolving sysroot error
From testing on a few dedicated server providers, I ran into an issue with GRUB not working properly after the upgrade to RockyLinux 8.5
Example error message:
This is fixable by remounting all drives within a rescue system, then rebuilding grub cfg.
We currently have some similar code in the fix_efi function to fix up EFI boot. Should we be doing tis for legacy boot as well?
We currently have some similar code in the fix_efi function to fix up EFI boot. Should we be doing this for legacy boot as well?
Perhaps we should. The servers I was having issues with were booting via legacy BIOS. I'll suggest those modifications perhaps?
Yes, it's unusual for legacy boot to require running grub2-mkconfig, but it shouldn't hurt. Do you want to open a new PR for that?
I would suggest a cleaner way of dealing with this: https://github.com/rocky-linux/rocky-tools/pull/162
Indeed, we do need to run grub2-mkconfig -o /boot/grub2/grub.cfg on a machine, that boots the BIOS way even if it is UEFI capable.
| gharchive/pull-request | 2022-01-29T23:00:15 | 2025-04-01T06:45:39.018612 | {
"authors": [
"komitov",
"markd69",
"pajamian"
],
"repo": "rocky-linux/rocky-tools",
"url": "https://github.com/rocky-linux/rocky-tools/pull/152",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1033746033 | Errors on new M1 macs
Snakemake segfaults on the newer Mac machines. I believe we can circumvent this by installing snakemake-minimal instead of snakemake, although that might break DAG generation.
Working on a patch in the M1 branch: https://github.com/rodluger/showyourwork/blob/M1/Makefile#L39-L44
@dfm, can you confirm that the command
sysctl -n machdep.cpu.brand_string
outputs a string containing M1 on your system?
Confirmed:
> sysctl -n machdep.cpu.brand_string
Apple M1
Thanks!
| gharchive/issue | 2021-10-22T16:15:22 | 2025-04-01T06:45:39.043009 | {
"authors": [
"dfm",
"rodluger"
],
"repo": "rodluger/showyourwork",
"url": "https://github.com/rodluger/showyourwork/issues/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1140812905 | pensar na arquitetura (visão macro) do minishell
visão macro. quais os principais fluxos e pontos de decisão
Refs:
https://cdn.discordapp.com/attachments/889658916970045520/890693779080773672/128610128-fa85c8ae-eff7-4b8a-9b29-90cdfc4aeaa5.png (chat da vila de Rods)
https://whimsical.com/minishell-R7ozXMj4gLMiR1nmeYoxDY
eu acho que talvez seja importante nesse ponto ler um pouco sobre isso aqu?
https://www.cs.purdue.edu/homes/grr/SystemsProgrammingBook/Book/Chapter5-WritingYourOwnShell.pdf
talvez essa seja A LEITURA, idk i'm so sleepy rn send help
| gharchive/issue | 2022-02-17T02:52:08 | 2025-04-01T06:45:39.054694 | {
"authors": [
"rodsmade"
],
"repo": "rodsmade/Minishell-42sp",
"url": "https://github.com/rodsmade/Minishell-42sp/issues/20",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
505632745 | WIP - Beginnings of nested settings class PreLoad
Not finished, but it's a start. At this point a bunch of tests are failing.
Reference: #21
@rofr All tests are passing so I think it's at least at a good point to give some code review. Can you take a look and see if there's anything that sticks out?
I changed some prefix / bindingPath bits to fix a bug and needed to make some changes that effects this PR.
| gharchive/pull-request | 2019-10-11T04:37:19 | 2025-04-01T06:45:39.059104 | {
"authors": [
"myty",
"rofr"
],
"repo": "rofr/Fig",
"url": "https://github.com/rofr/Fig/pull/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
192241193 | When something goes wrong creating a new user, the system never recovers
Problem
Khaled was missing a library for his PIL to work correctly. (Not sure if that's the problem, cause when I reinstalled the app and used new email error still persists)
When a created his user account something went wrong while creating his profile, resulting in the fact that has has no Identity QR code:
payload: {"a":"be-loc","p":10,"r":"20284368-b61b-11e6-a781-35e9c97479a4@jabber.mobicage.com","kid":"17f876e4-e733-4008-8b25-984b06e99d4f","t":["xmpp"]}
ESC[1;33mWARNING 2016-11-29 10:04:06,441 system.py:135] getIdentityQRCode ProfilePointer was NoneESC[0m
- request.email: mohamedk@greenitglobe.com
- request.size: None
Secondly much more things appear to be wrong with his profile, because you just cannot send any messages to him:
ESC[1;31mERROR 2016-11-29 10:12:08,077 add_1_monkey_patches.py:97] send_inbox_forwarders_message chat errorESC[0m
Traceback (most recent call last):
File "/home/khaled/gig/github/rogerthat/appengine/build/solutions/common/bizz/messaging.py", line 756, in _send_inbox_forwarders_message_by_app
chat_message_key = messaging.send_chat_message(message.message_key, body, answers, attachments, sender_member, None, True, tag=tag)
File "/home/khaled/gig/github/rogerthat/appengine/build/rogerthat/rpc/service.py", line 312, in wrapped
result = f(*args, **kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/mcfw/rpc.py", line 149, in typechecked_return
result = f(*args, **kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/mcfw/rpc.py", line 127, in typechecked_f
return f(**kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/rogerthat/service/api/messaging.py", line 122, in send_chat_message
message = bizz_send_chat_message(service_user, parent_key, message, answers, attachments, sender, priority, sticky, tag, alert_flags)
File "/home/khaled/gig/github/rogerthat/appengine/build/mcfw/rpc.py", line 149, in typechecked_return
result = f(*args, **kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/mcfw/rpc.py", line 127, in typechecked_f
return f(**kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/rogerthat/bizz/messaging.py", line 1643, in send_chat_message
priority=priority, default_alert_flags=alert_flags)
File "/home/khaled/gig/github/rogerthat/appengine/build/mcfw/rpc.py", line 149, in typechecked_return
result = f(*args, **kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/mcfw/rpc.py", line 127, in typechecked_f
return f(**kwargs)
File "/home/khaled/gig/github/rogerthat/appengine/build/rogerthat/bizz/messaging.py", line 1144, in sendMessage
sender=UserDetailsTO.fromUserProfile(profile_info),
File "/home/khaled/gig/github/rogerthat/appengine/build/rogerthat/to/service.py", line 557, in fromUserProfile
to.email = get_human_user_from_app_user(user_profile.user).email()
AttributeError: 'NoneType' object has no attribute 'user'
Expected behavior:
The creation of the profile should be done that it will succeed in any case (try_or_defer ?)
When reinstalling the app the system should inspect the profile to see if something is lacking
I've reinstalled and reconfigured the platform with no problems in the PIL package. But the later problem still persists.
| gharchive/issue | 2016-11-29T10:18:51 | 2025-04-01T06:45:39.082338 | {
"authors": [
"khaledkbadr"
],
"repo": "rogerthat-platform/rogerthat-backend",
"url": "https://github.com/rogerthat-platform/rogerthat-backend/issues/35",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2735847 | License?
Is there some kind of open source license allowing me to change and redistribute this?
closed with commit a8d73e44df4cc6e7bfb72c9f833798ed23c8fd4e
| gharchive/issue | 2012-01-05T14:40:20 | 2025-04-01T06:45:39.083395 | {
"authors": [
"rogerwang",
"thejh"
],
"repo": "rogerwang/node-webkit",
"url": "https://github.com/rogerwang/node-webkit/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
845862400 | Update style overrides for tables
This seems to have better support for tables in Redoc.
Thanks for the PR! But could you add a before/after screenshot of what exactly this is fixing?
@rohit-gohri, thanks for the great work in this library and taking the time, please take a look at my comments above
Thanks for the detailed explanations, will try to merge this tonight
| gharchive/pull-request | 2021-03-31T04:50:45 | 2025-04-01T06:45:39.092405 | {
"authors": [
"qin-guan",
"rohit-gohri"
],
"repo": "rohit-gohri/redocusaurus",
"url": "https://github.com/rohit-gohri/redocusaurus/pull/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
169758409 | Circular Image Elements
Example:
// .img-circle class
div(class="img-circle")
// img-sm img-md img-lg ... classes
div(class="img-circle img-sm")
Maybe sizes relate to location e.g. menu bar is img-sm
Supported http://hagrid.io/knowledge/images
| gharchive/issue | 2016-08-06T17:58:22 | 2025-04-01T06:45:39.099132 | {
"authors": [
"brianofrokk3r",
"gatodeveloper"
],
"repo": "rokk3rlabs/hagrid",
"url": "https://github.com/rokk3rlabs/hagrid/issues/7",
"license": "cc0-1.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1614112644 | 1063 feature add dropdown on user event list page
Description
Please provide a summary of the pull request and the issue it fixes. Please add necessary details, context, dependencies, explanation of when review is needed (see next section), etc.
Fixes #(add issue number here and remove parentheses)
Review Time Estimate
Please give your idea of how soon this pull request needs to be reviewed by selecting one of the options below. This can be based on the criticality of the issue at hand and/or other relevant factors.
[ ] Immediately
[ ] Within a week
[x] When possible
Type of changes
Please select a relevant option:
[ ] Bug fix (non-breaking change which fixes an issue).
[ ] New feature (non-breaking change which adds functionality).
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected).
[ ] Other (any another change that does not fall in one of the above categories.)
Checklist:
Please select all applicable options:
[ ] I have signed the Rokwire Contributor License Agreement (CLA). (Any contributor who is not an employee of the University of Illinois whose official duties include contributing to the Rokwire software, or who is not paid by the Rokwire project, needs to sign the CLA before their contribution can be accepted.)
[ ] I have updated the CHANGELOG.
[ ] I have read the Contributor Guidelines.
[ ] I have performed a self-review of my own code.
[ ] I have commented my code, particularly in hard-to-understand areas.
[ ] My change requires updating the documentation.
[ ] I have made necessary changes to the documentation.
[ ] I have added tests related to my changes.
[ ] My changes generate no new warnings.
[ ] New and existing unit tests pass locally with my changes.
[ ] Any dependent changes have been merged and published in downstream modules.
Hi, @bingzhang, the dimensions of the drop-down control don't seem to match that of the design. Also, the caret's position needs to be corrected. Please fix these. Thanks.
Hi, @bingzhang, the dimensions of the drop-down control don't seem to match that of the design. Also, the caret's position needs to be corrected. Please fix these. Thanks.
Yes, I found that too. I tried to fix it.
Hi, @bingzhang, the dimensions of the drop-down control don't seem to match that of the design. Also, the caret's position needs to be corrected. Please fix these. Thanks.
Yes, I found that too. I tried to fix it.
I was able to position the arrow with something similar to this :
/* Style the arrow inside the select element: */
.select-selected:after {
position: absolute;
content: "";
top: 14px;
right: 10px;
width: 0;
height: 0;
border: 6px solid transparent;
border-color: #ff552e transparent transparent transparent;
}
Hope this is useful
@bingzhang, could you please address the comments above? At least the height of the dropdown menu and caret's position (needs to be at the middle like in the create and edit pages) needs to be updated to match the design. The pagination dropdown also has the same issue.
Please fix the merge conflicts. Thanks.
| gharchive/pull-request | 2023-03-07T20:01:54 | 2025-04-01T06:45:39.113450 | {
"authors": [
"bingzhang",
"minump",
"sandeep-ps"
],
"repo": "rokwire/events-manager",
"url": "https://github.com/rokwire/events-manager/pull/1068",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
59705342 | Fix php < 5.4 compatibility
The proxy feature merged in #44 introduced the array syntax that only works on php versions >=5.4 per http://php.net/manual/en/language.types.array.php introducing an implicit version dependency. This changes the syntax to array() so it still works on older versions of php.
Thanks @unix1 !
Yeah, sorry guys, my bad. I work with version 5.5 now. @unix1 thanks :)
| gharchive/pull-request | 2015-03-03T21:05:50 | 2025-04-01T06:45:39.138764 | {
"authors": [
"Laykou",
"brianr",
"unix1"
],
"repo": "rollbar/rollbar-php",
"url": "https://github.com/rollbar/rollbar-php/pull/46",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
532305414 | Bug: babel.generated filename missing
When using babel.generated, the filename is not being provided to the babel options. https://github.com/rollup/rollup-plugin-babel/blob/master/src/index.js#L182. I think outputOptions.dir + chunk.fileName should give the appropriate filename.
We've not included this intentionally, see https://github.com/rollup/rollup-plugin-babel/pull/333#discussion_r347228883
Could you describe your use case? Why do you expect anything there and how do you want to use it?
in my use case, I have a babel plugin that is providing a stats file in the post hook that is providing meta about the current "filename".
Example code
post(state) {
results[state.opts.filename] = { ... };
writeFile(..., JSON.stringify(results, null, 2));
}
@Andarist any update or workaround you can provide? Maybe you can provide the entire chunk to the babel options or something? Or an alternative suggestion on how to stat an entry file on all its dependencies/async dependencies.
Note that if we don't want to provide this info as filename, we could abuse the caller option and use caller.chunkName.
my workaround
module.exports = function myCustomPlugin() {
let fileName;
const plugin = babel.generated.custom(() => ({
config(cfg) {
return { ...cfg.options, filename: fileName };
}
})({ /* my babel config */ })
return {
...plugin,
renderChunk(code, chunk) {
fileName = chunk.fileName;
return plugin.renderChunk.apply(plugin, arguments);
}
};
}
@lukastaegert @loganfsmyth could you take a look at the use case described here and share your thoughts about how this should be approached?
Hmm, @gingur could you expand more on how the filename is consumed from the JSON? Babel's filename option is generally the path of the original file being compiled, not the path of the output, so with your existing code, I'd expect it to already be extremely weird, since without .generated your JSON contains the paths of the original JS files, and with .generated your JSON would contain the path of the generated on-disk file.
How are the filenames used? Does it matter if the are the paths of the original files or the generated files?
The problem from my perspective is that in the future, filename may no longer be available in the renderChunk hook or in general when Babel runs on the output. The reason is that our current hashing is problematic because it runs BEFORE renderChunk is run and does not respect changes in that hook. A future version will probably change this, with the side effect that we only know that chunk names AFTER all chunks have been processed. This will make your approach impossible, at least when file hashes are involved.
My goal is to generate a JSON description of each entry/chunk, and the statics/dynamics that asset references. I do not care about the original src filenames, only exported output filenames and their content. My current workaround does seem to work, as chunk.fileName is properly formatted based on my entryFileNames/chunkFileNames specified as [name]/[hash].js
| gharchive/issue | 2019-12-03T22:19:51 | 2025-04-01T06:45:39.258766 | {
"authors": [
"Andarist",
"gingur",
"loganfsmyth",
"lukastaegert",
"nicolo-ribaudo"
],
"repo": "rollup/rollup-plugin-babel",
"url": "https://github.com/rollup/rollup-plugin-babel/issues/337",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
279860965 | fix sourcemap throwing error
Resolves https://github.com/roman01la/webpack-closure-compiler/issues/37 and https://github.com/roman01la/webpack-closure-compiler/issues/29
Please ensure you bump package.json version and re-deploy to NPM with new version please.
@roman01la any way to get this merged in? We use this plugin in multiple production projects and I don't really want to maintain my own version of this project. You could give me write access to help maintain this library if you like.
Done! Thanks! I've also added you as a collaborator, feel free to make this project better 😄
| gharchive/pull-request | 2017-12-06T18:36:23 | 2025-04-01T06:45:39.272123 | {
"authors": [
"bansawbanchee",
"roman01la"
],
"repo": "roman01la/webpack-closure-compiler",
"url": "https://github.com/roman01la/webpack-closure-compiler/pull/39",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
195674697 | Rollbac migration error Laravel 5.2
[Symfony\Component\Debug\Exception\FatalThrowableError] │Saving To...
Class 'CreatePermissionUserTable' not found
php artisan optimize
| gharchive/issue | 2016-12-14T23:24:05 | 2025-04-01T06:45:39.273482 | {
"authors": [
"bakiro"
],
"repo": "romanbican/roles",
"url": "https://github.com/romanbican/roles/issues/203",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
66039295 | Feature proposal - list roles
In order to list all roles of a current user I came up with crude, ugly solution:
I just list all of them like that:
@role('Admin')
Admin
@endrole
But in one of my project I will soon have more than 20 roles.
Are you planning any feature which would list all roles of the current user?
If yest, it should include a way to style each role with a css, such as this one:
{!!
join(', ',
array_map(function($o) {
return link_to_route('roles',
$o->name,
[$o->id],
['class' => 'ui blue tag label']
);},
$object->capacities->all())
) !!}
Thx
I think it should return collection so developer can do whatever with data not just show them...
@foreach (Auth::user()->roles()->get() as $role)
{{ $role->name }}
@endforeach
Your code works perfectly! Thank you.
I suggest you include it in your readme.md.
Peter
Yes, thanks.
| gharchive/issue | 2015-04-03T00:20:24 | 2025-04-01T06:45:39.277182 | {
"authors": [
"Kyslik",
"UnrulyNatives",
"romanbican"
],
"repo": "romanbican/roles",
"url": "https://github.com/romanbican/roles/issues/25",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
135803421 | Support for 3.5+?
Can you tell me if this plugin is supported on 3.5+?
@shanegriffiths yes it is. I’m not experiencing any issues with Sketch 3.5.2.
| gharchive/issue | 2016-02-23T17:24:15 | 2025-04-01T06:45:39.283496 | {
"authors": [
"romashamin",
"shanegriffiths"
],
"repo": "romashamin/sort-me-sketch",
"url": "https://github.com/romashamin/sort-me-sketch/issues/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
856541732 | pthome种子下载链接改变
PT 助手版本: 1.5.1
PT 助手安装方式:zip包安装
浏览器名称及版本:谷歌浏览器 89.0.4389.114 (正式版本) (64 位) (cohort: Stable)
浏览器是否安装了其他插件:是
停用其他插件后是否正常工作:否
问题描述:pthome种子下载链接从passkey变为downhash
例:https://www.pthome.net/download.php?id=&downhash=
我觉得这个问题至少在 dev 中已修复 https://github.com/ronggang/PT-Plugin-Plus/pull/706
请不要在该站点的设置中添加passkey,即 此处留空
| gharchive/issue | 2021-04-13T03:53:57 | 2025-04-01T06:45:39.311287 | {
"authors": [
"Rhilip",
"wxuuu"
],
"repo": "ronggang/PT-Plugin-Plus",
"url": "https://github.com/ronggang/PT-Plugin-Plus/issues/773",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1141087652 | 下载服务器暂时离线
您的功能请求是否与问题有关? 请描述一下。
有些下载器可能有段时间不会用到,比如带着笔记本出门一段时间,
但下载服务器多的话又很乱,删的话,后面需要还要重新添加回来
描述你想要的解决方案
所以想有个临时关闭不显示的选项
描述您考虑过的替代方案
其他附加信息
对,挺好的建议
服务器怎么添加
@ted423 已支持
| gharchive/issue | 2022-02-17T09:33:55 | 2025-04-01T06:45:39.313671 | {
"authors": [
"IITII",
"cocoer1",
"ted423",
"zhufeng"
],
"repo": "ronggang/PT-Plugin-Plus",
"url": "https://github.com/ronggang/PT-Plugin-Plus/issues/974",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
280029125 | Module is not compiling. Having some build errors
When I am using ng2-ui-auth, I am not able to build it successfully. getting error as shown below. Please help.
Seems like you have an old version of typescript. are you using 2.3 or above?
You are correct.
I was having typescript version 2.2.
Now I have upgraded the typescript version to 2.3.4. Issue resolved.
Thanks!
Hi Ron Zeidman,
I am using node.js as web service.
Did this module is only for UI authentication or Shall I use it with passport.js in node.js?
You should implement your own server-side endpoints, something similar to https://github.com/sahat/satellizer/blob/master/examples/server/node/server.js
only the parameters the server receives are more organized and look like this:
{
"authorizationData": {
//the same data sent to the auth endpoint
},
"oauthData": {
//the data received from the oauth endpoint
},
"userData": {
//additional data you've provided
}
}
| gharchive/issue | 2017-12-07T07:30:10 | 2025-04-01T06:45:39.334736 | {
"authors": [
"MalateshPatil",
"ronzeidman"
],
"repo": "ronzeidman/ng2-ui-auth",
"url": "https://github.com/ronzeidman/ng2-ui-auth/issues/132",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1481653852 | 🛑 TROJAN 🇸🇬 Singapore SGV 1 is down
In d1a87d2, TROJAN 🇸🇬 Singapore SGV 1 (https://sgt-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇸🇬 Singapore SGV 1 is back up in 0bc629a.
| gharchive/issue | 2022-12-07T11:11:20 | 2025-04-01T06:45:39.348344 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/13259",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1498559391 | 🛑 XRAY 🇬🇧 United Kingdom UKO 1 is down
In fd62c11, XRAY 🇬🇧 United Kingdom UKO 1 (https://ukx-1.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: XRAY 🇬🇧 United Kingdom UKO 1 is back up in 1f05b13.
| gharchive/issue | 2022-12-15T14:40:19 | 2025-04-01T06:45:39.350869 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/13940",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
959005506 | 🛑 TROJAN 🇸🇬 Singapore SGA 1 is down
In f63236b, TROJAN 🇸🇬 Singapore SGA 1 (https://sgt-3.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇸🇬 Singapore SGA 1 is back up in d78ed5a.
| gharchive/issue | 2021-08-03T11:13:24 | 2025-04-01T06:45:39.353687 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/1455",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1524355940 | 🛑 XRAY 🇸🇬 Singapore SGO 2 is down
In 1352093, XRAY 🇸🇬 Singapore SGO 2 (https://sgx-3.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: XRAY 🇸🇬 Singapore SGO 2 is back up in a682b63.
| gharchive/issue | 2023-01-08T06:30:55 | 2025-04-01T06:45:39.356143 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/16146",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1530379855 | 🛑 TROJAN 🇭🇰 Hong Kong HKE 1 is down
In 51460a9, TROJAN 🇭🇰 Hong Kong HKE 1 (https://hkt-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇭🇰 Hong Kong HKE 1 is back up in a216603.
| gharchive/issue | 2023-01-12T09:26:08 | 2025-04-01T06:45:39.358641 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/16514",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1533740111 | 🛑 TROJAN 🇨🇦 Canada CAO 1 is down
In a982030, TROJAN 🇨🇦 Canada CAO 1 (https://cat-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇨🇦 Canada CAO 1 is back up in d5bdb83.
| gharchive/issue | 2023-01-15T09:48:51 | 2025-04-01T06:45:39.361151 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/16824",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1538352140 | 🛑 TROJAN 🇩🇪 Germany DEH 1 is down
In f2b7b61, TROJAN 🇩🇪 Germany DEH 1 (https://det-1.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇩🇪 Germany DEH 1 is back up in 47c5f0a.
| gharchive/issue | 2023-01-18T17:00:26 | 2025-04-01T06:45:39.363652 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/17214",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1558400909 | 🛑 TROJAN 🇸🇬 Singapore SGO 1 is down
In fac7a2c, TROJAN 🇸🇬 Singapore SGO 1 (https://sgt-2.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇸🇬 Singapore SGO 1 is back up in 303017c.
| gharchive/issue | 2023-01-26T16:15:15 | 2025-04-01T06:45:39.366393 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/17546",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2019516465 | 🛑 V2RAY 🇮🇩 Indonesia IDA 2 is down
In 2ef3072, V2RAY 🇮🇩 Indonesia IDA 2 (https://idv-3.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: V2RAY 🇮🇩 Indonesia IDA 2 is back up in 6f5c6a8.
| gharchive/issue | 2023-11-30T21:12:01 | 2025-04-01T06:45:39.368883 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/21341",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2044793946 | 🛑 V2RAY 🇺🇸 United States USO 1 is down
In 5446a16, V2RAY 🇺🇸 United States USO 1 (https://usv-4.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: V2RAY 🇺🇸 United States USO 1 is back up in fdb5e3f.
| gharchive/issue | 2023-12-16T15:23:57 | 2025-04-01T06:45:39.371386 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/22567",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2087066476 | 🛑 TROJAN 🇬🇧 United Kingdom UKM 1 is down
In 0db9378, TROJAN 🇬🇧 United Kingdom UKM 1 (https://ukt-1.optnl.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇬🇧 United Kingdom UKM 1 is back up in 9e0978a.
| gharchive/issue | 2024-01-17T22:00:15 | 2025-04-01T06:45:39.373894 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/24752",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2091958810 | 🛑 V2RAY 🇸🇬 Singapore SGF 3 is down
In 49c7941, V2RAY 🇸🇬 Singapore SGF 3 (https://sgv-3.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: V2RAY 🇸🇬 Singapore SGF 3 is back up in ca97039.
| gharchive/issue | 2024-01-20T08:10:54 | 2025-04-01T06:45:39.376368 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/24904",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1011318166 | 🛑 OVPN 🇮🇩 Indonesia IDJ 1 is down
In 0728878, OVPN 🇮🇩 Indonesia IDJ 1 (http://ido-1.opensvr.net:8080/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: OVPN 🇮🇩 Indonesia IDJ 1 is back up in 5bac689.
| gharchive/issue | 2021-09-29T18:31:19 | 2025-04-01T06:45:39.379193 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/2853",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1065098228 | 🛑 XRAY 🇩🇪 Germany DEH 1 is down
In 43d54a1, XRAY 🇩🇪 Germany DEH 1 (https://dex-1.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: XRAY 🇩🇪 Germany DEH 1 is back up in 41b2d1a.
| gharchive/issue | 2021-11-27T18:00:03 | 2025-04-01T06:45:39.381632 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/4006",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1147378378 | 🛑 XRAY 🇺🇸 United States USF 1 is down
In d8908f1, XRAY 🇺🇸 United States USF 1 (https://usx-1.openv2ray.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: XRAY 🇺🇸 United States USF 1 is back up in b3ebaf2.
| gharchive/issue | 2022-02-22T21:19:46 | 2025-04-01T06:45:39.384087 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/5652",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1417309334 | 🛑 TROJAN 🇺🇸 United States USO 1 is down
In 93a1c6c, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: TROJAN 🇺🇸 United States USO 1 is back up in e0d7394.
| gharchive/issue | 2022-10-20T21:24:03 | 2025-04-01T06:45:39.386546 | {
"authors": [
"roosterkid"
],
"repo": "roosterkid/opentunnel-status-server",
"url": "https://github.com/roosterkid/opentunnel-status-server/issues/9812",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
151620636 | Conflicting autoprefixers?
Since we're using autoprefixer, shouldn't we add autoprefixer: false by default to cssNano since it will remove prefixes for IE etc. if we add them to the browsers list of autoprefixer?
The part of the gulp-file I'm talking about:
.pipe(autoprefixer, { browsers: [ 'last 2 versions', 'android 4', 'opera 12' ] }) .pipe(cssNano, { safe: true })
I had a problem with this today. The default autoprefixer browser list do not add prefix to justify-content, and Safari 8 need this.
So I edit this part to: .pipe(autoprefixer, {browsers: ['last 2 versions', 'safari 8', 'android 4', 'opera 12'] }).
But cssNano run autoprefixer too, and by default do not include prefix to justify-content. So what autoprefixer add, cssNano remove.
The solution was to disable autoprefixer in cssNano, as suggested above by @Nocklas.
closing as we're not focused on sage 8, but would be open to a PR if someone cares enough about it
| gharchive/issue | 2016-04-28T11:44:04 | 2025-04-01T06:45:39.422909 | {
"authors": [
"Nocklas",
"marcelo2605",
"retlehs"
],
"repo": "roots/sage",
"url": "https://github.com/roots/sage/issues/1639",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
461734991 | Prepare and Test for RaspberryPi 4
The RaspberryPi 4 got released and RaspiBlitz v1.3 should be ready for this great hardware update.
TODO: Build SD card based on "Raspbian Buster Lite"
https://www.raspberrypi.org/downloads/raspbian/
TODO: Test with RaspberryPi 4
TODO: Test with RaspberryPi 3 compatibility
Compatibility test with RaspberryPi 3 tests looking good
OK new v1.3 Raspiblitz test image build on Raspbian Buster is running fine one RaspberryPi 4 :)
Powertest needs to be updated:
On RaspberryPi4 the default seems with stable power 0.8718 V / 871800 microVolt - starting form 76 C temp it seems to throttle down the CPU speed.
TODO: How to detect a RaspberryPi4?
OK - Power/Stresstest can now detect RaspberryPi 4 and adapts the treshhold values.
Ordered https://shop.pimoroni.com/products/fan-shim for cooling testing
Fixed the "self-sync" option because RP4 gets in the range to make the job in around 2-3 - testing at the moment.
Running a "self-sync" test with RP4 .. first 20% was a overnight thing (that part where the blockchain was not that full yet) but then it gets slowly. From the look of htop CPU is just 50% and memory looks just a 1/3 is used (of 2GB). So it maybe HDD slowing things down.
Will also run "self-sync" test with SSD soon. If that works out in a fair time it may make sense to measure HDD/SSD speed on bootup and then advise to self-sync or not.
Running HDD speed tests...
sudo apt-get install hdparm
sudo hdparm -Tt /dev/sda
RP3 + HDD:
Timing cached reads: 1184 MB in 2.00 seconds = 592.29 MB/sec
Timing buffered disk reads: 92 MB in 3.05 seconds = 30.13 MB/sec
RP3 + SSD:
Timing cached reads: 1238 MB in 2.00 seconds = 619.16 MB/sec
Timing buffered disk reads: 94 MB in 3.03 seconds = 31.05 MB/sec
RP4 + HDD:
Timing cached reads: 1452 MB in 2.00 seconds = 726.00 MB/sec
Timing buffered disk reads: 372 MB in 3.01 seconds = 123.61 MB/sec
RP4 + SSD:
TODO
Testing the RPi4 4GB with the official 3A USB-C powersupply.
The HWtest was failing, so disconnected everything (only bare board with the power supply, ethernet and SDcard attached):
*** Hardware Test Report ***
RaspiBlitz Hardwaretest v0.2
RaspberryPi Model Version: 4
Starting sysbench to run for 60 seconds (--max-time=60 --cpu-max-prime=10000)
58.0'C 1500 MHz 0.8630V
--> Power WARN detected
65.0'C 1500 MHz 0.8630V
--> Power WARN detected
67.0'C 1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
68.0'C 1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
70.0'C 1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
--> Temp WARN detected
63.0'C 600/1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
62.0'C 600/1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
59.0'C 1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
59.0'C 1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
60.0'C 600/1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
58.0'C 600/1500 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
58.0'C 600/ 600 MHz 0.8630V
--> Power CRITICAL detected
--> Power WARN detected
It seems that my default voltage is 0.8630 V.
I have no heatsink on and as it looks the CPU is severely throttled above 60 degrees.
Running HDD speed tests...
sudo apt-get install hdparm
sudo hdparm -Tt /dev/sda
RP3 + HDD:
Timing cached reads: 1184 MB in 2.00 seconds = 592.29 MB/sec
Timing buffered disk reads: 92 MB in 3.05 seconds = 30.13 MB/sec
RP3 + SSD:
Timing cached reads: 1238 MB in 2.00 seconds = 619.16 MB/sec
Timing buffered disk reads: 94 MB in 3.03 seconds = 31.05 MB/sec
RP4 + HDD:
Timing cached reads: 1452 MB in 2.00 seconds = 726.00 MB/sec
Timing buffered disk reads: 372 MB in 3.01 seconds = 123.61 MB/sec
RP4 + SSD:
TODO
RPi4 + SSD: (Samsung 860 EVO 500GB)
Timing cached reads: 1706 MB in 2.00 seconds = 853.70 MB/sec
Timing buffered disk reads: 1104 MB in 3.00 seconds = 367.52 MB/sec
The new shopping list will be RP4 focused - just mentioning that RPi3B+ also can be used if you alread have one.
The HardwareTest is already adapted for the RP4 in v1.3 branch.
The HardwareTest is already adapted for the RP4 in v1.3 branch
I am not sure why but the voltage on my Pi is slightly lower: 863000 mV vs. 871800 mV. It could be because of the 4GB vs 2GB RAM? Are you using the official RPi power supply too?
I adapted the power test - I used my Macbook power supply ;)
Running the Initial Block Download (self-sync) behind Tor on the RPi4 4GB and SSD - first 30% was ~12 hours. Already quicker than my Odroid HC1/XU4 was on clearnet.
Really interesting tests happening here guys, hopefully, we have that shopping list updated really soon. Well done.
OK testes the Shim fan for active cooling. It fits under the LCD and the LCD is still working. Its effective cooling - even in stress test the temp gets just to upper 50s. In normal operations mine is 45 celsius. I recommend this fan to keep temp in check on the RP4.
Tested Self-Syncing/Validation: Was not able to finish 100% due to time ... but was able to sync up to 80% in about 3 days - over last night (half day) it was able todo 8% ... so its realistic to fully self-sync a RaspiBlitz with RP4&SSD in around 4-5 days. And the temp with the shim fan was during sync at around 47/48 celsius .. so super cool :)
There is some fun python code to control the fan shim and built in LED. Probably should not be included by default, but worth a mention in the readme, it is great to play with:
https://learn.pimoroni.com/tutorial/sandyj/getting-started-with-fan-shim
https://github.com/pimoroni/fanshim-python/tree/master/examples
Two finished IBD times on the RPi4 4GB with and SSD behind Tor: 80 and 96.5 hours https://github.com/rootzoll/raspiblitz/issues/592
There's bug with power with Pi4, it will be fixed in the next months.
https://www.techrepublic.com/article/your-new-raspberry-pi-4-wont-power-on-usb-c-cable-problem-now-officially-confirmed/
Hello all, new to Raspiblitz and testing out the new RC3 for my rpi4
got as far as raspian and
``wget https://raw.githubusercontent.com/rootzoll/raspiblitz/v1.3/build_sdcard.sh && sudo bash build_sdcard.sh
upon reboot it complains about not being able to find bitcoind
https://termbin.com/35bq
I noticed an issue with flashing the http://raspiblitz.com/raspiblitz-v1.3RC1-2019-07-04.img.gz to sdcard only when I did zcat raspiblitz-v1.3RC1-2019-07-04.img.gz | sudo dd bs=4M of=/dev/mmcblk0 conv=fsync status=progress' did it manage to flash the card correctly and boot.
I've created a new issue from RC3 regarding RTL hangs
The Initial Block Download should not be called ultra slow in case the acripts are running on any more powerful hw (2GB or more RAM and USB3) than the RPi3 . That is the best option to start the node and should be communicated to the user in the setup menu.
Even lower voltage results were spotted on Telegram:
HDMI external monitor, also the touch screen attached, keyboard and mouse, HDD: USB Seagate 1TB.
— WITH USB HDD attached —
67.0'C 1500 MHz 0000000000 0.8332V
66.0'C 1500 MHz 0000000000 0.8332V
66.0'C 1500 MHz 0000000000 0.8332V
65.0'C 1500 MHz 0000000000 0.8332V
65.0'C 1500 MHz 0000000000 0.8332V
64.0'C 1500 MHz 0000000000 0.8332V
— WITHOUT USB HDD attached —
69.0'C 1500 MHz 0000000000 0.8332V
67.0'C 1500 MHz 0000000000 0.8332V
67.0'C 1500 MHz 0000000000 0.8332V
66.0'C 1500 MHz 0000000000 0.8332V
66.0'C 1500 MHz 0000000000 0.8332V
66.0'C 1500 MHz 0000000000 0.8384V
65.0'C 1500 MHz 0000000000 0.8332V
As attaching the HDD seems to make no difference I suggested to carry on with the blockchain download. Will follow up. We might need to reduce the power limit further if there will be no issues here.
@openoms as long there are no "Bad Power" reports in the logs we should be fine to lower limit in the stress test for RP4 still a bit.
| gharchive/issue | 2019-06-27T20:31:55 | 2025-04-01T06:45:39.448093 | {
"authors": [
"CommanderPoe",
"annaparker",
"openoms",
"rootzoll",
"woeisme"
],
"repo": "rootzoll/raspiblitz",
"url": "https://github.com/rootzoll/raspiblitz/issues/673",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.