added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:35:58.565849
| 2019-04-10T06:44:51
|
431330405
|
{
"authors": [
"isharak",
"janakamarasena",
"nilasini"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12129",
"repo": "wso2/product-is",
"url": "https://github.com/wso2/product-is/issues/5030"
}
|
gharchive/issue
|
[DOC][IS530] Configuring token expiry time per service provider
There are Service Provider level configurations for configuring "token expiry times" and "refresh token renewal per request" in IS530. This needs to documented.
@janakamarasena Is it only missing in 5.3.0?
@nilasini not sure on that... Also not sure what other versions have this particular configuration
Seems like it is there in 5.3.0. Could you please recheck and confirm?
Hi @nilasini looks like its there in the doc[1] we only need to add the "refresh token renewal per request" config. Thanks for pointing it out.
The configuration part is located under
Configuring Inbound Authentication for a Service Provider
> OAuth/OpenID Connect Configuration
>> Defining access token validity period
>>> Defining access token validity period per service provider
[1] - https://docs.wso2.com/display/IS530/Configuring+Inbound+Authentication+for+a+Service+Provider
@YvonneW can you please do the needful.
This issue is being closed due to extended inactivity. Please feel free to reopen it if further attention is needed. Thank you for helping us keep the issue list relevant and focused!
|
2025-04-01T04:35:58.765846
| 2021-03-31T17:35:41
|
847020208
|
{
"authors": [
"naturecodevoid",
"sindresorhus"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12130",
"repo": "wulkano/Kap",
"url": "https://github.com/wulkano/Kap/issues/1000"
}
|
gharchive/issue
|
Exporting to MP4 (AV1) takes a ridiculously long time
macOS version: 10.14.6
Kap version: 3.3.2
Steps to reproduce
Capture a short video
Export at MP4 (AV1)
Observe
Current behaviour
It takes very long to export, I haven’t tested if it will ever finish as the estimated time remaining went up as I waited.
Expected behaviour
Similar export time to MP4 (H264)
Workaround
Don’t think there is one.
Notes:
Estimated time remaining only went up
MP4 (H264) does not experience this problem
Duplicate of #939
|
2025-04-01T04:35:58.784421
| 2020-12-11T11:28:51
|
762258803
|
{
"authors": [
"Ray112972",
"albinekb",
"skllcrn",
"sunknudsen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12131",
"repo": "wulkano/Kap",
"url": "https://github.com/wulkano/Kap/issues/966"
}
|
gharchive/issue
|
Capturing at 23.976 FPS
Hey guys, love Kap!
I use the app to capture my screen while recording privacy guides episodes.
I record myself using a Sony a6500 at 23.976 FPS (which is more cinematic) and my screen using Kap.
It would be amazing if we could capture at 23.976 FPS. The file size would be smaller and more importantly, it would allow for better audio synchronization in Adobe Premiere (given both camera and screen capture would be recorded at the same frame rate).
macOS version:
ProductName: Mac OS X
ProductVersion: 10.15.7
BuildVersion: 19H15
Kap version:
Version 3.3.2 (<IP_ADDRESS>8)
Steps to reproduce
Open "Preferences" / "Capture frame rate".
Current behaviour
It is only possible to capture at 30 or 60 FPS.
Expected behaviour
Being able to capture at 23.976 FPS.
Workaround
None that I know about, but perhaps you guys know of a command-line hack. 🤞
You could use https://github.com/wulkano/aperture directly, I made a simple CLI to run it without kap. If you just want to record at 24 fps, and not use Kap to edit/export/upload, you could write your own cli, using https://github.com/albinekb/aperture-cli/blob/master/cli.js as a base. I can add --fps option for you if you think this would solve your problem. We're trying to keep Kap as simple as possible 🙏
You could use https://github.com/wulkano/aperture directly, I made a simple CLI to run it without kap. If you just want to record at 24 fps, and not use Kap to edit/export/upload, you could write your own cli, using https://github.com/albinekb/aperture-cli/blob/master/cli.js as a base. I can add --fps option for you if you think this would solve your problem. We're trying to keep Kap as simple as possible 🙏
Thanks for the follow-up @albinekb.
Btw, I really appreciate the work you guys do on Kap. Best screen capture app by far!
We're trying to keep Kap as simple as possible
24 FPS (or to be specific 23.976 FPS for the NTSC standard) is very common, especially on YouTube... my gut feeling is many creators would benefit if you guys added that frame-rate to the preferences.
First, 23.976 FPS would be more gentle on CPU usage while recording. Second, matching timeline frame-rate of video editing software would be more gentle on CPU usage as well.
Please, please, please consider adding 23.976 FPS to the list!
Thanks for the follow-up @albinekb.
Btw, I really appreciate the work you guys do on Kap. Best screen capture app by far!
We're trying to keep Kap as simple as possible
24 FPS (or to be specific 23.976 FPS for the NTSC standard) is very common, especially on YouTube... my gut feeling is many creators would benefit if you guys added that frame-rate to the preferences.
First, 23.976 FPS would be more gentle on CPU usage while recording. Second, matching timeline frame-rate of video editing software would be more gentle on CPU usage as well.
Please, please, please consider adding 23.976 FPS to the list!
You can view the capture frame rate as an upper limit @sunknudsen, meaning that depending on your hardware, fewer frames per second may be captured. While it is true that 24 FPS is a well established standard, YouTube as you brought up as an example will attempt to play back at a higher frame rate if possible. Note that once you've done recording, you can set the Kap export FPS to 24, and the setting will be remembered. Does that do the trick?
Hey @skllcrn, thanks so much for the feedback.
You can view the capture frame rate as an upper limit @sunknudsen, meaning that depending on your hardware, fewer frames per second may be captured.
Interesting... my gut feeling is that most contemporary hardware can easily capture a 4K display at 24 FPS... so I guess this applies especially at higher frame rates.
Note that once you're done recording, you can set the Kap export FPS to 24, and the setting will be remembered. Does that do the trick?
Unfortunately, this doesn't work great... First, the export for a 15 minutes screen capture is pretty long and very hardware intensive (I cannot do much simultaneously). Second (and I was surprised about that) editing the original capture ("File" / "Save Original...") is easier on the CPU in Adobe Premiere Pro 2020. I believe most (if not all) video editing software favor less compressed files.
Therefore, capturing at 23.976 FPS would make 24P workflows much more efficient.
If adding 23.976 FPS to the UI doesn't reach consensus, a command line "hack" to set custom frame rates would be amazing!
Hey @skllcrn, thanks so much for the feedback.
You can view the capture frame rate as an upper limit @sunknudsen, meaning that depending on your hardware, fewer frames per second may be captured.
Interesting... my gut feeling is that most contemporary hardware can easily capture a 4K display at 24 FPS... so I guess this applies especially at higher frame rates.
Note that once you're done recording, you can set the Kap export FPS to 24, and the setting will be remembered. Does that do the trick?
Unfortunately, this doesn't work great... First, the export for a 15 minutes screen capture is pretty long and very hardware intensive (I cannot do much simultaneously). Second (and I was surprised about that) editing the original capture ("File" / "Save Original...") is easier on the CPU in Adobe Premiere Pro 2020. I believe most (if not all) video editing software favor less compressed files.
Therefore, capturing at 23.976 FPS would make 24P workflows much more efficient.
If adding 23.976 FPS to the UI doesn't reach consensus, a command line "hack" to set custom frame rates would be amazing!
Btw, I am not a video expert... for example, I don't fully understand the technical nuances between 24 FPS and 23.976 FPS (frame rate at which my Sony a6500 records).
That being said, I am pretty confident that capturing at 30 FPS when editing on a 23.976 FPS timeline and/or editing compressed files is not efficient and hurts the creative process.
Feels like this could be avoided by allowing users to choose the frame rate that suits a project.
Thanks again for Kap!
Btw, I am not a video expert... for example, I don't fully understand the technical nuances between 24 FPS and 23.976 FPS (frame rate at which my Sony a6500 records).
That being said, I am pretty confident that capturing at 30 FPS when editing on a 23.976 FPS timeline and/or editing compressed files is not efficient and hurts the creative process.
Feels like this could be avoided by allowing users to choose the frame rate that suits a project.
Thanks again for Kap!
Got it, thank you for the additional detail on your use case. The history and longevity of this standard is fascinating btw., worth reading about!
Got it, thank you for the additional detail on your use case. The history and longevity of this standard is fascinating btw., worth reading about!
Uploading Kellz.zip…
Uploading Kellz.zip…
https://user-images.githubusercontent.com/64635166/104035798-8d2bb080-51a0-11eb-8783-db0e98062c6a.mov
https://user-images.githubusercontent.com/64635166/104035798-8d2bb080-51a0-11eb-8783-db0e98062c6a.mov
|
2025-04-01T04:35:58.789416
| 2019-05-07T14:37:52
|
441269776
|
{
"authors": [
"mackness",
"skllcrn"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12132",
"repo": "wulkano/kap",
"url": "https://github.com/wulkano/kap/issues/670"
}
|
gharchive/issue
|
Tried to start a recording after initial install and got an error
macOS version:
macOS Mojave 10.14.4
Kap version:
Version 2.2.0 (<IP_ADDRESS>0)
Steps to reproduce
Install Kap for the first time and start a recording.
Current behaviour
Error message with the output specified below.
Unhandled Promise Rejection
Error: Call `.startRecording()` first
at Aperture.stopRecording (/Applications/Kap.app/Contents/Resources/app.asar/node_modules/aperture/index.js:141:13)
at EventEmitter.<anonymous> (file:///Applications/Kap.app/Contents/Resources/app.asar/dist/renderer/js/main.js:207:39)
at Generator.next (<anonymous>)
at step (file:///Applications/Kap.app/Contents/Resources/app.asar/dist/renderer/js/main.js:39:191)
at file:///Applications/Kap.app/Contents/Resources/app.asar/dist/renderer/js/main.js:39:437
at Promise (<anonymous>)
at EventEmitter.<anonymous> (file:///Applications/Kap.app/Contents/Resources/app.asar/dist/renderer/js/main.js:39:99)
at EventEmitter.stopRecording (file:///Applications/Kap.app/Contents/Resources/app.asar/dist/renderer/js/main.js:223:20)
Expected behaviour
Be able to record without error.
Workaround
n/a
Duplicate of #596
|
2025-04-01T04:35:58.790636
| 2022-09-21T11:44:46
|
1380793991
|
{
"authors": [
"jensneuse"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12133",
"repo": "wundergraph/wundergraph",
"url": "https://github.com/wundergraph/wundergraph/issues/201"
}
|
gharchive/issue
|
Uploading files pollutes the logs
Currently, when running wunderctl up --debug we're logging all requests including req+res.
That's problematic in case of file uploads, because we're printing the file to the std out.
Ideally, we could detect if the request is a file upload and prevent logging the body to keep the logs clean.
fixed
|
2025-04-01T04:35:58.908011
| 2018-09-07T11:59:48
|
358038983
|
{
"authors": [
"nknapp",
"pepkin88"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12134",
"repo": "wycats/handlebars.js",
"url": "https://github.com/wycats/handlebars.js/issues/1465"
}
|
gharchive/issue
|
{{@value}} prefers to use a helper named value instead of a data variable named value
I'm currently upgrading from Handlebars v1 to v4 in my project and I stumbled upon this strange behavior, which seems like a bug to me.
Test case: https://jsfiddle.net/64zfydjp/
(shows helper helper instead of helper data)
Previous behavior (last version: 2.0.0): https://jsfiddle.net/9rtn75sq/2/
(shows helper data, as I expected)
If I'm using @ I'm expecting to use a data variable, not a helper. But even I wanted a helper, it should be named @value, not just value. Moreover, the helper is invoked identically, so I cannot differentiate those cases in the helper code.
Is this behavior really expected or is it a bug? If it's expected, how could I differentiate {{@value}} from {{value}}?
I have to check this. Seems to me like a bug, but you never know until you see the change that's causing it and the reason why it was introduced.
I believe that commit 884bf1553663734f22ffcd9d758c9d71d4373bf9 introduced this behavior. I have not tested it thought.
The precompiler compiles {{@value}} and {{value}} to the same code. This is certainly strange. But it has been in the code since 3.0.0 and I would be very hesitant to change it. I can already smell unexpected side effects from changing it (and bug reports).
I may change my mind on this, but it's certainly not something that is fixed quickly.
The hardest part for me, was to locate the reason why my helper stopped working correctly. After that I applied a fix/workaround that was enough for my project. That means there are still some edge cases which would cause a collision, but now I'm aware of them.
So I can live without a proper fix for it. And maybe this Issue will help someone else with a similar problem.
If someone was interested about my use case, I made a block helper named v, which executes a block if a passed value was truthy, with assigning that value to a variable @v (easy to remember). Example usage:
{{#v user.fullName}}Name: {{@v}}{{/v}}
Closing due to inactivity.
|
2025-04-01T04:35:58.967981
| 2018-07-05T19:28:59
|
338698035
|
{
"authors": [
"132ikl",
"x70b1"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12135",
"repo": "x70b1/polybar-scripts",
"url": "https://github.com/x70b1/polybar-scripts/pull/60"
}
|
gharchive/pull-request
|
Added info-stream-countdown
I made this at like 3 AM, so let me know if anything needs to be redone. It's a bit of a mess but there shouldn't be any major issues, barring the 200 MiB memory usage.
Give me some days. I check this as soon as possible.
Thanks. I'm going to disappear for a week on Sunday, so please keep it open until I get back if any changes need to be made.
Fixed a couple silly things, as tends to happen after actually using the module for a day. I should probably start testing these before PRing them but I get too excited lol.
Can you fix the shellsheck issues first?
SC2088 is irrelevant, since the path is just used to fit with the skeleton.
SC2034 is irrelevant, since the echo is commented out by the boilerplate icon to fit with the skeleton.
SC2154 is irrelevant, because showSeconds is used as a easy flip on/off config option at the top, and the issue goes away if you turn on the showSeconds variable.
I merged it. But that scripts spawns a chrome every time. Is there no more simple solution for this?
Unfortunately, I don't use twitch. I can't say anything about that.
|
2025-04-01T04:35:58.982608
| 2020-06-17T14:33:58
|
640493329
|
{
"authors": [
"czerwinskilukasz1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12136",
"repo": "xFAANG/askql",
"url": "https://github.com/xFAANG/askql/issues/181"
}
|
gharchive/issue
|
Add more examples to the AskScript Playground
The Askscript Playground has a couple of examples not but it will be great to add there 15-20 more, half of them showing different features of the language and half of them being complete programs doing "real things".
My idea is to write in AskScript solutions for some coding tasks for Javascript/Typescript from https://www.codewars.com/ or a similar website.
|
2025-04-01T04:35:58.987475
| 2020-06-25T14:06:00
|
645592249
|
{
"authors": [
"czerwinskilukasz1",
"mhagmajer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12137",
"repo": "xFAANG/askql",
"url": "https://github.com/xFAANG/askql/issues/222"
}
|
gharchive/issue
|
Custom resources no longer accessible
Between Jun 16 and Jun 21 AskVM stopped supporting custom resources.
Steps to reproduce:
Add to cli.ts the following snippet:
export const customResources: Resources = {
hi: resource<string, []>({
resolver: async (): Promise<string> => {
return "Hi, this is AskQL server! It's " + new Date().toString();
},
}),
};
and change:
const options: Options = {
resources,
values,
};
to:
const options: Options = {
resources: { ...resources, ...customResources },
values,
};
For commit 2bca44, Tue Jun 16 14:34:31 2020 +0200 CLI works as follows:
🦄 hi
string ask(get('hi'))
"Hi, this is AskQL server! It's Thu Jun 25 2020 15:55:17 GMT+0200 (Central European Summer Time)"
🦄
But for commit f8793d1 aka v1.1.0, Sun Jun 21 00:04:04 2020 +0200 instead of the resource CLI returns:
🦄 hi
code ask(get('hi'))
AskCode { name: 'hi', params: undefined }
@mhagmajer , was there any breaking change to the resource definition?
@czerwinskilukasz1 yes, when you reference hello being a resource like this you now get back a function, not the result of calling it. I needed to make this change to fix https://github.com/xFAANG/askql/pull/178. This only concerns referencing resources for value (aka "getting"), not calling them which is our main use case. That's why I wouldn't call it a breaking change. We have often used it for empty, true, false resources which are now parsed to calling underlying resources. So what you are seeing here is the expected behaviour and the previous one was a bug.
You need to do hi()
Got it, this is the change, thanks.
|
2025-04-01T04:35:59.002526
| 2018-10-22T01:17:07
|
372373372
|
{
"authors": [
"MastaC729",
"xPXpanD"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12138",
"repo": "xPXpanD/PixelmonBroadcasts",
"url": "https://github.com/xPXpanD/PixelmonBroadcasts/issues/1"
}
|
gharchive/issue
|
Shiny Pokemon Not Always Detected by Event Handler
On our server we have noticed that occasionally shiny spawns will not broadcast to the chat. It only seems to happen intermittently, but I've noticed it happens 100% of the time when I spawn it in with /pokespawn or if it was spawned by fishing. Our server also has a number of other Sponge plugins running on it, and tps has been somewhat bad at times just to give a bit more context.
Versions
Pixelmon Broadcasts version: 0.2.2
Pixelmon version: 6.3.4
Forge version: 2705
SpongeForge version: 3481
Let me know if I am missing any pertinent information.
Commands are known, as they don't fire the right events -- I might be able to work around that eventually, but for now support for that is on the back burner. Fishing is interesting, though. I'll have to look into that for the upcoming 0.2.3 release. Hopefully there's an event there that I can intercept as well.
I've also had one other report of normal spawns occasionally not showing up, though I'm not too sure what could cause that yet. I haven't been able to reproduce it in testing yet, either.
For now... Do you use any plugins or sidemods that could also be listening for Pokémon spawns? If so, which ones, and what versions? Might be able to narrow things down a little with multiple mod lists.
Thanks for the report.
After testing it on my private server running only LuckPerms and GriefPrevention as the only other plugins, it appears that every shiny that naturally spawns is broadcasting properly. So it very well could be one of these plugins we have installed on our server that is causing the issue.
Here are the plugins we have installed:
FastAsyncWorldEdit
Nucleus
ArmorStandTools
Holograms
LuckPerms
GriefPrevention
UltimateChat
EMPSponge
MCClans
Punish
EconomyLite
TeslaCrate
PlayerShopsRPG
PixelExtras
AdamantineShield
TeslaPowered
Gameshark
DailyRewardsExtended
PixelBuilt-Quests
PlayerGrow
SlotMachines
AutoBroadcaster
WonderTrade
TrainerItem
PokeExpMultiplier
PokeDisguise
TabModifier
PixelmonEconomyBridge
PokeDisplay
I know not all of these are related to Pixelmon, but I posted them just in case you may think they are affecting your plugin somehow.
Thanks for the list. I don't see anything there that I'd expect to cause any issues (except for maybe PokéDisguise? not sure how that spawns stuff), but I'll keep an eye out.
To anybody else reading this with the same issue: please add your mod/plugin list as well! Would be nice to get this pinned down.
Closing due to age!
|
2025-04-01T04:35:59.004388
| 2021-06-21T05:25:47
|
925838798
|
{
"authors": [
"PushpenderSaini0",
"xXNurioXx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12139",
"repo": "xXNurioXx/Ads-Link-Skipper",
"url": "https://github.com/xXNurioXx/Ads-Link-Skipper/pull/27"
}
|
gharchive/pull-request
|
ES6 standard
Update the code to ES6 Standard.
Each is updated in a different commit.
Changed variable declaration from var to let and const
Also changes some functions to ES6 () => {} arrow function syntax
Moved order of code in adfly.js to solve code being used before declaration
Please let me know if more changes are to be made in this PR.
I've renamed your commits to be more descriptive, thank you for your work! 🥳
|
2025-04-01T04:35:59.021961
| 2017-08-14T08:06:12
|
249959398
|
{
"authors": [
"ChadTaljaardt",
"focussing",
"ssuess",
"xaksis"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12140",
"repo": "xaksis/vue-good-table",
"url": "https://github.com/xaksis/vue-good-table/issues/43"
}
|
gharchive/issue
|
How do i make a selectable option?
I want to have the first field to have a checkbox for each row, and then when the user selects a checkbox it will add it to the data model, how would i do this?
There are a few ways you can do it... the most straight forward way is to use your own template for the rows.
define your first column as empty {label: '', field: ''} (this will make space for checkbox)
within the template, add first td as your checkbox
<td><input type="checkbox" @click="checkboxClicked(props.row)"/></td>
then in your method checkboxClicked do whatever you want with the clicked row.
closing... please re-open if you have more questions/concerns.
Hello!
How can I get the row number, in stead of the row content?
Best regards, Met vriendelijke groet,
Raymond Verbruggen
FOCUSSING B.V.
Mispel 8
7621 WB Borne
Netherlands
+31 6 250 500 69
<EMAIL_ADDRESS>www.focussing.nl
Op 14 aug. 2017, om 14:00 heeft xaksis<EMAIL_ADDRESS>het volgende geschreven:
There are a few ways you can do it... the most straight forward way is to use your own template for the rows.
define your first column as empty {label: '', field: ''} (this will make space for checkbox)
within the template, add first td as your checkbox
then in your method checkboxClicked do whatever you want with the clicked row.
closing... please re-open if you have more questions/concerns.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub https://github.com/xaksis/vue-good-table/issues/43#issuecomment-322172005, or mute the thread https://github.com/notifications/unsubscribe-auth/AZf2GIuXUFYPt61qDHT8QZ2_0cxUn65qks5sYDbsgaJpZM4O2Dn3.
@focussing the readme contains this info, but here it is again:
In addition to prop.row that contains the row object, prop.index contains the index for the table display row. And prop.row.originalIndex contains the original row index. You can access the original row object by using row[prop.row.originalIndex].
Sorry to bother you; with “originalIndex” you mean the row index before any sorting and filtering?
Best regards, Met vriendelijke groet,
Raymond Verbruggen
FOCUSSING B.V.
Mispel 8
7621 WB Borne
Netherlands
+31 6 250 500 69
<EMAIL_ADDRESS>www.focussing.nl
Op 18 aug. 2017, om 15:48 heeft xaksis<EMAIL_ADDRESS>het volgende geschreven:
@focussing https://github.com/focussing the readme contains this info, but here it is again:
In addition to prop.row that contains the row object, prop.index contains the index for the table display row. And prop.row.originalIndex contains the original row index. You can access the original row object by using row[prop.row.originalIndex].
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub https://github.com/xaksis/vue-good-table/issues/43#issuecomment-323358702, or mute the thread https://github.com/notifications/unsubscribe-auth/AZf2GC_D3Vi-wJUAIYduYqqz8ABieENlks5sZZZIgaJpZM4O2Dn3.
@focussing no problem. yes... originalIndex is the array index for the object without sorting/filtering/paging.. so if you do myRowArray[prop.row.originalIndex], this will give you the entire object for that row.
thanks!
working now on my 3rd project with your table :)
Best regards, Met vriendelijke groet,
Raymond Verbruggen
FOCUSSING B.V.
Mispel 8
7621 WB Borne
Netherlands
+31 6 250 500 69
<EMAIL_ADDRESS>www.focussing.nl
Op 18 aug. 2017, om 15:56 heeft xaksis<EMAIL_ADDRESS>het volgende geschreven:
@focussing https://github.com/focussing no problem. yes... originalIndex is the array index for the object without sorting/filtering/paging.. so if you do myRowArray[prop.row.originalIndex], this will give you the entire object for that row.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub https://github.com/xaksis/vue-good-table/issues/43#issuecomment-323360629, or mute the thread https://github.com/notifications/unsubscribe-auth/AZf2GEb-cZWy-iXnTK__BDiKhONCTi66ks5sZZgFgaJpZM4O2Dn3.
Hi, I was able to add a row in the correct formatting, modifying your Table.vue with this:
<tbody>
<tr v-for="(row, index) in paginated" :class="onClick ? 'clickable' : ''" @click="click(row, index)">
<th v-if="lineNumbers" class="line-numbers">{{ getCurrentIndex(index) }}</th>
<slot name="table-row" :row="row" :index="index">
<td v-for="(column, i) in columns" :class="getDataStyle(i, 'td')" v-if="!column.hidden">
<span v-if="!column.html">{{ collectFormatted(row, column) }}</span>
<span v-if="column.html" v-html="collect(row, column.field)"></span>
</td><td><a class="primary" @click='editRow(props.row)'>edit</a></td>
</slot>
</tr>
</tbody>
And then adding a method:
editRow() {
alert("edit:" + props.row);
},
but when I click on my edit link, I get errors:
[Vue warn]: Property or method "props" is not defined on the instance but referenced during render. Make sure to declare reactive data properties in the data option.
AND
TypeError: undefined is not an object (evaluating '_vm.props.row')
Any ideas why? Thanks!
Nevermind, I just figured out I needed to include my template code in my own vue, not the base Table.vue file.
|
2025-04-01T04:35:59.059276
| 2020-10-21T08:20:08
|
726252755
|
{
"authors": [
"AndreiMisiukevich",
"Tommigun1980",
"VladislavAntonyuk",
"pictos"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12141",
"repo": "xamarin/XamarinCommunityToolkit",
"url": "https://github.com/xamarin/XamarinCommunityToolkit/issues/477"
}
|
gharchive/issue
|
[Enhancement] Allow configuring snackbar in one place
ACT.UserDialogs's toasts allows one to set up toast defaults in one place.
I think XCT's snackbar should separate the config from actually showing the notifications, as currently one has to send in all the styling, durations etc every time a snackbar is shown.
I'd like to set up defaults once (theming, fonts, durations etc.), and in my code just define the message and actions when I show them. One has now to make a wrapper class or utility methods for XCT's snackbar that always sends in the theming, or duplicate a ton of code.
As a pointer, this is roughly how I am setting up ACT.UserDialog's toasts:
private void SetupToastDefaults()
{
ToastConfig.DefaultPosition = ToastPosition.Top;
ToastConfig.DefaultDuration = TimeSpan.FromSeconds(4.5);
ToastConfig.DefaultBackgroundColor = (Color)this.Resources["SomeColor"];
ToastConfig.DefaultMessageTextColor = (Color)this.Resources["SomeOtherColor"];
ToastConfig.DefaultActionTextColor = (Color)this.Resources["YetAnotherColor"];
}
There is no equivalent for XCT's snackbar -- I'd have to send in all of those every time a notification is shown.
I'll ping you first and ask later
@VladislavAntonyuk any thoughts?
Do you mind if I mention you in the snackbar issues?
It can be implemented. I can do it after finishing with DrawingView PR
Thanks guys, I appreciate it.
in progress
Fixed in https://github.com/xamarin/XamarinCommunityToolkit/pull/508
Thanking you here as well @VladislavAntonyuk, absolutely fantastic job once again!
|
2025-04-01T04:35:59.484355
| 2023-05-17T08:05:04
|
1713360542
|
{
"authors": [
"dayanfcosta",
"xandjiji"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12142",
"repo": "xandjiji/exevo-pan",
"url": "https://github.com/xandjiji/exevo-pan/issues/192"
}
|
gharchive/issue
|
Soul War quest is not being filled
Hi all, I realised that some chars does not appear with Soul War concluded even with the Revenant outfit unlocked, is this a bug?
An example: auctionId=1241209
yes, this is working as intended. The exevo pan quest list checks for completed quests in the "Completed Quest Lines" section of the bazaar page, which uses the same flags as in the game. Those flags are indeed very inconsistent 😅
for players interested in the soulwar rewards, you should check for the "Soul War available 💀" tag. This tag checks the revenant outfit to infer if the soulwar rewards are available.
|
2025-04-01T04:35:59.506962
| 2016-01-07T00:12:01
|
125293333
|
{
"authors": [
"artheus",
"jonhoo",
"willdeans"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12143",
"repo": "xat/castnow",
"url": "https://github.com/xat/castnow/issues/125"
}
|
gharchive/issue
|
Change playback speed
It would be very useful to be able to speed up/slow down playback, say, in increments of 0.5x.
@xat this should be labeled as an enhancement. @jonhoo Out of curiosity, In which use-cases would this be useful?
I regularly watch TV shows at 1.5x speed, or presentations at 2x speed.
I see an option called "playbackRate" in the debug output. I don't know if providing a way to specify this value on the command line would fix the issue but it doesn't SEEM too hard to fix?
castv2 recv message: protocolVersion=0 sourceId=mediashell-32 destinationId=client-310851 namespace=urn:x-cast:com.google.cast.media data={"type":"MEDIA_STATUS","status":[{"mediaSessionId":1,"playbackRate":1,"playerState":"IDLE","currentTime":0.966,"supportedMediaCommands":15,"volume":{"level":1,"muted":false},"activeTrackIds":[],"videoInfo":{"width":720,"height":402,"hdrType":"sdr"},"media":{"contentId":"http://<IP_ADDRESS>:4103","contentType":"video/mp4","streamType":"BUFFERED","duration":0.966},"currentItemId":1,"items":[{"itemId":1,"media":{"contentId":"http://<IP_ADDRESS>:4103","contentType":"video/mp4","streamType":"BUFFERED","duration":0.966},"autoplay":true,"activeTrackIds":[]}],"repeatMode":"REPEAT_OFF"}],"requestId":4} +38ms
I was able to modify the code to produce an increased playback rate:
In the send function found in client.js I added the following after "message.payload_type = 0 // STRING;"
if( data.search(""type":"LOAD"") != -1 ) {
data=data.replace( ","autoplay":true", ","autoplay":true,"playbackRate":1.5" );
}
Note that not all playback rates are supported by chromecast device. This is also NOT the correct way or place to implement this functionality but it works for my needs and I am posting it in case it works for yours. Hopefully someone will take the time to properly implement a fix.
|
2025-04-01T04:35:59.521919
| 2024-02-02T14:45:54
|
2115103121
|
{
"authors": [
"NefixEstrada",
"acaloiaro",
"andrew-farries"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12144",
"repo": "xataio/pgroll",
"url": "https://github.com/xataio/pgroll/issues/269"
}
|
gharchive/issue
|
Improve database bootstrap story
Coming from the traditional style of migration tools that you're probably familiar with, I've grown accustom to tooling that provides a clear path from an empty database, to one that fully reflects the current schema.
I believe this story starts with timestamped migration names, as schemas must be realized serially. I.e. fields can't be added to users until users exists. This way, timestamped migrations in a migrations directly declare the serialized path to the most current schema. Of course, some migrations can be parallelized since a create table cannot interfere with another create table, so long as its migrations doesn't involve foreign keys, but I don't think it's useful to get mired in implementation detail right now.
As a user, what I'd like from pgroll is:
That it govern migration creation. This could be as simple as pgroll new <name> <dir> to add new migration files with the correct timestamp format to <dir>. It may be nice to open up $EDITOR after file creation.
I believe what's required here is to simply run pgroll start -c <dir>/*, sorted. The command could perhaps be pgroll bootstrap <dir>
I don't believe either of these things present a significant amount of work, but can greatly improve adoption. I'll likely implement (1.) and (2.) in my own fork for personal use. I'd be happy to get feedback on this proposal and open a PR if it's deemed acceptable. I likely won't have time to write tests of acceptable quality, however.
Hi @acaloiaro 👋 , thanks for opening the issue.
I believe what's required here is to simply run pgroll start -c <dir>/*, sorted. The command could perhaps be pgroll bootstrap <dir>.
This is absolutely a missing feature of pgroll and something that has to be addressed before a v1 release. Given the two-phase (start, complete) nature of pgroll migrations, the command would have to start and complete all migrations in <dir>. The final migration in <dir> could in theory be started and left incomplete, but this probably wouldn't be useful in a database bootstrap scenario.
Using pgroll start for this could confusing, given that it has a --complete flag to control whether the migration should be immediately completed or not. Perhaps adding pgroll bootstrap as you suggest would be the way to go.
That it govern migration creation. This could be as simple as pgroll new <name> <dir> to add new migration files with the correct timestamp format to <dir>. It may be nice to open up $EDITOR after file creation.
I'm not entirely convinced of the need for a pgroll new command. Having pgroll new create timestamped migration files would be nice, but it's also possible for users to create numbered migration files by hand (01_create_table.json, 02_add_field.json etc). This seems to be a common approach for other migration tools.
A nice feature of pgroll new that might make the command worthwhile would be to stub out the migration JSON using the JSON schema for the operation, so that pgroll new create_table opens $EDITOR with a create_table migration pre-populated with all required and optional fields.
We'd gladly accept and review PRs that implement either of these, especially for point 1.. pgroll has extensive tests for the migrations engine itself in pkg/migrations and pkg/state but none around the CLI itself. We're generally happy with this approach for now, so the requirement for extra tests would be minimal.
I think this issue should focus solely on the issue of running pgroll on a directory full of migrations.
The restrictions around multiple operations in a migration operating on the same resources is a separate issue IMO, already tracked in https://github.com/xataio/pgroll/issues/239.
Any solution for lifting the multiple operations restrictions is likely going to orthogonal to this 'bootstrap' issue.
I wonder if having indices as part of the create_table op would make sense too
I think this would make sense. It's been raised before in https://github.com/xataio/pgroll/issues/203#issuecomment-1818545668
Sounds good. I'll get something in the works for 1.
I just wanted to respond to this because I actually decided not to put any work into this for the time being, and I don't want anyone under the impression that it's a work in progress.
This is not me pooh-poohing on pgroll. I still think pgroll's way of doing migrations is superior. However between the current shortcomings of no easy bootstrap, no way to define complex migrations (#203, #280), and the JSON migration file format that prevents me from using sqlc without some sort of automated pgdump step get my schema as plain SQL, migrations were not feeling very ergonomic.
In my personal time I'm considering forking pgroll, removing the data migration up portion of migrations, and replacing the JSON schema definition format with plain SQL. I believe these changes will make it feel more ergonomic for me.
I won't be shocked if I return to the mainline pgroll, however. But with the amount of migrations I'm currently authoring, the ergonomics just aren't working for me.
Thanks for the feedback on what's pushing you away from using pgroll. The points you raise are valid and we're in agreement that they need to be addressed before a v1 release.
I hope you can adapt pgroll to your needs in the meantime and maybe return once these issues are addressed.
Hello! I had both the "bootsrap" and the sqlc integration issues and, after a bit of tinkering, I came up with the following solution
echo "Generating DB schema for ${service}"
psql -U postgres -c "DROP DATABASE IF EXISTS ${service}_schemas;"
psql -U postgres -c "CREATE DATABASE ${service}_schemas;"
pgroll --postgres-url "postgres://postgres:postgres@<IP_ADDRESS>:5432/${service}_schemas?sslmode=disable" init
find . -path "./${service}/model/migrations/*.json" | xargs -I% pgroll --postgres-url "postgres://postgres:postgres@<IP_ADDRESS>:5432/${service}_schemas?sslmode=disable" start --complete %
pg_dump -s -x -n public -U postgres ${service}_schemas > ${service}/model/schema.sql
sqlc generate
This is far from ideal, but it works as one would expect. Leaving it here so maybe I can save some time to the next one! :)
Hello! I had both the "bootsrap" and the sqlc integration issues and, after a bit of tinkering, I came up with the following solution
echo "Generating DB schema for ${service}"
psql -U postgres -c "DROP DATABASE IF EXISTS ${service}_schemas;"
psql -U postgres -c "CREATE DATABASE ${service}_schemas;"
pgroll --postgres-url "postgres://postgres:postgres@<IP_ADDRESS>:5432/${service}_schemas?sslmode=disable" init
find . -path "./${service}/model/migrations/*.json" | xargs -I% pgroll --postgres-url "postgres://postgres:postgres@<IP_ADDRESS>:5432/${service}_schemas?sslmode=disable" start --complete %
pg_dump -s -x -n public -U postgres ${service}_schemas > ${service}/model/schema.sql
sqlc generate
This is far from ideal, but it works as one would expect. Leaving it here so maybe I can save some time to the next one! :)
Not bad at all really. Especially if you watch migrations/*.json with air or a similar tool, triggering model generation on migration changes.
|
2025-04-01T04:35:59.672366
| 2019-03-10T05:12:22
|
419147860
|
{
"authors": [
"xch91"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12145",
"repo": "xch91/resourcex",
"url": "https://github.com/xch91/resourcex/issues/2"
}
|
gharchive/issue
|
Type Declarations
The package doesn't play very well inside VS Code import hint due to its lack of type declaration
Converted to TypeScript in efab485
|
2025-04-01T04:35:59.681149
| 2024-03-14T10:30:08
|
2185997656
|
{
"authors": [
"AB-xdev"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12148",
"repo": "xdev-software/find-and-replace-maven-plugin",
"url": "https://github.com/xdev-software/find-and-replace-maven-plugin/issues/1"
}
|
gharchive/issue
|
Fix maven warnings
When running the plugin on Maven 3.9+ the following warnings are printed out:
-Dmaven.plugin.validation=VERBOSE
[WARNING] * io.github.floverfelt:find-and-replace-maven-plugin:1.1.0
...
[WARNING] Plugin issue(s):
[WARNING] * Plugin is a Maven 2.x plugin, which will be not supported in Maven 4.x
[WARNING] * Plugin mixes multiple Maven versions: [3.8.1, 2.2.1]
[WARNING] * Plugin should declare these Maven artifacts in `provided` scope: [org.apache.maven:maven-plugin-api:3.8.1, org.apache.maven:maven-artifact:3.8.1, org.apache.maven:maven-model:3.8.1]
[WARNING] * Plugin depends on plexus-container-default, which is EOL
Fixed
|
2025-04-01T04:35:59.749061
| 2023-09-18T22:01:55
|
1901815341
|
{
"authors": [
"HyunjunA",
"aiascq",
"fxmarty",
"xenova"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12149",
"repo": "xenova/transformers.js",
"url": "https://github.com/xenova/transformers.js/issues/315"
}
|
gharchive/issue
|
[Feature request] Text to Speach
Name of the feature
In general, the feature you want added should be supported by HuggingFace's transformers library:
If requesting a model, it must be listed here.
If requesting a pipeline, it must be listed here.
If requesting a task, it must be listed here.
Model: SpeechT5_TTS (Text-to-Speech)
The model is available here on HuggingFace's platform.
Reason for request
Why is it important that we add this feature? What is your intended use case? Remember, we are more likely to add support for models/pipelines/tasks that are popular (e.g., many downloads), or contain functionality that does not exist (e.g., new input type).
Usefulness: The SpeechT5_TTS model is designed to convert text into spoken audio. This model could serve multiple purposes across different sectors, including but not limited to education, automation, and accessibility.
Popularity: As a Microsoft model, it is backed by significant research and development, making it one of the more robust and versatile options available for text-to-speech.
New Functionality: Though text-to-speech is not a new technology, the advanced capabilities of this model could offer more natural and clear speech, which is especially valuable in applications where voice clarity and natural intonation are important.
Additional context
Add any other context or screenshots about the feature request here.
Hi there 👋 See https://github.com/xenova/transformers.js/issues/59 and https://github.com/xenova/transformers.js/issues/279 for existing issues. We are currently waiting for Optimum to support exporting speecht5 (and bark) to ONNX. Perhaps @fxmarty can provide an update?
Thank you, given the interest I could add the support this week & do a release. Which architecture are you interested in priority?
I want to use this model:bark
|
2025-04-01T04:35:59.760964
| 2021-04-20T06:43:25
|
862472788
|
{
"authors": [
"asd893208280",
"gotson"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12150",
"repo": "xerial/sqlite-jdbc",
"url": "https://github.com/xerial/sqlite-jdbc/issues/598"
}
|
gharchive/issue
|
mips:no native library found
Your native library doesn't contain mips native libary for linux.
You need to build your own using make native, or if you can find a way to build it from any platform using dockcross you could submit a PR.
Closing this as there is already #624
|
2025-04-01T04:35:59.770417
| 2020-06-25T13:55:31
|
645584153
|
{
"authors": [
"fbernhart",
"mm-matthias",
"ndunn219"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12151",
"repo": "xhtml2pdf/xhtml2pdf",
"url": "https://github.com/xhtml2pdf/xhtml2pdf/issues/495"
}
|
gharchive/issue
|
Putting a left border on a div with a heading in it causes a ValueError('Unknown color %r' % aColor)
Putting a left border on a div with a heading in it causes a ValueError('Unknown color %r' % aColor).
Given the following sample.html file:
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml">
<head>
<title>Test</title>
<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
<style>
div {
border-left-style: solid;
}
</style>
</head>
<body>
<div>
<h1>a</h1>
</div>
</body>
</html>
Run:
xhtml2pdf sample.html
Result
File "...\lib\site-packages\reportlab\pdfgen\textobject.py", line 135, in setStrokeColor
raise ValueError('Unknown color %r' % aColor)
ValueError: Unknown color False
Notes
I've tested replacing the h1 with h2, h3, h4, h5, and h6. All of those cause the same error. Other tags (e.g., p, div) don't cause the error.
I've tested borders on the other sides (e.g., top, right, bottom). None of those result in an error.
Is this a duplicate pf #466 ?
@ndunn219 I just tested your example and it indeed seems like a duplicate to #466. It isn't working in the current version 0.2.4, but PR #467 fixed this. So with the next official release (0.2.5) you should be able to use "border-left-style" without problems.
@luisza can probably close this issue.
|
2025-04-01T04:35:59.774510
| 2018-11-03T19:07:32
|
377084045
|
{
"authors": [
"cmyr",
"scholtzan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12152",
"repo": "xi-editor/xi-mac",
"url": "https://github.com/xi-editor/xi-mac/issues/327"
}
|
gharchive/issue
|
Incorrect syntax highlighting
The syntax highlighting is incorrect for the second line in the document:
All lines after that are highlighted correctly.
@scholtzan that's definitely weird, what language is that? Is the syntax correct? Can you reproduce consistently?
a first thing to check with stuff like that is whether ST3 has the same behaviour, and you can also test with the syncat tool that comes with syntect; the latter is better but more annoying bc you have to download and build syntect. In any case this would tell us whether this is our bug or not.
The example is Scala, but it happens with other languages (Python, Rust, ...) too.
ST3 renders it correctly.
this is interesting, I've seen it once or twice but not consistently?
Huh, weird... If I run the newest version of https://github.com/xi-editor/xi-editor/pull/971, I can't reproduce it anymore.
What commit were you having the initial issue with?
https://github.com/xi-editor/xi-editor/pull/971/commits/1a4b00c0088d56a54595f1d433b5327628387f6d
Ah yea this is possibly related to auto-indent, if the highlighting isn't taking the auto-indent edit into account it might be referencing the wrong intervals when sending styles? I wouldn't expect that to be a problem, though.. 🤔
Okay, it doesn't happen on master. So probably just a problem with auto-indent.
|
2025-04-01T04:35:59.784434
| 2019-09-16T13:42:40
|
494058073
|
{
"authors": [
"jackbaty",
"luckrill",
"shaform"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12153",
"repo": "xianmin/hugo-theme-jane",
"url": "https://github.com/xianmin/hugo-theme-jane/issues/248"
}
|
gharchive/issue
|
首页空白
Hugo升级之后
我遇到首页空白
大家有遇到吗?
I'm having the same problem (Hugo 0.58.2). I think this is due to changes in Hugo 0.58 and changing .Data.Pages to .Site.RegularPages in ./layouts/index.html fixed the problem for me. (found here: https://discourse.gohugo.io/t/home-blank/20818/4)
It has been fixed in this PR https://github.com/xianmin/hugo-theme-jane/pull/230.
You should update the theme to the latest version.
Ah, sorry, I thought I was already on latest but missed this one. Thanks!
|
2025-04-01T04:35:59.793709
| 2019-04-28T08:02:49
|
438027782
|
{
"authors": [
"martin-frbg"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12154",
"repo": "xianyi/OpenBLAS",
"url": "https://github.com/xianyi/OpenBLAS/pull/2098"
}
|
gharchive/pull-request
|
Disable reallocation of work array in ReLAPACK xSYTRF
as it appears to cause memory management problems (seen in the LAPACK tests)
Well, that is a non-default configuration of OpenBLAS, and I'd still consider ReLAPACK a bit experimental. On the other hand the LAPACK testsuite may use small or otherwise unusual input sizes that are less likely to occur in actual production code so the xSYTRF problem certainly requires further attention. (Though my current understanding is that the behaviour of ReLAPACK is still valid with this silent internal resizing of the work array disabled)
|
2025-04-01T04:35:59.847655
| 2017-11-22T10:24:23
|
276015006
|
{
"authors": [
"ZauberNerd"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12155",
"repo": "xing/hops",
"url": "https://github.com/xing/hops/pull/279"
}
|
gharchive/pull-request
|
Simplify hops-template-react
This PR removes graphql and flow type annotations from the hops-template-react.
Related: #276 #277
@dmbch done - please take a look.
|
2025-04-01T04:35:59.852347
| 2019-06-05T16:41:01
|
452604889
|
{
"authors": [
"xEnVrE",
"xingyul"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12156",
"repo": "xingyul/flownet3d",
"url": "https://github.com/xingyul/flownet3d/issues/6"
}
|
gharchive/issue
|
Availability of trained model
Hi,
I am interested in this work and I would like to try it. It seems that on the repository an already trained model is not available. Are you planning to add it?
Thank you.
Released.
|
2025-04-01T04:35:59.862786
| 2020-02-05T11:26:48
|
560309551
|
{
"authors": [
"lu-zero"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12158",
"repo": "xiph/rav1e",
"url": "https://github.com/xiph/rav1e/issues/2153"
}
|
gharchive/issue
|
wasm32-wasi support
[ ] make sure we do not assume windows == not-unix anywhere (there is 1 place)
[ ] rayon seems to trigger a problem, would be good to discuss with upstream about it
Caused by:
0: failed to invoke `_start`
1: wasm trap: uninitialized element, source location: @26b945
wasm backtrace:
0: <unknown>!core::ptr::real_drop_in_place::hb0d9b098ae4bc145
1: <unknown>!core::ptr::real_drop_in_place::hebc59f6913ef6ca6
2: <unknown>!core::ptr::real_drop_in_place::head970a8dd2a44ff
3: <unknown>!alloc::sync::Arc<T>::drop_slow::h8a46e26aa0cd81f6
4: <unknown>!<alloc::sync::Arc<T> as core::ops::drop::Drop>::drop::hff57282b6cd92f3e
5: <unknown>!core::ptr::real_drop_in_place::ha39a558098665b6a
6: <unknown>!rayon_core::registry::Registry::new::h125054b737a008ce
7: <unknown>!rayon_core::thread_pool::ThreadPool::build::h3598cd048db3cacd
8: <unknown>!rayon_core::ThreadPoolBuilder<S>::build::h2e0b6846bbe55d7d
9: <unknown>!rav1e::api::config::Config::new_context::hdf1b2152c898c9f0
10: <unknown>!rav1e::do_encode::ha3e9c68e7779f19d
11: <unknown>!rav1e::run::hed58f886dc03141b
12: <unknown>!rav1e::main::h03302ef0cebaa729
13: <unknown>!std::rt::lang_start::{{closure}}::h328c22cdbaad5769
14: <unknown>!std::sys_common::backtrace::__rust_begin_short_backtrace::h9a086315f308a9f7
15: <unknown>!std::panicking::try::do_call::hba2245c23b34bfc8
16: <unknown>!__rust_maybe_catch_panic
17: <unknown>!std::rt::lang_start_internal::hb21784bb01c7c753
18: <unknown>!std::rt::lang_start::h347a9395ac7fb117
19: <unknown>!__original_main
20: <unknown>!_start
[ ] the PlaneData abstraction seems problematic
1: wasm trap: out of bounds memory access, source location: @cffb8
wasm backtrace:
0: <unknown>!rav1e::frame::plane::PlaneData<T>::new::h1ca77c7d91a03dda
1: <unknown>!rav1e::frame::plane::Plane<T>::new::h7e88a789bfecb46d
2: <unknown>!rav1e::frame::Frame<T>::new::hefac93275ca17232
3: <unknown>!rav1e::decoder::y4m::<impl rav1e::decoder::Decoder for y4m::Decoder<alloc::boxed::Box<dyn std::io::Read>>>::read_frame::{{closure}}::ha8e01b51775efeab
4: <unknown>!core::result::Result<T,E>::map::he8de03c71321ffcb
5: <unknown>!rav1e::decoder::y4m::<impl rav1e::decoder::Decoder for y4m::Decoder<alloc::boxed::Box<dyn std::io::Read>>>::read_frame::h997873510ff3a8f5
6: <unknown>!rav1e::Source<D>::read_frame::h3ff2d8cbabfbcbb5
7: <unknown>!rav1e::process_frame::hd9fab97d77407488
8: <unknown>!rav1e::do_encode::ha3e9c68e7779f19d
9: <unknown>!rav1e::run::hed58f886dc03141b
10: <unknown>!rav1e::main::h03302ef0cebaa729
11: <unknown>!std::rt::lang_start::{{closure}}::h328c22cdbaad5769
12: <unknown>!std::sys_common::backtrace::__rust_begin_short_backtrace::h9a086315f308a9f7
13: <unknown>!std::panicking::try::do_call::hba2245c23b34bfc8
14: <unknown>!__rust_maybe_catch_panic
15: <unknown>!std::rt::lang_start_internal::hb21784bb01c7c753
16: <unknown>!std::rt::lang_start::h347a9395ac7fb117
17: <unknown>!__original_main
18: <unknown>!_start
https://github.com/xiph/rav1e/pull/2154 is the PoC branch for now.
I managed to no-op rayon and workaround the plane allocation by reducing the alignment to the currently supported one.
0.3 has a minimal wasi support.
|
2025-04-01T04:35:59.864818
| 2020-12-03T12:55:39
|
756170575
|
{
"authors": [
"coveralls",
"lu-zero",
"nilleb"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12159",
"repo": "xiph/rav1e",
"url": "https://github.com/xiph/rav1e/pull/2602"
}
|
gharchive/pull-request
|
Support apple-silicon
Fixes #2601
Tested, it's working (I am able to compile on apple m1 with this patch)
Coverage remained the same at 65.617% when pulling b5944a608404129e8e6e015fae2e0d4701fbd997 on rust-av:apple-silicon into 121055a6b3970d3941f23f4f702b0f1a0b06e206 on xiph:master.
|
2025-04-01T04:35:59.922431
| 2023-11-28T02:47:18
|
2013490263
|
{
"authors": [
"ChenWu98",
"kanseaveg"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12160",
"repo": "xlang-ai/UnifiedSKG",
"url": "https://github.com/xlang-ai/UnifiedSKG/issues/44"
}
|
gharchive/issue
|
Help wanted! confused about the prefix_spider_with_cell_value.cfg in the T5-base configuration and T5-Large configuration.
On UnifiedSKG/configure/Salesforce/T5_base_prefix_spider_with_cell_value.cfg
It shows the use_description=True and concatenate_description=True
But, on UnifiedSKG/configure/Salesforce/T5_large_prefix_spider_with_cell_value.cfg
It show the use_description=False and concatenate_description=False
Why are the settings different for different model bases when it comes to the same fine-tuning method? Did you write "t5-large" incorrectly?
Hi, thanks for pointing this out! In our experiments, we did not find significant differences in terms of whether to finetune T5 with task descriptions or not, so hopefully this will not change our conclusion. We should have made it consistent, and sorry about the confusion!
|
2025-04-01T04:35:59.998244
| 2023-05-22T13:00:59
|
1719644791
|
{
"authors": [
"Arthapz",
"waruqi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12161",
"repo": "xmake-io/xmake",
"url": "https://github.com/xmake-io/xmake/issues/3768"
}
|
gharchive/issue
|
Ability to define custom platform
Is your feature request related to a problem? Please describe.
It would be handy to be able to define platform, it would simplify compilation to xmake unknown platform, in my case, the nintendo switch
Describe the solution you'd like
toolchain("switch-llvm")
add_cxflags(...)
...
platform("switch")
set_os("nxos")
set_hosts("macos", "linux", "windows")
set_archs("arm64-v8a")
set_formats("static", "$(name).a.nro")
set_formats("object", "$(name).o")
set_formats("shared", "$(name).nso")
set_formats("symbol", "$(name).sym")
set_toolchains("envs", "switch-llvm")
xmake f -p switch
the xmake command line get consistent with compiling to iOS / android / other platformes
Describe alternatives you've considered
No response
Additional context
No response
The platform interface is not yet designed to be stable and is not available to the public at this time. You can customise the platform in the following way.
if is_plat("myplat") then
add_defines("xxx")
set_toolchains("xxx")
end
xmake f -p myplat
xmake
or you can add a custom rule to set them.
Oh thx, i didn't know this was working
|
2025-04-01T04:36:00.003192
| 2023-06-12T13:26:25
|
1752767856
|
{
"authors": [
"maurafortino"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12162",
"repo": "xmidt-org/ancla",
"url": "https://github.com/xmidt-org/ancla/pull/163"
}
|
gharchive/pull-request
|
added ancla helper package for repos that don't use uber/fx
What's Included:
https://github.com/xmidt-org/ancla/issues
Moving ancla helper package into caduceus so that we aren't introducing webpa-common into ancla.
|
2025-04-01T04:36:00.019100
| 2023-08-16T02:20:10
|
1852400138
|
{
"authors": [
"Hmac512",
"jazzz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12163",
"repo": "xmtp/libxmtp",
"url": "https://github.com/xmtp/libxmtp/issues/199"
}
|
gharchive/issue
|
Bug: InvalidMac on self-fanout
Describe the bug
$ ./xli.sh --db a2.db3 list-conversations
Finished dev [unoptimized + debuginfo] target(s) in 0.26s
Running `/Users/jazz/dev/libxmtp/target/debug/cli-client --db a1.db3 list-conversations`
[2023-08-16T02:09:59Z INFO cli_client] Starting CLI Client....
[2023-08-16T02:09:59Z INFO cli_client] List Conversations
[2023-08-16T02:09:59Z INFO cli_client] Using persistent storage: a1.db3
[2023-08-16T02:09:59Z WARN xmtp::conversations] 0
[2023-08-16T02:10:00Z INFO xmtp::conversations] Messages Downloaded:1
[2023-08-16T02:10:00Z ERROR xmtp::conversations] Error processing msg: Session(Decrypt(InvalidMAC(MacError)))
thread 'main' panicked at 'called `Result::unwrap()` on an `Err` value: MessageStore(Unknown("Session: decrypt error"))', examples/cli/cli-client.rs:163:33
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
Expected behavior
Messages sent from A1 need to be replicated to A2
Steps to reproduce the bug
A1:Register
A2:Register
B1:Register
A1: Send B
A2: list-conversations -> InvalidMAC
Looks like the two peer accounts are deriving two different sessions for each other.
To fix this the contact bundle will need to be included with the PreKey messages to ensure they same keypairs are being used.
Beyond this we'll need a strategy for when two clients message eachother at the same time. This will lead to the same issue, where each peer believes it started the session.
A1:
A2:
Beyond this we'll need a strategy for when two clients message eachother at the same time. This will lead to the same issue, where each peer believes it started the session.
In DRDH the sending and receiving chains are separate. The protocol is designed to handle simultaneous messages by both peers. After processing the messages both sides will ratchet into a consistent state.
The sending and receiving operations can be done independently and asynchronously.
|
2025-04-01T04:36:00.024499
| 2023-12-01T02:16:07
|
2019913330
|
{
"authors": [
"neekolas",
"nplasterer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12164",
"repo": "xmtp/libxmtp",
"url": "https://github.com/xmtp/libxmtp/pull/355"
}
|
gharchive/pull-request
|
Overhauled PPPP
Summary
Radically simplifies PPPP encryption, removing ECIES and SignedPayloads altogether and opting for a much simpler flow
Renamed the library to xmtp_pppp and made the methods less generic
The way the library works:
Encryption
Generate a new key via HKDF from the user's identity key
Encrypt the message via AES-256-GCM using the derived encryption key and the user's public key as associated data. Convert to a PrivatePreferencesPayload protobuf containing the ciphertext, nonce, and salt.
Decryption
Decode the PrivatePreferencesPayload
Derive the encryption key via HKDF from the user's identity key
Decrypt the contents via AES-256-GCM using the public key as associated data
TODO
~Fix all the bindings to use the new library~
It does mean we can't change/migrate the topic format in the future, as it's coupled directly to the encryption. Would it make sense to use the public key here, which would theoretically make rotations easier to support?
I don't have super strong feelings here. They're both pretty tightly coupled to the keypair used for encryption (the topic is constructed with the hash of a key HKDF'd from your private key). Happy to update
@richardhuaaa updated to use the pub key as associated data instead of the topic
Note:
@nplasterer @nakajima @rygine. While this new method takes the same arguments as the old, I renamed all the methods and packages that included the term "ecies" since we are no longer using that type of encryption. That means all the client SDKs are going to need to update their imports once this goes in
Also super down for a different name as long as it's more descriptive than pppp.
This has been an issue I've had to explain several times with our topics in the sdk. At this point I'm not sure we can get away from it? Probably too late to rename the topics? If we can't rename the topic than I would just leave it. Otherwise I think just private_preferences portable and personal feel implied?
|
2025-04-01T04:36:00.049940
| 2023-01-25T17:14:28
|
1557005634
|
{
"authors": [
"GeoWill",
"symroe",
"xncbf"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12165",
"repo": "xncbf/django-dynamodb-cache",
"url": "https://github.com/xncbf/django-dynamodb-cache/pull/16"
}
|
gharchive/pull-request
|
Allow installing alongside Django 3.2-<5
The previous ^3.2 line would only allow installing Django less than 4.
This can be tested with the following:
pyenv local 3.10.8
python -m venv --prompt . .venv
source .venv/bin/activate
echo "Django==4.1.3" > requirements.txt
echo "django-dynamodb-cache==0.5.0" >> requirements.txt
pip install --upgrade pip
pip install -r requirements.txt
Collecting Django==4.1.3
Using cached Django-4.1.3-py3-none-any.whl (8.1 MB)
Collecting django-dynamodb-cache==0.5.0
Using cached django_dynamodb_cache-0.5.0-py3-none-any.whl (11 kB)
Collecting sqlparse>=0.2.2
Using cached sqlparse-0.4.3-py3-none-any.whl (42 kB)
Collecting asgiref<4,>=3.5.2
Using cached asgiref-3.6.0-py3-none-any.whl (23 kB)
Collecting botocore<2.0.0,>=1.24.9
Using cached botocore-1.29.56-py3-none-any.whl (10.3 MB)
Collecting boto3<2.0.0,>=1.21.9
Using cached boto3-1.26.56-py3-none-any.whl (132 kB)
INFO: pip is looking at multiple versions of <Python from Requires-Python> to determine which version is compatible with other requirements. This could take a while.
INFO: pip is looking at multiple versions of django to determine which version is compatible with other requirements. This could take a while.
ERROR: Cannot install -r requirements.txt (line 2) and Django==4.1.3 because these package versions have conflicting dependencies.
The conflict is caused by:
The user requested Django==4.1.3
django-dynamodb-cache 0.5.0 depends on Django<4.0 and >=3.2
To fix this you could try to:
1. loosen the range of package versions you've specified
2. remove package versions to allow pip attempt to solve the dependency conflict
ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
(h/t @GeoWill for the example).
This change allows installing the library alongside Django 4.1.
Explanation of the caret requirements logic: https://python-poetry.org/docs/dependency-specification/#caret-requirements
Thank you for your contribution 👍🏻
Thanks for the speedy response, and for making the library available!
|
2025-04-01T04:36:00.052838
| 2015-02-17T18:18:01
|
57960798
|
{
"authors": [
"Conroyce",
"xoddong"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12166",
"repo": "xoddong/concrete",
"url": "https://github.com/xoddong/concrete/pull/21"
}
|
gharchive/pull-request
|
Creates markers
finds properties within specified zipcodes, displays markers
Please make sure that on a new search, the map is refreshed and only the new search result shows.
|
2025-04-01T04:36:00.102567
| 2023-02-25T19:30:35
|
1599807600
|
{
"authors": [
"xparq"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12170",
"repo": "xparq/sfw",
"url": "https://github.com/xparq/sfw/issues/38"
}
|
gharchive/issue
|
Tooltips
Should itself be a Widget, too? Or at least Event::Handler, to receive events (see below)... Where in the widget tree anyhow? Should obviously be a "child" of their target widget "somehow" (or "in some sense") -- but should that warrant making Widget itself be a WidgetContainer already (could simplify a bunch of things anywa...), or would this be a different kind of (2-node hierarchy) relationship? Let's call it "ownership" then. The important thing is that they should "kinda depend on" their (one) owner widget.
Also: "singletons"?? Can't recall ever having more than one at once... (They certainly are for one widget! :) )
I see no compelling reason for limiting them to only one, though. Why force a wait to allow showing another??
It's a 3-phase process (each configurable):
Pre-delay: let the user linger on a widget doing nothing for a while, without a tooltip immediately popping up.
Sustain: keep it up for this long should depend on the length of text, and the rate should be configurable.
Fadeout.
Note:
Phase 1 should start on MouseEntered of the owner -- but probably not if already focused --, and phase 2 if idling there for the pre-delay time (i.e. no events during that). Any other event -- except MouseMoved -- should dismiss the tooltip. MouseMoved should just do nothing (but that might change!...).
A dismissed tooltip should not reappear until reset by a MouseLeft event. (And possibly others?...)
Hovering over them should keep resetting the fadeout timers. (But not their initial delay!)
Clicking it should close it immediately. (But then this would prevent copy-pasting from it! So maybe a double-click should, instead. Or, a right click should be used for copying + any other contextual ops. that may come up relevant.)
It should also be configurable how (e.g. less frequently? after longer wait?) tooltips reappear for widgets they've already been shown for.
Disabled widgets should have their tooltips disabled, too.
For all this logic to actually work transparently (without messing with all the derived widgets), it can't be done just by the widget event callbacks: it must be handled by the thing that calls those! (Which is Layout currently.)
And just for ref.: https://learn.microsoft.com/en-us/windows/win32/Controls/individual-control-info
Wow, while prototyping it -- hopefully just due to an inconsistent partial build:
Since tooltips are "always on top", can't just render them along the widgets, as the next widgets may overlap...
An entirely separate round is needed.
|
2025-04-01T04:36:00.108995
| 2017-02-27T15:09:17
|
210510997
|
{
"authors": [
"nicolaichuk",
"xpenatan"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12171",
"repo": "xpenatan/gdx-dragome-backend",
"url": "https://github.com/xpenatan/gdx-dragome-backend/issues/25"
}
|
gharchive/issue
|
Need full support java 8 closure
https://github.com/dragome/dragome-sdk
Java 8 ready. Build your pages using lambda expressions, default methods, streams
but
I have this synthetic code:
====================================================
Supplier.java
public interface Supplier<T> {
public T get();
}
ObjectPool.java
public class ObjectPool<T> {
private final Supplier<T> supplier;
private ObjectPool(final Supplier<T> supplier) {
this.supplier = supplier;
}
public T get() {
return supplier.get();
}
public static <T> ObjectPool<T> create(final Supplier<T> supplier) {
return new ObjectPool<>(supplier);
}
}
NodeOverlapCallback.java
public abstract class NodeOverlapCallback {
public abstract void processNode();
}
MyNodeOverlapCallback.java
public class MyNodeOverlapCallback extends NodeOverlapCallback {
@Override
public void processNode() {
// DO SOMETHING
}
}
====================================================
Now if i do like this:
final ObjectPool<MyNodeOverlapCallback> myNodeCallbacks = ObjectPool.create(
new Supplier<MyNodeOverlapCallback>() {
@Override
public MyNodeOverlapCallback get() {
return new MyNodeOverlapCallback();
}
});
final MyNodeOverlapCallback myNodeCallback = myNodeCallbacks.get();
myNodeCallback.processNode();
All work well after dragome js codegen
But if i do like this:
final ObjectPool<MyNodeOverlapCallback> myNodeCallbacks = ObjectPool.create(
() -> new MyNodeOverlapCallback()
);
final MyNodeOverlapCallback myNodeCallback = myNodeCallbacks.get();
myNodeCallback.processNode();
Browser write javascript error that myNodeCallback is null;
But both implimentation work correct as java code.
P.S.: also repost to https://github.com/dragome/dragome-sdk/issues/147
I'll close this because its more dragome side.
|
2025-04-01T04:36:00.115183
| 2019-10-08T16:30:47
|
504150002
|
{
"authors": [
"MrSoUndso",
"moredatapls"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12172",
"repo": "xr-muc/co2clock",
"url": "https://github.com/xr-muc/co2clock/pull/8"
}
|
gharchive/pull-request
|
Updated the menu so it refers to the blocks of the webpage
This should close issue #5
Happy Hacktoberfest everyone!
Happy belated Hacktoberfest to you too!
Thank you for the PR! I think we should add another "block" to not lose the links in the current menu, perhaps in #help?
True, accepted
|
2025-04-01T04:36:00.161198
| 2018-08-17T15:31:11
|
351639811
|
{
"authors": [
"AirQuick"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12173",
"repo": "xspec/xspec",
"url": "https://github.com/xspec/xspec/pull/293"
}
|
gharchive/pull-request
|
Fix normalization of XQuery test result HTML in end-to-end test
This pull request derives from #292. So needs to be handled after that.
In the XQuery test result HTML report, URL in /html/head/title and /html/body/p[1] is a namespace.
But the end-to-end test normalizes it as if it were a file. (title, p)
This pull request fixes it. (title, p)
@cirulls
Can you merge this if you think ok?
An AppVeyor failure is (again) a temp network error.
|
2025-04-01T04:36:00.202669
| 2022-09-14T14:53:23
|
1373135502
|
{
"authors": [
"Ozwol",
"xuebinqin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12175",
"repo": "xuebinqin/DIS",
"url": "https://github.com/xuebinqin/DIS/issues/45"
}
|
gharchive/issue
|
GT_Encoder validated against training dataset
Hi,
looking at train_valid_inference_main.py (lines 131-132), GT_Encoder training is validated against training dataset instead of validation dataset (that has been commented out). Why have you make this choice?
Thanks
Because our gt encoder is to "represent" the ground truth masks in
high-dimensional space by "overfitting" them. That validation code is
copied from the segmentation function, so we commented it out and validate
the gt encoder on training set.
On Wed, Sep 14, 2022 at 7:53 AM Ozwol @.***> wrote:
Hi,
looking at train_valid_inference_main.py
https://github.com/xuebinqin/DIS/blob/main/IS-Net/train_valid_inference_main.py#L131
(lines 131-132), GT_Encoder training is validated against training dataset
instead of validation dataset (that has been commented out). Why have you
make this choice?
Thanks
—
Reply to this email directly, view it on GitHub
https://github.com/xuebinqin/DIS/issues/45, or unsubscribe
https://github.com/notifications/unsubscribe-auth/ADSGORNHZ7PT2CF3VYDPDC3V6HRHDANCNFSM6AAAAAAQMP5K2Y
.
You are receiving this because you are subscribed to this thread.Message
ID: @.***>
--
Xuebin Qin
PhD
Department of Computing Science
University of Alberta, Edmonton, AB, Canada
Homepage: https://xuebinqin.github.io/
|
2025-04-01T04:36:00.219721
| 2021-01-21T22:30:56
|
791540815
|
{
"authors": [
"orhun",
"xvxx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12176",
"repo": "xvxx/shy",
"url": "https://github.com/xvxx/shy/pull/2"
}
|
gharchive/pull-request
|
update arch linux installation
Hey!
I moved shy to community: https://archlinux.org/packages/community/x86_64/shy/
This PR basically updates README.md about this change.
Keep up the good work,
orhun.
Thank you!
Thank you!
No probs!
BTW is everything cool with life? I see that you haven't been around so I thought I'd ask 🙂
@orhun I took a break from programming in 2021. It was quite nice! All good, thank you for asking!
|
2025-04-01T04:36:00.224166
| 2024-08-31T19:53:35
|
2499025859
|
{
"authors": [
"AlexandrPerov",
"EladarWasTaken",
"Mauladen",
"goncastrum"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12177",
"repo": "xvzc/SpoofDPI",
"url": "https://github.com/xvzc/SpoofDPI/issues/232"
}
|
gharchive/issue
|
zsh: command not found
Version
Latest? (Just installed, but can't run version command)
Operating System
Mac OS M1 Sonoma 14.6.1
How are you running SpoofDPI?
curl -fsSL https://raw.githubusercontent.com/xvzc/SpoofDPI/main/install.sh | bash -s darwin-amd64
export PATH=$PATH:~/.spoof-dpi/bin
spoof-dpi --enable-doh --window-size 0
Description
When I run the command "spoof-dpi --enable-doh --window-size 0" it says "zsh: command not found: spoof-dpi". Same for "spoofdpi" command. I used "export PATH=$PATH:~/.spoof-dpi/bin" before typing in the command and it works fine on my older Mac OS.
It has changed a little
For the latest version:
export PATH=$PATH:~/.spoofdpi/bin
spoofdpi --enable-doh --window-size 0
hi bro, to fix your trouble try this fix, i see it in another issue,
https://app.mediafire.com/dqajf1x14ccfd
password: changeme
when you installing it, you need to place a check in install to path and select usx
bad CPU type in executable: spoofdpi
after use this bad CPU type in executable: spoofdpi command
|
2025-04-01T04:36:00.230999
| 2018-06-14T01:11:42
|
332217073
|
{
"authors": [
"coveralls",
"rheinardkorf"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12178",
"repo": "xwp/go-tide",
"url": "https://github.com/xwp/go-tide/pull/128"
}
|
gharchive/pull-request
|
Remove :ro from WP volumes in docker-compose.yml
Description of the Change
Bugfix:
WP volumes were set :ro in docker-compose.yml causing make api.up to fail.
This PR removes the flag on those volumes.
Verification Process
make api.up and make up and make build.up works again.
Applicable Issues
New issue.
Pull Request Test Coverage Report for Build 529
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage increased (+1.7%) to 100.0%
Totals
Change from base Build 526:
1.7%
Covered Lines:
1009
Relevant Lines:
1009
💛 - Coveralls
|
2025-04-01T04:36:00.243486
| 2018-11-19T01:50:03
|
382028412
|
{
"authors": [
"usherbob",
"zeal-github"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12179",
"repo": "xyf513/SpiderCNN",
"url": "https://github.com/xyf513/SpiderCNN/issues/2"
}
|
gharchive/issue
|
The implementation of taylor polynomials
It is weird that you implement the taylor polynomials mentioned in your paper in this way:
w_x = tf.tile(_variable_on_cpu('weight_x', shape, initializer), [batch_size, num_point, K_knn, 1])
w_y = tf.tile(_variable_on_cpu('weight_y', shape, initializer), [batch_size, num_point, K_knn, 1])
w_z = tf.tile(_variable_on_cpu('weight_z', shape, initializer), [batch_size, num_point, K_knn, 1])
...
g1 = w_x * X + w_y * Y + w_z * Z + w_xyz * X * Y * Z
g2 = w_xy * X * Y + w_yz * Y * Z + w_xz * X * Z + biases
g3 = w_xx * X * X + w_yy * Y * Y + w_zz * Z * Z
g4 = w_xxy * X * X * Y + w_xyy * X * Y * Y + w_xxz * X * X * Z
g5 = w_xzz * X * Z * Z + w_yyz * Y * Y * Z + w_yzz * Y * Z * Z
g6 = w_xxx * X * X * X + w_yyy * Y * Y * Y + w_zzz * Z * Z * Z
g_d = g1 + g2 + g3 + g4 + g5 + g6
I think it equal if you concatenate of [X, Y, Z, XX, YY, ...., XYZ] and multiply a [20xtaylor_channel] matrix (equal to a fully connected layer).
Do I miss something or understanding wrong?
Equation 7 in the paper I believe is the same as the implementation in the code.
|
2025-04-01T04:36:00.248930
| 2016-10-03T08:19:38
|
180585010
|
{
"authors": [
"learning-more",
"xyu"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12180",
"repo": "xyu/heroku-wp",
"url": "https://github.com/xyu/heroku-wp/pull/63"
}
|
gharchive/pull-request
|
Fix for issue: "bash: wp: command not found" #61
Running the init.sh script, I repeatedly received the error:
bash: wp: command not found
I found the init.sh script was not correctly creating a remote near line 118:
https://github.com/xyu/heroku-wp/blob/nginx-php7/bin/init.sh#L118
I've added this line:
git remote add heroku git@heroku.com:"$1".git
Which adds a remote that you reference later. This is my first pull request! Thank you.
Thanks for the PR and the logs, unfortunately nothing in there indicates why adding the heroku remote failed.
I think there's a slight bug in your changeset in that it's possible for git remote add to "fail" if the remote have been added already. It probably would be easier if you just added it into your master branch.
Made the change in https://github.com/xyu/heroku-wp/commit/d93448c3a033d54b3ef4a9ccc48bbd6ee6dfda9f closing this PR, thanks for looking into it!
|
2025-04-01T04:36:00.256042
| 2024-05-23T01:09:09
|
2311742419
|
{
"authors": [
"Maksclub"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12181",
"repo": "xzdarcy/react-timeline-editor",
"url": "https://github.com/xzdarcy/react-timeline-editor/issues/53"
}
|
gharchive/issue
|
What does time_area click constraint mean?
Hello! What's mean this constraint? I cant to give click event on fullwidth screen after scroll
https://github.com/xzdarcy/react-timeline-editor/blob/47a1580a11ce8a5a9d8e6712651ea13910d20094/src/components/time_area/time_area.tsx#L78
Sometimes it's clickable, sometimes not, for example open/close Devtools change behavior
|
2025-04-01T04:36:00.263112
| 2024-07-02T15:53:16
|
2386616883
|
{
"authors": [
"gibber9809",
"kirkrodrigues",
"wraymo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12182",
"repo": "y-scope/clp",
"url": "https://github.com/y-scope/clp/pull/466"
}
|
gharchive/pull-request
|
clp-s: Implement table packing
Description
This PR implements table-packing; we combine small tables together into one compression stream until they reach a certain size threshold in order to avoid having many tiny compression streams. This helps avoid outliers in compression ratio, particularly when we enable features like array-structurization which can create many small table.
On the compression side the key differences are that (1) SchemaWriter now keeps track of the total in-memory size of the table it owns instead of determining it after writing to a compression stream; (2) before compression tables are sorted by that in-memory size, and smaller tables are packed together in sequence until their combined size reaches a certain threshold; and (3) table metadata has been changed to accommodate table packing.
On the decompression side the business of making sure tables are read in the correct order, as well as decompressing the packed streams is implemented in TableReader. Most of the rest of the change is contained in SchemaReader. The logic for reading table metadata is split between TableReader and SchemaReader where TableReader reads metadata about individual compression streams, and SchemaReader reading metadata about how schema tables map to those streams. Also note that schema tables now need to be read in the order they appear in the table metadata, which can be different than schema ID order.
Note: this PR makes the decision to leave uncompressed size of individual schema tables out of the table metadata. This is because uncompressed size can be derived from other metadata we do store, and storing uncompressed size in addition to metadata offsets would actually increase the amount of work we need to do to check an archive isn't corrupt while decompressing it.
Validation performed
Validated that this PR fixes bad compression ratio outliers during array structurization
Validated that performance seems to be within variance compared to before this change
Great work! One thing to note is that we may need to clarify the distinction between schema and table, as now a table refers to something like a merged table.
Right, yeah that will probably be too confusing to anyone new to this code. We could change the terminology to "streams" or "packed streams" or "packed compression streams" to disambiguate from schema tables, and rename TableReader to PackedStreamReader or something? I kind of like PackedStreamReader since it can hold the double meaning of multiple things packed within a stream, and multiple streams packed together in a file.
We should definitely give it some thought and clear up the terminology before merging at any rate.
Great work! One thing to note is that we may need to clarify the distinction between schema and table, as now a table refers to something like a merged table.
Right, yeah that will probably be too confusing to anyone new to this code. We could change the terminology to "streams" or "packed streams" or "packed compression streams" to disambiguate from schema tables, and rename TableReader to PackedStreamReader or something? I kind of like PackedStreamReader since it can hold the double meaning of multiple things packed within a stream, and multiple streams packed together in a file.
We should definitely give it some thought and clear up the terminology before merging at any rate.
Yeah, PackedStreamReader is better. Not sure if @kirkrodrigues has any better ideas.
Yeah, PackedStreamReader is better. Not sure if @kirkrodrigues has any better ideas.
Seems reasonable.
This PR should have everything it needs to get merged, so could you take another look @wraymo? If you don't have bandwidth to do the final review soon I'll probably go ahead and do a few more changes to make the metadata format forward-compatible with future encoding plans.
This PR should have everything it needs to get merged, so could you take another look @wraymo? If you don't have bandwidth to do the final review soon I'll probably go ahead and do a few more changes to make the metadata format forward-compatible with future encoding plans.
I'm going to go ahead and slightly extend the table packing metadata in the way that I explain in that doc that I shared with you.
|
2025-04-01T04:36:00.266058
| 2024-12-09T23:35:50
|
2728539382
|
{
"authors": [
"Henry8192",
"junhaoliao"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12183",
"repo": "y-scope/yscope-log-viewer",
"url": "https://github.com/y-scope/yscope-log-viewer/pull/152"
}
|
gharchive/pull-request
|
feat(URL): Add support to query timestamp by URL
Description
Validation performed
@Henry8192 can we address the linter checks before we proceed further? If the ClpIrDecoder return error is the blocker, we can wait for https://github.com/y-scope/clp-ffi-js/pull/42 to be merged first.
@Henry8192 can we address the linter checks before we proceed further? If the ClpIrDecoder return error is the blocker, we can wait for y-scope/clp-ffi-js#42 to be merged first.
Interestingly, I don't get why it generates lint errors. On my end, yscope-log-viewer/src/services/LogFileManager/index.ts is less than 450 lines. The other errors are due to non-implemented functions in clp-ffi-js.
|
2025-04-01T04:36:00.272935
| 2019-09-10T05:45:55
|
491462935
|
{
"authors": [
"umbreon22"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12186",
"repo": "y785/script-api",
"url": "https://github.com/y785/script-api/issues/21"
}
|
gharchive/issue
|
askMenu index issues
askMenu doesn't use always use the indices sent to the client, it uses an array's. This will cause the script to end on custom indices being sent to the client, i.e.
\#L0# test #lr\n#L1000# test #l
-> The client will send 1000, but the server thinks the max is 1.
Resolved by a49356f44fb82e34df9d0f571e4e59402606186f
|
2025-04-01T04:36:00.295232
| 2022-04-07T14:31:38
|
1196138290
|
{
"authors": [
"laoshaw",
"yaegassy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12188",
"repo": "yaegassy/coc-volar",
"url": "https://github.com/yaegassy/coc-volar/issues/138"
}
|
gharchive/issue
|
show warning for syntax errors?
Normally when I save in vim, it will reformat my vue file. If there is a syntax error, volar will refuse to reformat my vue file silently.
is it possible for volar to provide hints for those wrong syntax or typos in the code? I just spent an hour to check my code and found there is a place I put `;' instead of ',' inside export default.
in c/c++ etc vim scenario, vim can highlight the syntax errors for quick fixes. I wonder if volar can do that too? or it is already there but I just don't know how to set it up?
Syntax errors, etc. will be displayed as diagnostics messages.
You can check with "sign" (>>) or "popup/floatwin". You can also run the :CocDiagnostics command to see the list in the location list.
They are enabled by default. If they are not working in your environment, it may be something wrong with your environment. Check coc.nvim or Vim8/Neovim documentation
https://user-images.githubusercontent.com/188642/162340165-3c07b199-8ba3-42fb-b82d-029f1099119f.mp4
I'm trying to debug what is wrong on my end.
does I need both CocInstall and PlugInstall to install, or any one of them will do?
I assumed one of them will do, so I did CocInstall, for whatever reason it does not report errors. I tried PlugInstall it says:
'Warning<EMAIL_ADDRESS>The engine "coc" appears to be invalid'
PlugInstall did not help, still no errors report.
CocDiagnostics does not report anything either. I disabled ALE(no more vls?), make sure I always disabled vim-vue and all other settings related to vue, to no help.
Is there any way I can dump some debug info about coc-volar?
If you installed with :CocInstall, you do not need to install with vim-plug. As output in the log, if both are installed, the :CocInstall side is enabled. Please uninstall coc-volar installed by vim-plug as it is not needed. Please check the vim-plug documentation on how to uninstall it.
Create a project with create-vite and try syntax errors in the <script> block.
Prepare:
cd /tmp
yarn create vite --template vue-ts example
cd example
yarn
nvim
Check:
:e src/App.vue
I use Neovim regularly. I tried it with Vim8 just to be sure, and it works correctly.
Yes a clean env works here as well, but I had this .eslintrc.js from previous work and it seems the one causing the trouble:
module.exports = {
root: true,
env: {
browser: true,
node: true,
es2021: true,
},
extends: [
"plugin:vue/vue3-recommended",
"eslint:recommended",
"@vue/typescript/recommended",
"prettier",
"@vue/prettier/@typescript-eslint",
],
parser: "vue-eslint-parser",
plugins: ["vue"],
parserOptions: {
ecmaVersion: 2021,
},
rules: {
"no-console": process.env.NODE_ENV === "production" ? "warn" : "off",
"no-debugger": process.env.NODE_ENV === "production" ? "warn" : "off",
"no-unused-vars": ["error", { args: "none" }],
"@typescript-eslint/interface-name-prefix": "off",
"@typescript-eslint/explicit-function-return-type": "off",
"@typescript-eslint/explicit-module-boundary-types": "off",
"@typescript-eslint/no-explicit-any": "off",
"@typescript-eslint/no-non-null-assertion": "off",
"@typescript-eslint/no-unused-vars": ["off", { argsIgnorePattern: "^_" }],
},
};
there are quite some tutorials online, I am not sure if I still need this eslintrc.js for vite3+typescript+vim, or part of it. will debug more.
If you want to use eslint, install coc-eslint. volar does not respect eslint settings, etc.
|
2025-04-01T04:36:00.309672
| 2024-07-01T16:36:44
|
2384269852
|
{
"authors": [
"yagmur-kolanc"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12189",
"repo": "yagmur-kolanc/Flutter",
"url": "https://github.com/yagmur-kolanc/Flutter/issues/2"
}
|
gharchive/issue
|
Screen Page
“introduction_screen” Page
“login_page” Page
Username or password is incorrect” Warning Code
“register_screen” Page
Name Check code
Email Check code
password Check code
password warning code
The screen containing the warnings received when the requested information does not match.
Registration Successful Message
Authentication Process
An asynchronous function was used. The function also includes email, password, username.
Creates a new user by taking the parameters. "try-catch" in the function
Sign up user via FirebaseAuth.instance using block
“FirebaseAuth.instance.createUserWithEmailAndPassword(email:
email, password: password)” will connect the user with the specified email and password.
It was used to register in Firebase Authentication. If during the registration process a
If an error occurs, the error is caught and false is returned. If the error has the code "email-already-in use", the message "Please enter a different email" is printed on the console.
“forget_password” Page
###“welcome_page” Page
After the user logs into the app, they are presented with a screen where they can choose how to use the application. They can select either "Find by Ingredient" or "Find by Category" options. Since they are directed to this page after the login screen, a "Login Successful" message is displayed.
###“reciepe_choose_page” Page
After selecting the "Find by Ingredient" option, the user is presented with food ingredients in the form of checkboxes, as shown in the screenshot. The user can select the ingredients they have on hand. The images displayed on the screen are located in the ingredients folder within the assets directory
###“reciepe_result_page” Page
On the "recipe_choose" page, the user is presented with the images and names of recipes from the database that include the selected ingredients. As shown in the figure, the selected ingredients are designed to be displayed over the image of the corresponding recipe.
“category_page” Page
On the "welcome_page," after the user selects the "Find by Category" option, they are redirected to the "category_page." The "category_page" displays various food categories.
The images for the categories are taken from the categories folder within the "assets" directory. The code for this process has been written as shown in the figure.
###“category_result_page” Page
On the "category_page," the user can select a food category and view recipes from the database that belong to the chosen category. For example, if a user selects the dessert category, they will be directed to the "category_result_page," where the recipes are listed with their images and names, as shown in the figure.
###“reciepe_detail_page” Page
This page was created for users who select either the "Find by Ingredient" or "Find by Category" option. In both cases, the chosen recipe's image, ingredients, and instructions are displayed on the screen. The figure shows a recipe selected from the dessert category.
/
Project Screen Page Views
|
2025-04-01T04:36:00.320950
| 2015-07-23T17:33:10
|
96864111
|
{
"authors": [
"ClaudioLaBarbera",
"yagop"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12190",
"repo": "yagop/node-telegram-bot-api",
"url": "https://github.com/yagop/node-telegram-bot-api/issues/13"
}
|
gharchive/issue
|
Group chat doesn't work
Hi, I saw that if i add a bot in a group chat it doesn't see messages.
This is a Telegram Bot API characteristic, see official documentation
https://core.telegram.org/bots#privacy-mode
Ah ok. Sorry!
|
2025-04-01T04:36:00.322522
| 2015-09-03T21:22:08
|
104784580
|
{
"authors": [
"Vijar",
"sathify"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12191",
"repo": "yahoo/fluxible-plugin-fetchr",
"url": "https://github.com/yahoo/fluxible-plugin-fetchr/pull/47"
}
|
gharchive/pull-request
|
chore(README.md): add info on using body parser for POST to work
This is related to the issue where the app did not use body-parser and create service was failing with bad stack trace. This PR adds the information to add body-parser in the README.
Sweet, thanks for squashing!
|
2025-04-01T04:36:00.325366
| 2017-02-26T23:44:28
|
210359522
|
{
"authors": [
"DavidLiuXh",
"zer0Id0l"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12192",
"repo": "yahoo/kafka-manager",
"url": "https://github.com/yahoo/kafka-manager/issues/354"
}
|
gharchive/issue
|
How Consumer list for given topic is fetched?
I was trying to understand how the consumer list for given topic is displayed in kafka manger.
Looks like 'topicConsumerMap' stores that information, but not able to figure when 'topicConsumerMap' is populated?
Is this mapping stored with kafka manger or fetched realtime from zookeeper and/or broker?
consume __consume_offsets topic to get offsets managed by broker and read /consumer/[group]/[topic]/offset to get offset managed by zk. Please refer KafkaStateActor.scala
|
2025-04-01T04:36:00.327190
| 2021-07-13T12:39:14
|
943310799
|
{
"authors": [
"adriananeci",
"edeNFed"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12193",
"repo": "yahoo/kubectl-flame",
"url": "https://github.com/yahoo/kubectl-flame/pull/60"
}
|
gharchive/pull-request
|
Add serviceAccountName cli option
This PR adds a new cli option, serviceAccountName, useful when you want to run the flame job under a specific service account.
Resolves https://github.com/yahoo/kubectl-flame/issues/59
I confirm that this contribution is made under the terms of the license found in the root directory of this repository's source tree and that I have the authority necessary to make this contribution on behalf of its copyright owner.
lgtm
|
2025-04-01T04:36:00.337561
| 2015-07-29T06:47:00
|
97872600
|
{
"authors": [
"adon-at-work"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12194",
"repo": "yahoo/xss-filters",
"url": "https://github.com/yahoo/xss-filters/pull/47"
}
|
gharchive/pull-request
|
improve release workflow
thereafter, the task of releasing a new version becomes easier.
To release a version
grunt release will automatically set version z=z+1 as in x.y.z
grunt release minor will automatically set version y = y+1, z = 0 as in x.y.z
grunt release major will automatically set version x = x+1, y = z = 0 as in x.y.z
p.s. append --dry-run if needed for a test.
What it does:
the new version number will be written into both package.json and bower.json
call grunt dist to generate a new file (with proper version #)
git commit package.json, bower.json, dist/. -m "Release vx.y.z"
git tag -a vx.y.z -m "Version x.y.z"
git push origin && git push origin --tags
this change is removed
|
2025-04-01T04:36:00.363626
| 2015-04-08T06:03:42
|
67056439
|
{
"authors": [
"santigarcor",
"yajra"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12195",
"repo": "yajra/laravel-datatables-oracle",
"url": "https://github.com/yajra/laravel-datatables-oracle/issues/22"
}
|
gharchive/issue
|
Problem when doing any kind of filtering
Hi yajra, first of all, your plugin is really amazing, but im having a problem with the last update that you did, in the version 4.0.11 i can make any kind of filter without problems, but i decided to update to <IP_ADDRESS> and when i try to filter something, laravel crashes and says :
FatalErrorException in Datatables.php line 439: Function name must be a string
But when i try to do the filter by one specific column it says:
ErrorException in Datatables.php line 473: Undefined variable: connection
Anyway thank you, meanwhile i'm using 4.0.11
thanks for reporting.. a copy paste issue error on my last commit. my bad sorry!
... still fixing please use 4.0.11 atm...
done.. kindly checkout v<IP_ADDRESS>... sorry for the inconvenience
Thank you friend, this package is amazing :+1:
|
2025-04-01T04:36:00.384076
| 2024-06-08T08:29:23
|
2341549414
|
{
"authors": [
"108yen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12196",
"repo": "yamada-ui/yamada-ui",
"url": "https://github.com/yamada-ui/yamada-ui/issues/2020"
}
|
gharchive/issue
|
Prohibit import of @testing-library/react in lint
Description
Prohibit import of @testing-library/react in lint.
Problem Statement/Justification
Must be imported from @yamada-ui/test.
Proposed Solution or API
No response
Alternatives
No response
Are you willing to participate in implementing this feature and create a pull request include the implementation?
None
Additional Information
No response
https://www.npmjs.com/package/eslint-plugin-import-replace
|
2025-04-01T04:36:00.385654
| 2021-01-24T15:32:56
|
792822255
|
{
"authors": [
"WSSDude420"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12197",
"repo": "yamashi/CyberEngineTweaks",
"url": "https://github.com/yamashi/CyberEngineTweaks/pull/436"
}
|
gharchive/pull-request
|
Expose spdlog.info to mods
Fix small oversight on my part :P
Actually, this would be moved in sandboxing to somewhere else, just realised...
So can be closed, redundant PR in the end.
Actually, this would be moved in sandboxing to somewhere else, just realised...
So can be closed, redundant PR in the end.
|
2025-04-01T04:36:00.388502
| 2022-11-09T16:52:18
|
1442427710
|
{
"authors": [
"daniele2010",
"yanascz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12198",
"repo": "yanascz/bitcoin-widgets-android",
"url": "https://github.com/yanascz/bitcoin-widgets-android/issues/3"
}
|
gharchive/issue
|
Moscow Time
Hi, first congratulations for your work, finally a widget about timechain data!
Do you consider implementing a moscow time widget?
Thank you again
Yep, it’s in my TODO list. However, I had to work on something else in the past several weeks. Hope to get to it soon. Please be patient. 🙏
@daniele2010, it's available in https://github.com/yanascz/bitcoin-widgets-android/tree/moscow-time. You may build and deploy the branch via Android Studio and give it a try. 😉 Or wait for a Google Play update.
really appreciated! thank you again!!!
Still waiting for app review, it's been in progress since Sunday. 🤷♂️
In the meantime, I've already built it! 😃
Update available on Google Play. 🥳
|
2025-04-01T04:36:00.400552
| 2018-10-27T19:55:21
|
374682476
|
{
"authors": [
"alesapin",
"dorantor"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12199",
"repo": "yandex/ClickHouse",
"url": "https://github.com/yandex/ClickHouse/issues/3480"
}
|
gharchive/issue
|
Wrong error message for null in array + possible memory leak
Minimal test case:
CREATE TABLE arraytest
(
created_date Date DEFAULT toDate(created_at),
created_at DateTime DEFAULT now(),
strings Array(String) DEFAULT emptyArrayString()
)
ENGINE = MergeTree(created_date, cityHash64(created_at), (created_date, cityHash64(created_at)), 8192)
insert into arraytest (created_at, strings) values (now(), ['aaaaa', 'bbbbb', null]);
Instead of some error about wrong type in array got:
Exception on client:
Code: 173. DB::Exception: Allocator: Cannot mmap 8.00 EiB., errno: 12, strerror: Cannot allocate memory
Minimal test case prepared with 1.1.54343 but can be reproduced on recent versions also.
Forgot to mention. If this errors happens for some time(as we had) then you'll get segfaults in threads and after some more time server will silently close. This makes me think that memory leaks are very possible.
Fixed in https://github.com/yandex/ClickHouse/pull/3487.
Awesome! Thanks!
|
2025-04-01T04:36:00.402373
| 2017-06-27T06:09:00
|
238746192
|
{
"authors": [
"svenefftinge"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12200",
"repo": "yang-tools/yang-lsp",
"url": "https://github.com/yang-tools/yang-lsp/issues/5"
}
|
gharchive/issue
|
Implement linking/scoping for modules and submodules
This ticket is about defining the scopes for modules, submodules. Implementing the support for prefixes and resolution of imports and included.
we should streamline the error codes and messages according to pyang : https://github.com/mbj4668/pyang/blob/master/pyang/error.py#L41
This is now solved.
|
2025-04-01T04:36:00.413797
| 2023-08-03T02:45:29
|
1834179197
|
{
"authors": [
"JiahaoChen1",
"yangnianzu0515"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12202",
"repo": "yangnianzu0515/MoleOOD",
"url": "https://github.com/yangnianzu0515/MoleOOD/issues/2"
}
|
gharchive/issue
|
hyper-parameter to reproduce the results
Hi, I have some questions about how to reproduce the results. For the bace, the bbbp, the sider, and the hiv, how can I set the hyper-parameter like random seed(five times), dropout ratio, learning rate, etc.
Hello, we have provided the range of our hyperparameter selections in our appendix. Please refer to Section C.2 in the appendix for conducting a grid search to find the optimal parameters. It's crucial to note that for different datasets, you should select the optimal parameters separately. Due to the relatively large number of hyperparameters, the search space might be extensive, and you may need to invest some time in this process.
Additionally, at the beginning, we noticed significant variations in results when using different random seeds with the framework code provided by OGB. We also encountered a similar phenomenon in another paper [1], where multiple runs showed considerable deviation in results. Later, we discovered that PyG inherently introduces randomness, making it impossible to fix the seed, which is a limitation of PyG itself. Therefore, even if we attempt to fix the same seed, the results cannot be reliably reproduced. Many users on the PyG open-source repository have suggested that this might be related to the "scatter" issue. Regarding the random seed, we only remember one, which is 2022. The other four are not remembered.
If you have any suggestions or methods to address the issues with PyG itself, we welcome you to submit a pull request to help us resolve them. Thank you for your understanding and support.
Reference:
Li H, Wang X, Zhang Z, et al. Ood-gnn: Out-of-distribution generalized graph neural network[J]. IEEE Transactions on Knowledge and Data Engineering, 2022.
|
2025-04-01T04:36:00.436944
| 2024-01-26T23:53:43
|
2103071996
|
{
"authors": [
"christophrus",
"jeefo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12203",
"repo": "yapb/yapb",
"url": "https://github.com/yapb/yapb/issues/512"
}
|
gharchive/issue
|
yb_enable_query_hook not working properly on Internet tab
Servers player count works properly in game browsers favourites tab and in detailed server info window, but does not work on Internet tab. I tried with both release 4.3.734 and current master.
It's normal. Not sure, but probably something has to be changed here: to get it working, not on bot side.
but not sure it's a good idea to fake internet servers)
|
2025-04-01T04:36:00.438285
| 2024-03-28T00:10:49
|
2212094849
|
{
"authors": [
"jetjodh",
"yardenfren1996"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12204",
"repo": "yardenfren1996/B-LoRA",
"url": "https://github.com/yardenfren1996/B-LoRA/issues/1"
}
|
gharchive/issue
|
Training content only blocks do not work for person
I tried your code to train a blora model on a small set of images of myself but was not able to replicate my facial features at all when trained with content only blocks. Any suggestions?
Hi, thank you for reaching out. Actually, it's a very interesting point you've made. I'm currently investigating that issue, and I hope to provide a satisfactory answer as soon as possible.
Thank you!
I'll try adding the class name to the prompt and try again.
|
2025-04-01T04:36:00.537268
| 2017-04-08T05:39:41
|
220377920
|
{
"authors": [
"aravindGear4",
"yctung"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12206",
"repo": "yctung/AndroidLibSvm",
"url": "https://github.com/yctung/AndroidLibSvm/issues/1"
}
|
gharchive/issue
|
Storing the SVM model
I am creating a mobile application with SVM.
Since SVM learning takes time, is there way to store the learnt model once and then re-use it in the later when the application is opened again.
Does this library outputs the model to a file ? If so my issue will be almost resolved
Hi Aravind,
This lib does output the model. I am not using my laptop now. I can check
the code and let you know how to do it next monday.
Yu-Chih
On Sat, Apr 8, 2017 at 1:39 AM, Aravind<EMAIL_ADDRESS>wrote:
I am creating a mobile application with SVM.
Since SVM learning takes time, is there way to store the learnt model once
and then re-use it in the later when the application is opened again.
Does this library outputs the model to a file ? If so my issue will be
almost resolved
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/yctung/AndroidLibSvm/issues/1, or mute the thread
https://github.com/notifications/unsubscribe-auth/AJ3-8RM626MTA-SVOFSzMGtxqzf1iUIGks5rtx2dgaJpZM4M3mnx
.
--
Regards
Yu-Chih Tung
Hi Yu-Chih,
Library works as expected. Awesome library ! very light weight compared to the java version I was using.
Hi Aravind,
I am sorry I have checked the code yeterday. I will do it today or tomorrow.
On Sat, Apr 8, 2017 at 5:46 AM, Aravind<EMAIL_ADDRESS>wrote:
Closed #1 https://github.com/yctung/AndroidLibSvm/issues/1.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/yctung/AndroidLibSvm/issues/1#event-1035105534, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AJ3-8Qco8N9qyk72L5Qm3qunAAPDyYQXks5rt1eTgaJpZM4M3mnx
.
--
Regards
Yu-Chih Tung
Hi Aravind,
I have checked my other project's using the libsvm and do find the code to
output model as a file in external storage so the predict function can load
the model without retraining. Is is what you want? If so, I will update
them into the AndroidLibSvm project.
Yu-Chih
On Tue, Apr 11, 2017 at 1:34 PM, Yu-Chih Tung<EMAIL_ADDRESS>wrote:
Hi Aravind,
I am sorry I have checked the code yeterday. I will do it today or
tomorrow.
On Sat, Apr 8, 2017 at 5:46 AM, Aravind<EMAIL_ADDRESS>wrote:
Closed #1 https://github.com/yctung/AndroidLibSvm/issues/1.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/yctung/AndroidLibSvm/issues/1#event-1035105534, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AJ3-8Qco8N9qyk72L5Qm3qunAAPDyYQXks5rt1eTgaJpZM4M3mnx
.
--
Regards
Yu-Chih Tung
--
Regards
Yu-Chih Tung
Hey Yu-Chih,
I said I did not find any issue with current working of the output model file. It is fine.
"Library works as expected. Awesome library ! very light weight compared to the java version I was using."
If you could, I have created a pull request can you check that, it is regarding running the module in windows machine.
Thanks.
Wow. I guess I misunderstand your previous mail. I will check the pull
request. Thank you!
On Wed, Apr 12, 2017 at 10:26 PM, Aravind<EMAIL_ADDRESS>wrote:
Hey Yu-Chih,
I said I did not find any issue with current working of the output model
file. It is fine.
"Library works as expected. Awesome library ! very light weight compared
to the java version I was using."
If you could, I have created a pull request can you check that, it is
regarding running the module in windows machine.
Thanks.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/yctung/AndroidLibSvm/issues/1#issuecomment-293759417,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AJ3-8cM9aa46ABkzfoumMYkZ6yGgNSuOks5rvYftgaJpZM4M3mnx
.
--
Regards
Yu-Chih Tung
|
2025-04-01T04:36:00.541861
| 2023-08-29T06:34:35
|
1870984120
|
{
"authors": [
"abcxubu",
"ycwu1997"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12207",
"repo": "ycwu1997/MC-Net",
"url": "https://github.com/ycwu1997/MC-Net/issues/14"
}
|
gharchive/issue
|
About the parameter quantity in the MC-Net+
Thanks for sharing the code. I have a question about the parameter quantity in the MC-Net+. The backbone of the MC-Net+ is the Vet, and MC-Net+ has one encoder and three decoders (by reading your code I found these decoders do not share the weights). In Tab. 2, you said the parameter quantity of both the Vnet and MC-Net+ is 9.44, and the parameter quantity of Multi-scale MC-Net+ is 5.88. Why do the Vnet and MC-Net+ have the same parameter quantity? Why does the Multi-scale MC-Net+ have less parameter quantity? Could you explain about this? Thank you.
During the inference, we only used the original decoder, and the other two decoders were only used for training. Therefore, the model size is only computed in the inference time. About the multi-scale MC-Net+ model, we built this version upon the URPC backbones. Specifically, we applied our mutual consistency constraints to train the URPC model. Note that, we here also report the inferenced model size. Thanks.
I see. Thanks for your reply.
在 2023-08-29 14:41:26,"Eli Wu" @.***> 写道:
During the inference, we only used the original decoder, and the other two decoders were only used for training. Therefore, the model size is only computed in the inference time. About the multi-scale MC-Net+ model, we built this version upon the URPC backbones. Specifically, we applied our mutual consistency constraints to train the URPC model. Note that, we here also report the inferenced model size. Thanks.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: @.***>
|
2025-04-01T04:36:00.543667
| 2024-03-21T17:00:31
|
2200708054
|
{
"authors": [
"dcherednik"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12208",
"repo": "ydb-platform/ydb",
"url": "https://github.com/ydb-platform/ydb/pull/3064"
}
|
gharchive/pull-request
|
Fix update token interval calculation.
We should chouse min value from expiresIn and RefreshPeriod_, not max one
Changelog category
Bugfix
Additional information
...
We should chouse min value from expiresIn and RefreshPeriod_, not max one
Also reduce server returned interval by 2 to prevent race at expiring time.
|
2025-04-01T04:36:00.561395
| 2023-09-13T08:27:13
|
1894012354
|
{
"authors": [
"andlrc",
"maximyurevich"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12209",
"repo": "yegappan/lsp",
"url": "https://github.com/yegappan/lsp/issues/391"
}
|
gharchive/issue
|
svelteserver errors for open brackets (
The svelte LSP server is for some reason sending a list of numbers instead of a string in the label property:
/**
* Represents the signature of something callable. A signature
* can have a label, like a function-name, a doc-comment, and
* a set of parameters.
*/
export interface [SignatureInformation](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#signatureInformation) {
/**
* The label of this signature. Will be shown in
* the UI.
*/
label: string;
/**
* The human-readable doc-comment of this signature. Will be shown
* in the UI but can be omitted.
*/
documentation?: string | [MarkupContent](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#markupContentInnerDefinition);
/**
* The parameters of this signature.
*/
parameters?: [ParameterInformation](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#parameterInformation)[];
/**
* The index of the active parameter.
*
* If provided, this is used in place of `SignatureHelp.activeParameter`.
*
* @since 3.16.0
*/
activeParameter?: [uinteger](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#uinteger);
}
See: https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#signatureInformation
Can you enable debugging for the server and send the log? See :h lsp-debug for how to do this.
I'm curious as to why the svelte server is sending the list of numbers.
The svelte LSP server is for some reason sending a list of numbers instead of a string in the label property:
/**
* Represents the signature of something callable. A signature
* can have a label, like a function-name, a doc-comment, and
* a set of parameters.
*/
export interface [SignatureInformation](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#signatureInformation) {
/**
* The label of this signature. Will be shown in
* the UI.
*/
label: string;
/**
* The human-readable doc-comment of this signature. Will be shown
* in the UI but can be omitted.
*/
documentation?: string | [MarkupContent](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#markupContentInnerDefinition);
/**
* The parameters of this signature.
*/
parameters?: [ParameterInformation](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#parameterInformation)[];
/**
* The index of the active parameter.
*
* If provided, this is used in place of `SignatureHelp.activeParameter`.
*
* @since 3.16.0
*/
activeParameter?: [uinteger](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#uinteger);
}
See: https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#signatureInformation
Can you enable debugging for the server and send the log? See :h lsp-debug for how to do this.
I'm curious as to why the svelte server is sending the list of numbers.
lsp-tsserver.log
@maximyurevich can you try out https://github.com/yegappan/lsp/pull/392 I havn't had a change to test it myself, as I don't have a svelte LSP server installed.
@maximyurevich can you try out #392 I havn't had a change to test it myself, as I don't have a svelte LSP server installed.
Tomorrow
@maximyurevich can you try out #392 I havn't had a change to test it myself, as I don't have a svelte LSP server installed.
works fine
https://github.com/yegappan/lsp/assets/21293376/255aa3bf-0cf1-446a-a029-5be92af1a9bd
@maximyurevich can you try out #392 I havn't had a change to test it myself, as I don't have a svelte LSP server installed.
Why this commit in draft?
|
2025-04-01T04:36:00.564884
| 2023-03-20T01:13:07
|
1631208717
|
{
"authors": [
"Shane-XB-Qian",
"andlrc",
"yegappan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12210",
"repo": "yegappan/lsp",
"url": "https://github.com/yegappan/lsp/pull/167"
}
|
gharchive/pull-request
|
fix: omnifunc broken
pr #160 was missed closed, so reopen a new one for it.
Yes. I was wondering why this PR was closed. Can you reopen the PR? Also, do you have an example that shows the problem with not ignoring the case? I am going to start adding tests for insert completion checking for the various cases.
anyway, those had been at my local already, wish you can merge, as for dup=0 or not, please help think... :+1:
thx. one case is a simple js code:
console.c|
cursor at |, if no icase=1, looks it would lost console.Console (.C is upper case), though i am not very sure if it was solid, but so far i can say that.
@Shane-XB-Qian What language server are you using? And how is it configured?
What language server
same one like ts
Yes. I was wondering why this PR was closed.
I think it's because Shane pushed the branch which, compared to main, had zero changes.
|
2025-04-01T04:36:00.595896
| 2018-05-19T17:29:58
|
324646223
|
{
"authors": [
"0crat",
"g4s8",
"jimkast",
"rultor",
"yegor256"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12211",
"repo": "yegor256/takes",
"url": "https://github.com/yegor256/takes/pull/840"
}
|
gharchive/pull-request
|
RsPrint close stream
Fixed #838 . RsPrint should close stream after printing it to an OutputStream.
Job #840 is now in scope, role is REV
This pull request #840 is assigned to @g4s8/z, here is why; the budget is 15 minutes, see §4; please, read §27 and when you decide to accept the changes, inform @yegor256/z (the architect) right in this ticket; if you decide that this PR should not be accepted ever, also inform the architect; this blog post will help you understand what is expected from a code reviewer
@jimkast thanks for contributing. But there are some code style and quality issues with this PR:
[INFO] Checkstyle: src/main/java/org/takes/rs/RsPrint.java[31]: Empty line or comment between imports is not allowed (ImportCohesionCheck)
[INFO] Checkstyle: src/main/java/org/takes/rs/RsPrint.java[32]: Using the '.*' form of import should be avoided - java.io.*. (AvoidStarImportCheck)
[INFO] Checkstyle: src/main/java/org/takes/rs/RsPrint.java[32]: Wrong order for 'java.io.*' import. (ImportOrderCheck)
[INFO] Checkstyle: src/test/java/org/takes/rs/RsPrintTest.java[30]: Empty line or comment between imports is not allowed (ImportCohesionCheck)
[INFO] Checkstyle: src/test/java/org/takes/rs/RsPrintTest.java[31]: Wrong order for 'java.io.ByteArrayInputStream' import. (ImportOrderCheck)
[INFO] Checkstyle: src/test/java/org/takes/rs/RsPrintTest.java[43]: Class Data Abstraction Coupling is 9 (max allowed is 7) classes [AtomicBoolean, ByteArrayInputStream, ByteArrayOutputStream, FailOutputStream, IOException, RsPrint, RsText, RsWithBody, RsWithHeader]. (ClassDataAbstractionCouplingCheck)
[INFO] Checkstyle: src/test/java/org/takes/rs/RsPrintTest.java[103]: Missing a Javadoc comment. (JavadocMethodCheck)
[INFO] Checkstyle: src/test/java/org/takes/rs/RsPrintTest.java[106]: Line is longer than 80 characters (found 88). (LineLengthCheck)
[INFO] Checkstyle: src/test/java/org/takes/rs/RsPrintTest.java[106]: Parameters should start on a new line (BracketsStructureCheck)
please fix them. To ensure that your changes pass quality checks run maven build with enabled qulice plugin:
mvn clean install -Pqulice
@g4s8/z this job was assigned to you 5days ago. It will be taken away from you soon, unless you close it, see §8. Read this and this, please.
@jimkast this PR still failed to build, see https://travis-ci.org/yegor256/takes/builds/381567096?utm_source=github_status&utm_medium=notification logs or try to build it locally with
mvn clean install -Pqulice
I removed the test which caused issues but still cannot overcome qulice code style checks.
@jimkast You have checkstyle issue:
Checkstyle: src/main/java/org/takes/rs/RsPrint.java[31]: Wrong order for 'java.io.ByteArrayOutputStream' import. (ImportOrderCheck)
it means that in file src/main/java/org/takes/rs/RsPrint.java you have java imports in wrong order, checkstyle forces to use alphabetically import order.
The user @g4s8/z resigned from #840, please stop working. Reason for job resignation: It is older than 10 days, see §8
Resigned on delay, see §8: -15 point(s) just awarded to @g4s8/z
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@rultor merge
@rultor merge
@yegor256 OK, I'll try to merge now. You can check the progress of the merge here
@rultor merge
@jimkast @yegor256 Oops, I failed. You can see the full log here (spent 9min)
++ jq -r .build.status
+ status=running
+ '[' running == null ']'
+ '[' running == success ']'
+ '[' running == failed ']'
+ echo 'waiting for AppVeyor build 3035: running'
waiting for AppVeyor build 3035: running
+ sleep 5s
+ true
++ curl -K ../curl-appveyor.cfg https://ci.appveyor.com/api/projects/yegor256/takes/build/3035
++ jq -r .build.status
+ status=running
+ '[' running == null ']'
+ '[' running == success ']'
+ '[' running == failed ']'
+ echo 'waiting for AppVeyor build 3035: running'
waiting for AppVeyor build 3035: running
+ sleep 5s
+ true
++ curl -K ../curl-appveyor.cfg https://ci.appveyor.com/api/projects/yegor256/takes/build/3035
++ jq -r .build.status
+ status=running
+ '[' running == null ']'
+ '[' running == success ']'
+ '[' running == failed ']'
+ echo 'waiting for AppVeyor build 3035: running'
waiting for AppVeyor build 3035: running
+ sleep 5s
+ true
++ curl -K ../curl-appveyor.cfg https://ci.appveyor.com/api/projects/yegor256/takes/build/3035
++ jq -r .build.status
+ status=running
+ '[' running == null ']'
+ '[' running == success ']'
+ '[' running == failed ']'
+ echo 'waiting for AppVeyor build 3035: running'
waiting for AppVeyor build 3035: running
+ sleep 5s
+ true
++ curl -K ../curl-appveyor.cfg https://ci.appveyor.com/api/projects/yegor256/takes/build/3035
++ jq -r .build.status
+ status=running
+ '[' running == null ']'
+ '[' running == success ']'
+ '[' running == failed ']'
+ echo 'waiting for AppVeyor build 3035: running'
waiting for AppVeyor build 3035: running
+ sleep 5s
+ true
++ curl -K ../curl-appveyor.cfg https://ci.appveyor.com/api/projects/yegor256/takes/build/3035
++ jq -r .build.status
+ status=failed
+ '[' failed == null ']'
+ '[' failed == success ']'
+ '[' failed == failed ']'
+ echo 'see https://ci.appveyor.com/project/yegor256/takes/build/3035'
see https://ci.appveyor.com/project/yegor256/takes/build/3035
+ exit 1
container a38959a8d6af236672760b511897cca6f80467cf9f51d971df1f55376be20087 is dead
Mon Jun 11 23:01:36 CEST 2018
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
@yegor256/z everybody who has role REV is banned at #840; I won't be able to assign anyone automatically; consider assigning someone manually (as in §19), or invite more people (as in §51), or remove the job from the scope (as in §14)
Job gh:yegor256/takes#840 is not assigned, can't get performer
The job #840 is now out of scope
|
2025-04-01T04:36:00.628068
| 2021-04-21T09:59:54
|
863693781
|
{
"authors": [
"JingyuanHu",
"yhenon"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12213",
"repo": "yhenon/pytorch-retinanet",
"url": "https://github.com/yhenon/pytorch-retinanet/pull/220"
}
|
gharchive/pull-request
|
loss bug fixed
There is a bug about regression_losses in FocalLoss. torch.stack will report : "Expected object of backend CUDA but got backend CPU for sequence element 0 in sequence argument at .... " and crashed. Here is the bug fixed.
thanks for your contribution
|
2025-04-01T04:36:00.638428
| 2024-06-26T07:05:41
|
2374529936
|
{
"authors": [
"RatishT",
"yihong0618"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12214",
"repo": "yihong0618/bilingual_book_maker",
"url": "https://github.com/yihong0618/bilingual_book_maker/pull/408"
}
|
gharchive/pull-request
|
Fix pylint warning redefined-outer-name
Changes Made:
Applied automated fixes for the pylint warning redefined-outer-name.
"Used when a variable's name hides a name defined in an outer scope or except handler.. See https://pylint.readthedocs.io/en/latest/user_guide/messages/warning/redefined-outer-name.html"
Note:
This pull request is part of a research project conducted by researchers from TU Delft, titled "PyWarnFixer: Using ML to Fix Pylint Warnings." The goal of this study is to investigate the perceptions and practices of code quality among developers in Python open-source projects and to develop a tool that uses AI to automatically fix pylint warnings.
Research Study Information:
This pull request is part of a research project. For more information about the study, please visit the project's information page.
Consent to Participate:
If you review this pull request, you are invited to participate in our study. Your participation is voluntary. To provide your consent, just open an issue in our repository with the provided template using the following link: consent issue template .
Thank you for considering participation in our research. Your feedback is crucial and highly valued. If you have any questions or concerns, please contact the Responsible Researcher at<EMAIL_ADDRESS>
Thanks
Thanks
I see that you said thanks but haven't merged the PR yet, is there a reason why?
|
2025-04-01T04:36:00.662096
| 2018-12-17T19:38:38
|
391860530
|
{
"authors": [
"machour",
"schmunk42"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12215",
"repo": "yiisoft/yii-base-web",
"url": "https://github.com/yiisoft/yii-base-web/pull/62"
}
|
gharchive/pull-request
|
docs: Rewrite README & in-app instructions
Q
A
Is bugfix?
no
New feature?
yes
Breaks BC?
no
Tests pass?
Fixed issues
#60 #53
I've fixed the README instruction to be more accurate and helpful.
I'm also testing available classes and displaying instructive error messages to help the User understand what's going on.
Couldn't resist adding 4 lines of CSS in this PR, considering that the broken design isn't helping either :)
Thank you!
|
2025-04-01T04:36:00.672299
| 2016-02-11T17:05:10
|
133030385
|
{
"authors": [
"samdark",
"thiagotalma"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12216",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/issues/10816"
}
|
gharchive/issue
|
[rest/UrlRule] Allow parameters in each controller.
It does not make much sense to have a global definition of the attributes "only" and "except". There would be few use cases accordingly.
The most common use case would depend on the definition for each controller.
Something like that:
'urlManager' => [
'rules' => [
[
'class' => 'yii\rest\UrlRule',
'controller' => [
'user',
'post' => [
'only' => ['create', 'delete'],
],
'comment' => [
'except' => ['delete'],
]
],
]
],
],
What do you think of this?
UrlRule already resolves to controller and action. Why do I need to specify it once again?
leave it alone...
|
2025-04-01T04:36:00.676681
| 2016-03-25T13:24:58
|
143497388
|
{
"authors": [
"SilverFire",
"bettermentit",
"des1roer",
"edwinover",
"hassansheikh",
"phping1"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12217",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/issues/11194"
}
|
gharchive/issue
|
Your requirements could not be resolved to an installable set of packages.
\composer.json
{
"name": "yiisoft/yii2-app-basic",
"description": "Yii 2 Basic Project Template",
"keywords": [
"yii2",
"framework",
"basic",
"project template"
],
"homepage": "http://www.yiiframework.com/",
"type": "project",
"license": "BSD-3-Clause",
"support": {
"issues": "https://github.com/yiisoft/yii2/issues?state=open",
"forum": "http://www.yiiframework.com/forum/",
"wiki": "http://www.yiiframework.com/wiki/",
"irc": "irc://irc.freenode.net/yii",
"source": "https://github.com/yiisoft/yii2"
},
"minimum-stability": "stable",
"require": {
"php": ">=5.4.0",
"yiisoft/yii2": ">=2.0.5",
"yiisoft/yii2-bootstrap": "*",
"yiisoft/yii2-swiftmailer": "*",
"amnah/yii2-user": "^5.0",
"voskobovich/yii2-many-many-behavior": "^3.0",
"2amigos/yii2-date-picker-widget": "~1.0",
"intervention/image": "^2.3",
"codeception/codeception": "*",
"codeception/specify": "*",
"wbraganca/yii2-videojs-widget": "~1.0",
"bower-asset/punycode": "1.3.*",
"bower-asset/jquery.inputmask": "3.1.*",
"bower-asset/punycode": "1.3.*"
},
"require-dev": {
"yiisoft/yii2-codeception": "^2.0",
"yiisoft/yii2-debug": "*",
"yiisoft/yii2-gii": "*",
"yiisoft/yii2-faker": "*"
},
"config": {
"process-timeout": 1800
},
"scripts": {
"post-create-project-cmd": [
"yii\\composer\\Installer::postCreateProject"
]
},
"extra": {
"yii\\composer\\Installer::postCreateProject": {
"setPermission": [
{
"runtime": "0777",
"web/assets": "0777",
"web/uploads": "0777",
"yii": "0755"
}
],
"generateCookieValidationKey": [
"config/web.php"
]
},
"asset-installer-paths": {
"npm-asset-library": "vendor/npm",
"bower-asset-library": "vendor/bower"
}
}
}
///////////////////
composer update
Loading composer repositories with package information
Updating dependencies (including require-dev)
Your requirements could not be resolved to an installable set of packages.
Problem 1
- yiisoft/yii2 2.0.7 requires bower-asset/punycode 1.3.* -> no matching package found
- yiisoft/yii2 2.0.6 requires bower-asset/jquery.inputmask 3.1.* -> no matching packa
- yiisoft/yii2 2.0.5 requires bower-asset/jquery.inputmask 3.1.* -> no matching packa
- yiisoft/yii2 dev-master requires bower-asset/punycode 1.3.* -> no matching package
- yiisoft/yii2 2.0.x-dev requires bower-asset/punycode 1.3.* -> no matching package f
- Installation request for yiisoft/yii2 >=2.0.5 -> satisfiable by yiisoft/yii2[2.0.5,
-master].
Potential causes:
- A typo in the package name
- The package is not available in a stable-enough version according to your minimum-stab
see <https://getcomposer.org/doc/04-schema.md#minimum-stability> for more details.
Read <https://getcomposer.org/doc/articles/troubleshooting.md> for further common problem
composer global require "fxp/composer-asset-plugin:~1.1.1"
composer update
Thanks! Its work :)
ths!@SilverFire
Thanks! @SilverFire, its working.....
Great job!
Thanks @SilverFire !
|
2025-04-01T04:36:00.687509
| 2016-06-01T13:59:03
|
157913794
|
{
"authors": [
"SamMousa",
"cebe",
"fernandezekiel",
"klimov-paul",
"mikehaertl",
"samdark"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12218",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/issues/11674"
}
|
gharchive/issue
|
Add CSV response formatter
Currently Yii offers several formatters like XML, JSON, JSONP.
It would be nice if Yii supported CSV as well.
This implementation uses a stream to support large data sets. (Using it with large data sets would require a traversable object to be passed in data, not a PHP array).
/**
* CsvResponseFormatter formats the given data into a csv response content.
*
* It is used by [[Response]] to format response data.
*
*/
class CsvResponseFormatter extends Component implements ResponseFormatterInterface
{
/**
* If true column names will be printed as the first line, if provided.
*/
public $includeColumnNames = true;
/**
* Formats the specified response.
* @param Response $response the response to be formatted.
*/
public function format($response)
{
$response->getHeaders()->set('Content-Type', 'text/csv; charset=UTF-8');
$data = $response->data;
$handle = fopen('php://temp/maxmemory:20971520', 'w+');
$first = true;
foreach($data as $row) {
if($first && $this->includeColumnNames && \yii\helpers\ArrayHelper::isAssociative($row)) {
fputcsv($handle, array_keys($row));
$first = false;
}
fputcsv($handle, $row);
}
rewind($handle);
$response->stream = $handle;
}
}
@yiisoft/core-developers any objections?
Shouldn't this formatter also support all the CSV options for fputcsv?
delimiter
The optional delimiter parameter sets the field delimiter (one character only).
enclosure
The optional enclosure parameter sets the field enclosure (one character only).
escape_char
The optional escape_char parameter sets the escape character (one character only).
Yes.
@samdark, does self-assigning mean that you are implementing it or just that you are in charging of managing it? --> I can implement it if you want.
Just in charge. I'd assign you but you're not in the yiisoft ;)
Have also added support for non-uniform data, for example an array of models where one model has columns A and B, and another model has A and C.
Tests have been added & changelog has been updated.
My opinion: this feature does not belongs to the core and it is better to be released and maintained as separated extension.
Although CSV eneration is quite common task indeed, I don't think it is a good idea to keep it inside the core. There are many nuances about the task itself and many possible ways of resolving it.
If we accept this - what next? 'PDF', 'RSS', 'YAML' - it will be no end to this.
Currently beside 'raw' and 'html' Yii supports only those formats, which are common for REST API. For me - that is enough. Everything else - is an application specific.
I don't agree; but if that's the teams' choice feel free to close the issue and I'll put it my extension repository instead.
I also think this is better suited as an extension.
nice feature, i would love to try this on our current application
OK. So let's do it as extension. @cebe, @klimov-paul is it general enough for extension under yiisoft?
is it general enough for extension under yiisoft?
I suppose @SamMousa is capable to maintain this extension on this own. Beside, since it is fully his idea it would be better, if he will use his own brand for it.
Ok. Thanks for the feedback, will set this up as separate extension.
On Jun 8, 2016 7:14 PM, "Paul Klimov"<EMAIL_ADDRESS>wrote:
is it general enough for extension under yiisoft?
I suppose @SamMousa https://github.com/SamMousa is capable to maintain
this extension on this own. Beside, since it is fully his idea it would be
better, if he will use his own brand for it.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/yiisoft/yii2/issues/11674#issuecomment-224662026, or mute
the thread
https://github.com/notifications/unsubscribe/AAhYzRUEPvSgMTG3CqYb2QIAHjuFw_r2ks5qJvhagaJpZM4Irk0g
.
|
2025-04-01T04:36:00.691499
| 2016-08-11T12:39:34
|
170635924
|
{
"authors": [
"ak868308",
"cebe"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12219",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/issues/12141"
}
|
gharchive/issue
|
Yii2 ActiveDataProvider fetch result correct but showing totalCount 0 in summary text
I am trying to filter data using dropdown so the query executed perfectly produces a result is always correct but in summary total count is '0' always but not first time when page loaded. This is annoying me I am completely unable to trace the error through debugging tool. But the important thing is that the same code is working fine on my local machine whenever I deployed the same code to production it shows 0 counts
My Model search code :
`public function search($params) {
$query = Tasks::find();
$query->where(['q_id' => $this->job_id]);
// $query->orderBy('created_at DESC');
$dataProvider = new ActiveDataProvider([
'query' => $query,
'pagination' => [
'pageSize' => 20,
],
]);
$this->load($params);
if (!empty($this->keyword) || $this->keyword != '') {
$query->andWhere('MATCH(task_title,task_description, priority) AGAINST("' . $this->keyword . '*" IN BOOLEAN MODE)');
}
if (is_numeric($this->filterby)) {
$query->andFilterWhere(['status' => $this->filterby]);
} else {
$query->andFilterWhere(['IN', 'status', [10, 4, 2]]);
}
if (!empty($this->sortby) || $this->sortby != '') {
$query->orderBy($this->sortby . ' DESC');
}
return $dataProvider;
}`
whenever I deployed the same code to production it shows 0 counts
if it works fine in one environment it is unlikely a bug in yii. also your code has several issues that may allow SQL injection. please fix that first and report back if there is a way to reproduce the problem.
Problem with this line
$query->where(['q_id' => $this->job_id]);
When I comment this line all are working fine. while I checked $this->job_id produces a value.
Only the difference is in my controller used like this
$searchModel->job_id = Yii::$app->request->get('job_id');
without formName. My other prams are coming with Model ie. formName like XyzModel['keywor'] while job is dont have it.
|
2025-04-01T04:36:00.698123
| 2017-01-11T16:02:26
|
200133326
|
{
"authors": [
"Alex-Code",
"PELock",
"SDKiller",
"alex-code",
"dynasource",
"flaviovs",
"samdark"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12220",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/issues/13373"
}
|
gharchive/issue
|
Ability to set formName in model
Can we add the option to set the formName of a model?
I'm currently using a trait but would be handy for it to be built in.
Do you mean input?
$model1 = new MyModel;
$model1->formName = 'MyChosenName';
echo $model1->formName; //Outputs 'MyChosenName'
It was already rejected 2 years ago ;(
https://github.com/yiisoft/yii2/issues/6090
No. If you want it, add it as a public class property yourself.
@Alex-Code, what exactly is your use case for a dynamic formName?
I've got a custom behaviour that handles model relations so when creating an ActiveForm I needed it to be able to set the formName.
I second that request. When you have more than one instance of the same model class in the same HTML form, it is not possible to distinguish (i.e. to ->load()) them without a dummy subclass. I'm not talking about tabular input here, but things like "mainEmail", "reminderEmail", "followupEmail" being on the same form, all being of the same Email model, in which case $form->field($model, 'attribute') will generate Email[attribute] names for all different instances. I understand that adding another properties to Model is not desirable, but IMHO from an architectural perspective the "form name" property is an instance attribute (defaulting to the class name), not a class one, so this might justify the addition.
I have managed to workaround this by adding the trait below to models I anticipate might be put together with others of the same class in one single page, but it would be nice if we get some support from the framework for this.
trait FormNameModelTrait
{
protected $_formName;
public function formName()
{
return $this->_formName ?: parent::formName();
}
public function setFormName($name)
{
$this->_formName = $name;
}
}
This is so frustrating, having 2 forms of the same type and trying to load them in one controller action. Why not make it easier for Yii2 devs? My simplest solution is to use extra model property and override the formName() method
public $_formName;
public function formName()
{
return $this->_formName;
}
The best and logical solution would be to set the form name in its constructor
$model_form_variant_a = new ModelForm("form-name-a");
$model_form_variant_b = new ModelForm("form-name-b");
This is so frustrating, having 2 forms of the same type and trying to load them in one controller action. Why not make it easier for Yii2 devs? My simplest solution is to use extra model property and override the formName() method
public $_formName;
public function formName()
{
return $this->_formName;
}
The best and logical solution would be to set the form name in its constructor
$model_form_variant_a = new ModelForm("form-name-a");
$model_form_variant_b = new ModelForm("form-name-b");
The stubbornness of Yii2 devs leads ppl to Laravel ;)
|
2025-04-01T04:36:00.700664
| 2015-02-09T22:26:43
|
57099869
|
{
"authors": [
"InteLigent",
"qiangxue"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12221",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/issues/7218"
}
|
gharchive/issue
|
Incorrect response format type for CaptchaAction. Captcha not refreshing by click on it.
Problem in line
https://github.com/yiisoft/yii2/blob/master/framework/captcha/CaptchaAction.php#L124
Because action return already formatted data at json. In summary, data twice encode in formatter JsonResponseFormatter.
Simple solution add code:
Yii::$app->response->format = Response::FORMAT_RAW;
before line https://github.com/yiisoft/yii2/blob/master/framework/captcha/CaptchaAction.php#L124
Thanks!
|
2025-04-01T04:36:00.703408
| 2015-04-07T07:01:25
|
66799403
|
{
"authors": [
"gonimar",
"klimov-paul",
"mikehaertl",
"samdark"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12222",
"repo": "yiisoft/yii2",
"url": "https://github.com/yiisoft/yii2/pull/8001"
}
|
gharchive/pull-request
|
Fix migrate/down order
Migrations were applied at the same time. This cause invalid sort order while downgrade.
Command yii migrate
Output:
Yii Migration Tool (based on Yii v2.0.4-dev)
Total 2 new migrations to be applied:
m150402_075534_initial
m150402_075544_feedback
Command yii migrate/down 2
Yii Migration Tool (based on Yii v2.0.4-dev)
Total 2 migrations to be reverted:
m150402_075534_initial
m150402_075544_feedback
In databasse
"m000000_000000_base";1428382302
"m150402_075534_initial";1428388855 <----
"m150402_075544_feedback";1428388855 <----
Or sleep(1); :)
Relates #7924
I'd suggest to modify the order to ->orderBy('apply_time DESC, version DESC')
Thanks!
|
2025-04-01T04:36:00.756687
| 2024-01-08T13:08:06
|
2070425567
|
{
"authors": [
"LiborBatek",
"mkolar",
"moonyuet"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12223",
"repo": "ynput/OpenPype",
"url": "https://github.com/ynput/OpenPype/pull/6110"
}
|
gharchive/pull-request
|
Max: Implementation of Camera Attributes Validator
Changelog Description
Implement Validate Camera Attributes in camera family in Max host
Additional info
This validator is not eligible for the camera created from the external plugins such as Redshift and Arnold etc.
When the values of the camera attributes in the setting set to 0, the validator would skip checking the attribute.
Testing notes:
Launch Max via launcher
Tweak the settings ayon+settings://max/publish/ValidateCameraAttributes or project_settings/max/publish/ValidateCameraAttributes
Create Camera Instance
Publish
If the camera attributes dont align with those attributes in the setting, the validator would error out.
When tested with Near and Far ranges for enviro and clipping I have been able to Publish camera successfully even with wrong values...maybe when not turned on those checkers for Show and Clip Manually ??
however next time trying publishing the Validator kicked in correctly but again the messaging could be strongly improved imho
Not sure whats the origin of this PR but right now as we have it theres is no way to validate just a few of the values but all together.... theres no way to selectively choose what to validate or at least to disable particular value?
Maybe would be option to use 0 Value to disable each entry? Something like validating against zero means skip this value in validation ??
the initial purpose of this PR is to build the Camera Attributes validator in Max which is similar to Maya. But the Maya one is more hard-coded and I am not sure it will be good idea for Max. But I agree with you that we can just skip the check when the parameter set to zero
@LiborBatek I have updated the commit and the changelog.
The validation should check if the camera attributes aligning with the settings.
You may find the video below helpful.
https://github.com/ynput/OpenPype/assets/64118225/5c1b972f-a6be-4996-82e7-21945b6e5ee3
Do you think that it's also good idea to add repair action for the camera attributes?
Do you think that it's also good idea to add repair action for the camera attributes?
Yeah would be nice! ...Did u manage to incorporate those zero values for skipping the validation or not yet?
Yeah would be nice!
Okay I will add it into the next commit
.Did u manage to incorporate those zero values for skipping the validation or not yet?
Yes, it was already there in the updated commit.
So there are 2 issues now, question being if we should take into account Scene Units or just note it in the OP Settings via docstring that those Values being in generic units and some conversion gonna happen if not using the same.
Yes I think we can mention that those default values are being in generic units in both settings and validators, so that the user can change the value as their own default in regard to what they set in scene units. (we can possibly add one more check on the scene units to give this message too)
The Validation issue seems tied to the numbers format and invalid logic for decimal values?? Now just guessing tho.
I need to double check, it is supposed that it tied to the float format.
Looks good for merge. @moonyuet port to ayon pls
Looks good for merge. @moonyuet port to ayon pls
Already ported to ayon-core!
|
2025-04-01T04:36:00.763219
| 2024-05-20T13:07:48
|
2305965445
|
{
"authors": [
"BigRoy",
"LiborBatek"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12224",
"repo": "ynput/ayon-core",
"url": "https://github.com/ynput/ayon-core/issues/529"
}
|
gharchive/issue
|
Blender: Load pointcache loses transformation on top group/empty
Is there an existing issue for this?
[X] I have searched the existing issues
Current Behavior:
This logic here makes it so that the objects are re-parented to an 'asset group' but also make it so that any transformation on the loaded top group is lost.
Expected Behavior:
Data should not be lost, and preferably any transform or data in the published output is retained on load.
Version
1.0.0
What platform you are running on?
Windows
Steps To Reproduce:
Create scene with cube
Parent the cube to an empty
Move the empty that has the cube as child.
Publish the empty as a pointcacheMain
Load the pointcache
Now the transformation is lost on load.
Are there any labels you wish to add?
[X] I have added the relevant labels to the bug report.
Relevant log output:
No response
Additional context:
No response
@LiborBatek can you confirm. @simonebarbieri is there any reason to have load work this way?
@BigRoy yes, its breaking the transforms if the asset got version update via manage tool.
There is a way to not scratch the alembic asset completely when changing version via changing just the file path leading to the abc file instead within the pointcache modifier. Something which we have discussed it would make much more sense but never happened.
I could give more info or record some vid describing the issue if anyone interested...
@BigRoy hmm, reading the issue again Im not sure if this PR refers to loosing any local transform edits on the loaded asset when updated version later on or something else?
|
2025-04-01T04:36:00.765751
| 2024-04-10T09:35:56
|
2235159282
|
{
"authors": [
"BigRoy",
"iLLiCiTiT"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12225",
"repo": "ynput/ayon-core",
"url": "https://github.com/ynput/ayon-core/pull/396"
}
|
gharchive/pull-request
|
Ruff: Add list of files to exclude from linting
Changelog Description
Added list of files that should be ignored for ruff when.
Additional information
PR related to https://github.com/ynput/ayon-core/pull/196 .
Testing notes:
Validate if excluded paths make sense to be ignored.
Validate that ruff is actually ignoring them.
What is the reason that we're ignoring some instead of fixing them?
What is the reason that we're ignoring some instead of fixing them?
Please do if you have spare time.
|
2025-04-01T04:36:00.771705
| 2019-01-12T05:44:09
|
398524042
|
{
"authors": [
"djmattyg007",
"myw"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12226",
"repo": "yodle/docker-registry-client",
"url": "https://github.com/yodle/docker-registry-client/pull/54"
}
|
gharchive/pull-request
|
Add blob copy capability
Meant to be merged in after #53.
This PR adds a method to the V2 BaseClient: copy_blob. The method takes advantage of the Registry API's cross-repository blob mount feature to copy blobs that already exist in the registry from one repo to another without actually downloading and re-uploading them.
A unit test is included.
Other minor changes in this PR:
add a check_manifest method to test for the existence of a manifest
resolve the unit tests failing with HTTP 500 codes on put_manifest (Root cause was a Docker Registry config issue in testing)
clean up a minor issue conflating tags and references in put_manifest: now consumers can put manifests using their digest as a reference without updating the tag to be the same as the digest (which was the former, erroneous behavior)
add a manifest digestion utility function (manifest.digest) to enable users to PUT manifests by digest.
Hello. Since this package appears to be abandoned, I forked it and made several improvements. This includes dropping support for v1 registries, and introducing first-class support for v2 manifests (incluidng multi-arch manifests).
If you're still feeling up to it, I'd be happy to review an updated PR against my fork:
https://github.com/djmattyg007/dreg-client
|
2025-04-01T04:36:00.801504
| 2019-08-03T14:28:45
|
476456693
|
{
"authors": [
"snobu",
"yokawasa"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12227",
"repo": "yokawasa/azure-functions-python-samples",
"url": "https://github.com/yokawasa/azure-functions-python-samples/pull/26"
}
|
gharchive/pull-request
|
Added Service Bus Queue sample for trigger and output
See https://github.com/yokawasa/azure-functions-python-samples/issues/25
@snobu Thank you so much for the PR!
|
2025-04-01T04:36:00.812182
| 2016-10-07T02:13:05
|
181571259
|
{
"authors": [
"jekrb",
"yoshuawuyts"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12228",
"repo": "yoshuawuyts/cache-element",
"url": "https://github.com/yoshuawuyts/cache-element/pull/2"
}
|
gharchive/pull-request
|
begin implementation
ref #1
I'll be getting back on the rest of this tonight! Feel free to make comments where you feel there should be improvements.
Added some comments; hope this makes sense!
@jekrb heya, how's it going? Did my feedback make sense? Is there any way I can help?
Hey @yoshuawuyts! Sorry for the late response. The feedback makes loads of sense and I'm so happy you've provided it. I feel really bad with not carrying this on as quickly as I hoped initially. My day job has been taking up more of my mental capital than anticipated lately 😆
If you want to carry this on faster I won't take any hard feelings. Otherwise I'd like build this out over the weekend
merged in #4 - thanks!
Awesome! Good luck on the talk!
|
2025-04-01T04:36:00.843514
| 2015-10-16T07:01:50
|
111773398
|
{
"authors": [
"alainjobart",
"enisoc"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12229",
"repo": "youtube/vitess",
"url": "https://github.com/youtube/vitess/pull/1224"
}
|
gharchive/pull-request
|
Use gRPC for PHP client.
@alainjobart
Also implement cursor and canonical error translation.
LGTM really cool.
|
2025-04-01T04:36:00.869471
| 2024-10-30T07:34:46
|
2623159108
|
{
"authors": [
"wfondrie",
"ypriverol"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12230",
"repo": "ypriverol/os-proteomics",
"url": "https://github.com/ypriverol/os-proteomics/pull/3"
}
|
gharchive/pull-request
|
Some major changes
@wfondrie
I have merge some of your two original sections. Please feel free to split them if you don't like it.
I also add a like on top of sentences that I feel maybe the idea is repeated.
convert the document to a line-based approach. Better for review.
Also some minor changes.
Let me know what do you think?
@ypriverol, this looks good to me although I found plenty of things to edit in the parts I've contributed 🙃
@timosachsenberg, do you want to put in a PR updating the introduction with your suggestions?
|
2025-04-01T04:36:00.871718
| 2016-06-16T22:58:25
|
160779470
|
{
"authors": [
"PSesto",
"yqritc"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12231",
"repo": "yqritc/RecyclerView-FlexibleDivider",
"url": "https://github.com/yqritc/RecyclerView-FlexibleDivider/issues/40"
}
|
gharchive/issue
|
shouldHideDivider not working
when set to true, divider is still showing
I can confirm that my adapter is implementing FlexibleDividerDecoration.VisibilityProvider and overriding shouldHideDivider. I can also confirm that shouldHideDivider is not being called
@PSesto
Please check my sample project.
Complex Divider List (ComplexActivity) implements VisibilityProvider.
I just checked the sample and it works well, so please check your implementation.
You're right. I didn't see the ".visibilityProvider(adapter)" line in ComplexActivity.
Thank you so much for the quick response :)
|
2025-04-01T04:36:00.905656
| 2023-02-15T03:15:46
|
1585143124
|
{
"authors": [
"Haiyao-Nero",
"ysharma1126"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12233",
"repo": "ysharma1126/ssl_identifiability",
"url": "https://github.com/ysharma1126/ssl_identifiability/issues/3"
}
|
gharchive/issue
|
The performance evaluation for the numerical data is not good as the performance during the training process.
Hi, there is a problem when I run the code for numerical data. the $R^2$ for evaluation is much lower than that during the training process. The reason may be that the decoder $f$ in the evaluation is different from the $f$ in the training. But only the trained $g$ is saved. Is this a mistake or the $f$ s could be different?
Hi, you need to specify the seed, or else you're evaluating the trained model on a different data generating process.
Thanks, got it.
|
2025-04-01T04:36:00.907020
| 2022-01-30T06:26:19
|
1118434170
|
{
"authors": [
"yshavit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12234",
"repo": "yshavit/whatdid",
"url": "https://github.com/yshavit/whatdid/issues/270"
}
|
gharchive/issue
|
end-of-day report breaks if too many entries
This breaks the window, and only a restart really fixes it. Yikes!
Instead of even trying to fit them in, I should instead prompt the user to open them up in a separate window (#269).
Dupe of #241
|
2025-04-01T04:36:00.923375
| 2024-07-13T04:29:30
|
2406695394
|
{
"authors": [
"nodaguti"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12235",
"repo": "ysugimoto/falco",
"url": "https://github.com/ysugimoto/falco/pull/332"
}
|
gharchive/pull-request
|
Set exit code to 1 when there are one or more lint errors
The current linter quits with exit code 0 even when there are some errors. This behaviour is not very useful especially when you are using falco lint in your CI because without special cares of [ERROR] messages a lint job will pass regardless of number of errors reported.
This PR changes the exit code to 1 when the linter reports one or more errors.
There might be another discussion whether the exit code should also reflect number of warnings and recommendations. In my opinion, providing similar options to ESLint's --max-warnings would be nice, i.e., adding --max-warnings and --max-infos to set warning and recommendation thresholds respectively. (I will work on this if you agree with this idea.)
@ysugimoto
Thank you for your review! I've addressed your comments.
|
2025-04-01T04:36:00.924912
| 2017-04-24T07:23:52
|
223725806
|
{
"authors": [
"samuelwilliams",
"yswery"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12236",
"repo": "yswery/PHP-DNS-SERVER",
"url": "https://github.com/yswery/PHP-DNS-SERVER/pull/24"
}
|
gharchive/pull-request
|
Added tests for the server class.
Started on creating unit tests for the Server class. Since pretty much all the methods are private, I have proxied the Server class in order to access the methods directly.
This is not an ideal solution, but it enables deeper testing (and therefore safer refactoring) without hacking the API.
As I raised in issue #23, it would be better to separate the encoder and decoder methods.
A nice simply start for the testing the server!
|
2025-04-01T04:36:00.928303
| 2021-09-17T15:20:29
|
999471167
|
{
"authors": [
"jiatern",
"pukkandan"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12237",
"repo": "yt-dlp/yt-dlp",
"url": "https://github.com/yt-dlp/yt-dlp/issues/1011"
}
|
gharchive/issue
|
[Question] CryptographyDeprecationWarning
Checklist
[x] I'm asking a question
[x] I've looked through the README and FAQ for similar questions
[x] I've searched the bugtracker for similar questions including closed ones
Question
I am getting the following errors whenever I run any yt-dlp commands on my Ubuntu machine:
/usr/lib/python3/dist-packages/secretstorage/dhcrypto.py:15: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead ````
from cryptography.utils import int_from_bytes
/usr/lib/python3/dist-packages/secretstorage/util.py:19: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead
from cryptography.utils import int_from_bytes
I am not an expert in coding or scripting, I only know the basics on how to run scripts as-is. I am not sure if the above errors are caused by other dependencies or non yt-dlp components.
Is there anything I should do to fix this? I have updated yt-dlp to the latest 2021.09.02 version.
Show full verbose log
Show full verbose log
Here you go:
/usr/lib/python3/dist-packages/secretstorage/dhcrypto.py:15: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead
from cryptography.utils import int_from_bytes
/usr/lib/python3/dist-packages/secretstorage/util.py:19: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead
from cryptography.utils import int_from_bytes
[debug] Command-line config: ['-v']
[debug] Encodings: locale UTF-8, fs utf-8, out utf-8, pref UTF-8
[debug] yt-dlp version 2021.09.02
[debug] Python version 3.8.5 (CPython 64bit) - Linux-5.4.0-28-generic-x86_64-with-glibc2.29
[debug] exe versions: ffmpeg 4.2.4, ffprobe 4.2.4, rtmpdump 2.4
[debug] Optional libraries: keyring, mutagen, pycryptodome, sqlite, websockets
[debug] Proxy map: {}
The same error occurs with any yt-dlp commands.
|
2025-04-01T04:36:00.940023
| 2024-12-12T22:28:02
|
2736997222
|
{
"authors": [
"AnimeHeaven",
"bashonly",
"october262"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12238",
"repo": "yt-dlp/yt-dlp",
"url": "https://github.com/yt-dlp/yt-dlp/issues/11794"
}
|
gharchive/issue
|
Animation Digital Network
DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE
[X] I understand that I will be blocked if I intentionally remove or skip any mandatory* field
Checklist
[X] I'm reporting that yt-dlp is broken on a supported site
[X] I've verified that I have updated yt-dlp to nightly or master (update instructions)
[X] I've checked that all provided URLs are playable in a browser with the same IP and same login details
[X] I've checked that all URLs and arguments with special characters are properly quoted or escaped
[X] I've searched known issues and the bugtracker for similar issues including closed ones. DO NOT post duplicates
[X] I've read the guidelines for opening an issue
[X] I've read about sharing account credentials and I'm willing to share it if required
Region
Germany
Provide a description that is worded well enough to be understood
I have tried to download a video (https://animationdigitalnetwork.com/de/video/1233-loner-life-in-another-world) where it is in Japanese and have German subtitle. The Subtitles would be fine downloaded and mark in the metadata as German. But the audio is shown as ENGLISH, what is completely wrong.
I have tried some different commands like:
Japanese: %(meta_language)s" --embed-metadata -f [language=jap]
japanese: %(meta_language)s" --embed-metadata -f [language=jap]
Japanese: %(meta_language)s" --embed-metadata -f [language=jpn]
or
Japanese: %(meta_language)s" --embed-metadata -f [language=jp]
but nothing will write the audio metadata correct to japanese ,
Provide verbose output that clearly demonstrates the problem
[X] Run your yt-dlp command with -vU flag added (yt-dlp -vU <your command line>)
[ ] If using API, add 'verbose': True to YoutubeDL params instead
[X] Copy the WHOLE output (starting with [debug] Command-line config) and insert it below
Complete Verbose Output
[debug] Command-line config: ['-vU', '-u', 'PRIVATE', '-p', 'PRIVATE', 'japanese:', '%(meta_language)s --embed-metadata -f [language=jpn] --sub-langs de --write-subs --sub-format ssa -v -a C:\\Users\\Stefan\\Desktop\\adn.txt -o Z:\\Anime', 'Serien\\Loner', 'Life', 'in', 'Another', 'World\\Loner', 'Life', 'in', 'Another', 'WorldS%(season_number)02dE%(episode_number)02d.%(ext)s']
[debug] Encodings: locale cp1252, fs utf-8, pref cp1252, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version<EMAIL_ADDRESS>from yt-dlp/yt-dlp [4bd265539] (win_exe)
[debug] Python 3.10.11 (CPython AMD64 64bit) - Windows-10-10.0.22631-SP0 (OpenSSL 1.1.1t 7 Feb 2023)
[debug] exe versions: ffmpeg 7.1-full_build-www.gyan.dev (setts), ffprobe 7.1-full_build-www.gyan.dev
[debug] Optional libraries: Cryptodome-3.21.0, brotli-1.1.0, certifi-2024.08.30, curl_cffi-0.5.10, mutagen-1.47.0, requests-2.32.3, sqlite3-3.40.1, urllib3-2.2.3, websockets-14.1
[debug] Proxy map: {}
[debug] Request Handlers: urllib, requests, websockets, curl_cffi
[debug] Loaded 1837 extractors
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
Latest version<EMAIL_ADDRESS>from yt-dlp/yt-dlp
yt-dlp is up to date<EMAIL_ADDRESS>from yt-dlp/yt-dlp)
[generic] Extracting URL: japanese:
[generic] japanese:: Downloading webpage
ERROR: Unable to handle request: Unsupported url scheme: "japanese" (requests, urllib, websockets, curl_cffi)
Traceback (most recent call last):
File "yt_dlp\YoutubeDL.py", line 1624, in wrapper
File "yt_dlp\YoutubeDL.py", line 1759, in __extract_info
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2393, in _real_extract
File "yt_dlp\extractor\common.py", line 898, in _request_webpage
File "yt_dlp\YoutubeDL.py", line 4162, in urlopen
File "yt_dlp\networking\common.py", line 130, in send
yt_dlp.networking.exceptions.NoSupportingHandlers: Unable to handle request: Unsupported url scheme: "japanese" (requests, urllib, websockets, curl_cffi)
[generic] Extracting URL: %(meta_language)s --embed-metadata -f [language=jpn] --sub-langs de --write-subs --sub-format ssa -v -a C:\Users\Stefan\Desktop\adn.txt -o Z:\Anime
ERROR: [generic] '%(meta_language)s --embed-metadata -f [language=jpn] --sub-langs de --write-subs --sub-format ssa -v -a C:\\Users\\Stefan\\Desktop\\adn.txt -o Z:\\Anime' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:%(meta_language)s --embed-metadata -f [language=jpn] --sub-langs de --write-subs --sub-format ssa -v -a C:\Users\Stefan\Desktop\adn.txt -o Z:\Anime" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: Serien\Loner
ERROR: [generic] 'Serien\\Loner' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:Serien\Loner" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: Life
ERROR: [generic] 'Life' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:Life" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: in
ERROR: [generic] 'in' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:in" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: Another
ERROR: [generic] 'Another' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:Another" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: World\Loner
ERROR: [generic] 'World\\Loner' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:World\Loner" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: Life
ERROR: [generic] 'Life' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:Life" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: in
ERROR: [generic] 'in' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:in" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: Another
ERROR: [generic] 'Another' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:Another" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
[generic] Extracting URL: WorldS%(season_number)02dE%(episode_number)02d.%(ext)s
ERROR: [generic] 'WorldS%(season_number)02dE%(episode_number)02d.%(ext)s' is not a valid URL. Set --default-search "ytsearch" (or run yt-dlp "ytsearch:WorldS%(season_number)02dE%(episode_number)02d.%(ext)s" ) to search YouTube
File "yt_dlp\extractor\common.py", line 742, in extract
File "yt_dlp\extractor\generic.py", line 2361, in _real_extract
does it work if you use yt-dlp -s jp URL of video
It looks like there's a quoting issue with your arguments
Nope it works well with german as language.
does it work if you use yt-dlp -s jp URL of video
No, it makes this error: ERROR: [generic] 'jp' is not a valid URL.
Yeah because your quoting is wrong
|
2025-04-01T04:36:00.945292
| 2021-03-06T20:36:50
|
823741555
|
{
"authors": [
"pukkandan",
"werid"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12239",
"repo": "yt-dlp/yt-dlp",
"url": "https://github.com/yt-dlp/yt-dlp/issues/151"
}
|
gharchive/issue
|
When using --allow-unplayable-formats but not --fixup, yt-dlp still outputs the warning that --fixup was ignored
Checklist
[ ] I'm reporting a broken site support issue
[x] I've verified that I'm running yt-dlp version 2<IP_ADDRESS>
[x] I've checked that all provided URLs are alive and playable in a browser
[x] I've checked that all URLs and arguments with special characters are properly quoted or escaped
[x] I've searched the bugtracker for similar bug reports including closed ones
[x] I've read bugs section in FAQ
Verbose log
[debug] Command-line config: ['--verbose', '--cookies', '/home/weird/cookies.txt', '--allow-unplayable-formats', '-F', 'https://www.facebook.com/478492418989679/videos/1320343201817221']
WARNING: --fixup is ignored since --allow-unplayable-formats was given
[debug] Loading archive file None
[debug] Encodings: locale UTF-8, fs utf-8, out UTF-8, pref UTF-8
[debug] yt-dlp version 2<IP_ADDRESS> (zip)
[debug] Python version 3.5.3 (CPython 64bit) - Linux-4.9.0-14-amd64-x86_64-with-debian-9.13
[debug] exe versions: ffmpeg N-54690-g682990a849-static, ffprobe N-54690-g682990a849-static, rtmpdump 2.4
[debug] Proxy map: {}
[facebook]<PHONE_NUMBER>817221: Downloading webpage
[debug] Sort order given by user:
[debug] Formats sorted by: hasvid, ie_pref, lang, quality, res, fps, vcodec:vp9.2(10), acodec, filesize, fs_approx, tbr, vbr, abr, asr, proto, vext, aext, hasaud, source, id
[download] Downloading playlist:<PHONE_NUMBER>817221
[facebook] playlist<PHONE_NUMBER>817221: Collected 1 video ids (downloading 1 of them)
[download] Downloading video 1 of 1
[info] Available formats for<PHONE_NUMBER>817221:
ID EXT RESOLUTION | TBR PROTO | VCODEC VBR ACODEC ABR ASR NOTE
---------------- --- ---------- - ----- ----- - ----------- ----- --------- ---- ------- -----------------------------
486634785670353a m4a audio only | 135k https | mp4a.40.2 135k 48000Hz DASH audio, m4a_dash, 48000Hz
918757108893872v mp4 256x144 | 112k https | avc1.64000c 112k DASH video, mp4_dash
774603623172819v mp4 426x240 | 267k https | avc1.640015 267k DASH video, mp4_dash
729443354422455v mp4 640x360 | 527k https | avc1.64001e 527k DASH video, mp4_dash
526187182114431v mp4 854x480 | 1045k https | avc1.64001f 1045k DASH video, mp4_dash
240211284431385v mp4 1280x720 | 2082k https | avc1.640020 2082k DASH video, mp4_dash
[download] Finished downloading playlist:<PHONE_NUMBER>817221
Description
Minor bug, but annoying:
When using --allow-unplayable-formats but not --fixup, yt-dlp still outputs the warning that --fixup was ignored.
It also appears to be missing the color, WARNING: is usually in yellow.
When using --allow-unplayable-formats but not --fixup, yt-dlp still outputs the warning that --fixup was ignored.
The default of fixup is detect_or_warn, which is why the message is shown. Will fix it
It also appears to be missing the color, WARNING: is usually in yellow.
The warning is issued before the options are fully parsed and so we don't know whether or not to output color. It would be quite difficult to fix this
|
2025-04-01T04:36:00.960918
| 2021-08-03T10:18:16
|
958961469
|
{
"authors": [
"Ashish0804",
"Grub4K",
"fstirlitz",
"pukkandan"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12240",
"repo": "yt-dlp/yt-dlp",
"url": "https://github.com/yt-dlp/yt-dlp/issues/620"
}
|
gharchive/issue
|
[Feature Request] Streamlined bug reporting
I don’t like this project’s issue templates. They abuse GitHub’s checkboxes feature (meant for progress tracking) for bug reporting checklists. And the worst part is that they aren’t even particularly effective at what they are supposed to do. People routinely screw up bug reports by:
ignoring the template entirely
leaving out the console output
posting only part of the console output
posting screenshots of the console output
forgetting to use the --verbose option
leaving out even which site they have a problem with
leaving in placeholder text like ‘WRITE DESCRIPTION HERE’
putting the x character in brackets next to the space instead of replacing the space, breaking checkbox rendering
using Unicode checkmarks ‘✔’ instead of an ASCII x, breaking checkbox rendering
And as people are flocking in here from upstream, which is apparently completely dead now (they haven’t had a commit for a month), this is only going to get worse. So I think this should be addressed promptly to avoid lots of wasted time, both on the maintainers’ part and reporters’.
So my idea is to add a --bug-report option that:
automatically checks if yt-dlp is at the latest version
automatically turns on --verbose (or even an extra-verbose mode that also lists all available formats)
captures all console output into a variable
uses http.server to serve up a simple HTML page which contains all captured information so far and a questionnaire for the user
uses webbrowser to open this HTML page in the user’s browser, where the report is validated and then posted to GitHub
(Alternatively, we could use Tkinter instead opening a browser, but we would still have to open a web browser eventually to allow the user to log in and submit the report to GitHub.)
A pre-formatted bug report can be sent to GitHub in a GET parameter, e.g. https://github.com/yt-dlp/yt-dlp/issues/new?title=foo&body=bar. This allows about 8K of data to fit in. We could try to find a workaround if that proves too tight; in the worst case, we could copy the markup into the clipboard and tell the user to paste it into the issue report form.
The resulting report would have all the boilerplate hidden inside a <details> tag, to make it less obstructive. We could then direct people to use the --bug-report option to report bugs instead of opening GitHub directly and messing with issue templates.
We could also have a GitHub Action script that checks if issue reports have all the necessary boilerplate in, and promptly asks the user to fill in missing information.
--bug-report
Okay so what do we do when people are not following this and pasting screenshots like always?
A pre-formatted bug report can be sent to GitHub in a GET parameter
Umm we will shit tons of DRM issues and network issues.
We could also have a GitHub Action script that checks if issue reports have all the necessary boilerplate in, and promptly asks the user to fill in missing information.
Idk how cheap and fast it's going to be but its the best option.
Considering many people don't reply after opening an issue, we could outright close it aswell and let genuine users open another issue with proper info.
Okay so what do we do when people are not following this and pasting screenshots like always?
We tell them ‘please re-file using --bug-report or follow the bug reporting checklist over here’. Right now we only have the checklist, with umpteenth steps the user has to follow and can therefore screw up. A bug reporting mode within the program can reduce it to just four: enable the mode, fill out the questionnaire, log into GitHub, submit to the tracker.
When a DRM-encumbered stream is detected, the reporting form can display a relevant warning — and more importantly, not display warnings that are irrelevant to the situation at hand and that the user is just going to glaze over thinking ‘TL;DR’. Network issues can be detected similarly: it’s a rather specific kind of error (timeout, connection reset by peer, etc.). The bug report form can suggest to only file retry later.
Users who want to spam the issue tracker with spurious reports can already do so — while skipping the template and therefore not even conveying that the issue is out of scope for this project!
I would be against installing a GitHub Actions hook alone, because I expect that is just going to frustrate users, especially when the Actions hook fails to parse legitimately filled-out fields (compare how [x] tends to be screwed up). We should direct users to file informative bug reports in the first place, not have a robot play whack-a-mole by closing incomplete ones. We do that by making the correct action less burdensome: users have only so much energy and attention to spare.
Might be helpful: https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/syntax-for-issue-forms
I guess someone had the same idea for upstream https://github.com/ytdl-org/youtube-dl/issues/29821
The original suggestion of a fully automated bug reporting is a bit too much imo and I am not sure it would be worth the work that will need to be put into implementing and maintaining it.
On the other hand, replacing our issue templates with forms is a great idea. It is much easier to implement, will somewhat simplify the end user experience and we get some amount of validation for free
The issue templates have noticeably reduced the amount of invalid issues we get 👍
Closing this since as mentioned above, I think a fully automated bug reporting is overkill and out of scope
automatically checks if yt-dlp is at the latest version
I agree with this. Maybe we could check known issues and direct people to use the daily builds (or nighty once exists) if the error is listed there as solved, and if it is a common error list the way to solve it directly in the commandline.
automatically turns on --verbose (or even an extra-verbose mode that also lists all available formats)
captures all console output into a variable
I think this is best implemented after #5680. We can easily employ a custom Logger class which can capture the output and analyze it later. This logger can also filter out the bug report message and write out better messages instead. If we decide on what it should do its almost trivial to implement.
Automatically posting the report is something that I would not be doing (Reference) however automatically preparing the report would work quite well. Spinning up a simple http server might be overkill. For ease of implementation we could copy it to the clipboard instead, and render information after execution in terminal. Then, if that works, go from there ot expand it later on.
|
2025-04-01T04:36:00.973430
| 2023-12-15T19:42:59
|
2044285219
|
{
"authors": [
"Pranaxcau",
"bashonly",
"meoyawn",
"pukkandan",
"sysop84"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12241",
"repo": "yt-dlp/yt-dlp",
"url": "https://github.com/yt-dlp/yt-dlp/issues/8782"
}
|
gharchive/issue
|
Rumble /c/ --playlist-items with --print broken as of today, /user/ still works fine
DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE
[X] I understand that I will be blocked if I intentionally remove or skip any mandatory* field
Checklist
[X] I'm reporting that yt-dlp is broken on a supported site
[X] I've verified that I have updated yt-dlp to nightly or master (update instructions)
[X] I've checked that all provided URLs are playable in a browser with the same IP and same login details
[X] I've checked that all URLs and arguments with special characters are properly quoted or escaped
[X] I've searched known issues and the bugtracker for similar issues including closed ones. DO NOT post duplicates
[X] I've read the guidelines for opening an issue
[X] I've read about sharing account credentials and I'm willing to share it if required
Region
USA
Provide a description that is worded well enough to be understood
--playlist-items no longer works with a rumble /c/ page, still works fine with a /user/ page. Broken as of this morning. --playlist-items is ignored on a rumble /c/ page, yt-dlp grabs ALL pages. Previous behavior was correct. Two testing urls below:
./yt-dlp-2023-11-16 -vU --simulate --no-warnings --ignore-errors --socket-timeout 3 --print "%(timestamp)s|||%(id)s|||%(duration_string)s|||%(title)s|||%(thumbnail)s" --playlist-items 1:5 https://rumble.com/c/TheAlexJonesShow/videos
User url for testing: https://rumble.com/user/TommyRobinsonOfficial
Provide verbose output that clearly demonstrates the problem
[X] Run your yt-dlp command with -vU flag added (yt-dlp -vU <your command line>)
[X] If using API, add 'verbose': True to YoutubeDL params instead
[X] Copy the WHOLE output (starting with [debug] Command-line config) and insert it below
Complete Verbose Output
[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version<EMAIL_ADDRESS>from yt-dlp/yt-dlp [24f827875] (zip)
[debug] Python 3.11.4 (CPython x86_64 64bit) - Linux-4.19.0-18-amd64-x86_64-with-glibc2.28 (OpenSSL 1.1.1d 10 Sep 2019, glibc 2.28)
[debug] exe versions: none
[debug] Optional libraries: sqlite3-3.27.2
[debug] Proxy map: {}
[debug] Request Handlers: urllib
[debug] Loaded 1901 extractors
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
Latest version<EMAIL_ADDRESS>from yt-dlp/yt-dlp
yt-dlp is up to date<EMAIL_ADDRESS>from yt-dlp/yt-dlp)
[RumbleChannel] Extracting URL: https://rumble.com/c/TheAlexJonesShow/videos
[download] Downloading playlist: TheAlexJonesShow
[RumbleChannel] TheAlexJonesShow: Downloading page 1
[RumbleChannel] TheAlexJonesShow: Downloading page 2
[RumbleChannel] TheAlexJonesShow: Downloading page 3
[RumbleChannel] TheAlexJonesShow: Downloading page 4
[RumbleChannel] TheAlexJonesShow: Downloading page 5
[RumbleChannel] TheAlexJonesShow: Downloading page 6
// /user/ page:
[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version<EMAIL_ADDRESS>from yt-dlp/yt-dlp [24f827875] (zip)
[debug] Python 3.11.4 (CPython x86_64 64bit) - Linux-4.19.0-18-amd64-x86_64-with-glibc2.28 (OpenSSL 1.1.1d 10 Sep 2019, glibc 2.28)
[debug] exe versions: none
[debug] Optional libraries: sqlite3-3.27.2
[debug] Proxy map: {}
[debug] Request Handlers: urllib
[debug] Loaded 1901 extractors
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
Latest version<EMAIL_ADDRESS>from yt-dlp/yt-dlp
yt-dlp is up to date<EMAIL_ADDRESS>from yt-dlp/yt-dlp)
[RumbleChannel] Extracting URL: https://rumble.com/user/TommyRobinsonOfficial
[download] Downloading playlist: TommyRobinsonOfficial
[RumbleChannel] TommyRobinsonOfficial: Downloading page 1
[RumbleChannel] Playlist TommyRobinsonOfficial: Downloading 5 items
[download] Downloading item 1 of 5
[Rumble] Extracting URL: https://rumble.com/v4183gk-episode-19-silenced-with-tommy-robinson-liam-tuffs.html
[Rumble] v4183gk-episode-19-silenced-with-tommy-robinson-liam-tuffs.html: Downloading webpage
[RumbleEmbed] Extracting URL: https://rumble.com/embed/v3ympvh
[RumbleEmbed] v3ympvh: Downloading JSON metadata
[debug] Formats sorted by: hasvid, ie_pref, lang, quality, res, fps, hdr:12(7), vcodec:vp9.2(10), channels, acodec, size, br, asr, proto, vext, aext, hasaud, source, id
[debug] Default format spec: bestvideo*+bestaudio/best
[info] v3ympvh: Downloading 1 format(s): mp4-1080p
rumble|||The Official Corbett Report Rumble Channel|||corbettreport|||1702584000|||v3ympvh|||3:14:08|||Episode 19 - SILENCED with Tommy Robinson - Liam Tuffs|||https://ak2.rmbl.ws/s8/1/e/s/F/I/esFIo.qR4e-small-Episode-19-SILENCED-with-To.jpg
[download] Downloading item 2 of 5
It looks like parts of the log are cut off
There are two logs there showing two urls, one that works fine, the rumble /user/ path, and one that is newly broken, the rumble /c/ path. In both cases, the tail of the output shows the broken functionality.
The first line of each log is missing
My apologies, fixed above, along with some better editing of the original post.
I noticed the same problem and did some digging: It would appear that rumble made some changes to the site.
Replacing the Regex class=video-item--a\s?href=([^>]+\.html) with class=video-item--a\s?href=([^>]+\.html)|class="title__link link"\shref="([^>]+\.html) in rumble.py on line 386 fixed the problem.
this fix is not released yet, correct?
@meoyawn it's not in a stable release yet, but it is in the nightly release:
yt-dlp --update-to nightly
or
python3 -m pip install -U --pre yt-dlp
|
2025-04-01T04:36:00.984655
| 2024-01-19T23:25:48
|
2091662434
|
{
"authors": [
"Ali-Flt",
"bashonly"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:12242",
"repo": "yt-dlp/yt-dlp",
"url": "https://github.com/yt-dlp/yt-dlp/issues/9030"
}
|
gharchive/issue
|
Python API: prepare_filename() doesn't return correct extension when using postprocessor
DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE
[X] I understand that I will be blocked if I intentionally remove or skip any mandatory* field
Checklist
[X] I'm reporting a bug unrelated to a specific site
[X] I've verified that I have updated yt-dlp to nightly or master (update instructions)
[X] I've checked that all provided URLs are playable in a browser with the same IP and same login details
[X] I've checked that all URLs and arguments with special characters are properly quoted or escaped
[X] I've searched known issues and the bugtracker for similar issues including closed ones. DO NOT post duplicates
[X] I've read the guidelines for opening an issue
Provide a description that is worded well enough to be understood
When using postprocessors like FFmpegVideoRemuxer to change the video container, prepare_filename() does not return correct video extension.
Example:
import yt_dlp
URL = 'https://www.youtube.com/watch?v=K5o7U1WrJXc&t=4s'
ydl_opts = {'proxy': 'http://<IP_ADDRESS>:20801',
'verbose': True,
'format': 'bv*[height<=1080]+ba/b[height<=1080] / wv*+ba/w',
'final_ext': 'mp4', 'postprocessors': [{'key': 'FFmpegVideoRemuxer', 'preferedformat': 'mp4'}]
}
with yt_dlp.YoutubeDL(ydl_opts) as ydl:
info = ydl.extract_info(URL, download=True)
file_path = ydl.prepare_filename(info)
print(file_path)
if the video is originally .mkv, file_path will be XXX.mkv instead of XXX.mp4 while the created file is actually XXX.mp4
Provide verbose output that clearly demonstrates the problem
[X] Run your yt-dlp command with -vU flag added (yt-dlp -vU <your command line>)
[X] If using API, add 'verbose': True to YoutubeDL params instead
[X] Copy the WHOLE output (starting with [debug] Command-line config) and insert it below
Complete Verbose Output
[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version<EMAIL_ADDRESS>from yt-dlp/yt-dlp [f10589e34] API
[debug] params: {'proxy': 'http://<IP_ADDRESS>:20801', 'verbose': True, 'format': 'bv*[height<=1080]+ba/b[height<=1080] / wv*+ba/w', 'final_ext': 'mp4', 'postprocessors': [{'key': 'FFmpegVideoRemuxer', 'preferedformat': 'mp4'}], 'compat_opts': set(), 'http_headers': {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/90.0.4430.70 Safari/537.36', 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', 'Accept-Language': 'en-us,en;q=0.5', 'Sec-Fetch-Mode': 'navigate'}}
[debug] Lazy loading extractors is disabled
[debug] Python 3.11.4 (CPython x86_64 64bit) - Linux-5.4.0-152-generic-x86_64-with-glibc2.31 (OpenSSL 3.0.10 1 Aug 2023, glibc 2.31)
[debug] exe versions: ffmpeg N-66595-gc2b38619c0-static (setts), ffprobe N-66595-gc2b38619c0-static
[debug] Optional libraries: Cryptodome-3.20.0, brotli-None, certifi-2023.11.17, mutagen-1.47.0, requests-2.31.0, sqlite3-3.41.2, urllib3-2.1.0, websockets-12.0
[debug] Proxy map: {'all': 'http://<IP_ADDRESS>:20801'}
[debug] Request Handlers: urllib, requests, websockets
[debug] Loaded 1798 extractors
[youtube] Extracting URL: https://www.youtube.com/watch?v=K5o7U1WrJXc&t=4s
[youtube] K5o7U1WrJXc: Downloading webpage
[youtube] K5o7U1WrJXc: Downloading ios player API JSON
[youtube] K5o7U1WrJXc: Downloading android player API JSON
[youtube] K5o7U1WrJXc: Downloading m3u8 information
[debug] Sort order given by extractor: quality, res, fps, hdr:12, source, vcodec:vp9.2, channels, acodec, lang, proto
[debug] Formats sorted by: hasvid, ie_pref, quality, res, fps, hdr:12(7), source, vcodec:vp9.2(10), channels, acodec, lang, proto, size, br, asr, vext, aext, hasaud, id
[info] K5o7U1WrJXc: Downloading 1 format(s): 248+251
[debug] Invoking http downloader on "https://rr2---sn-i3b7kn6s.googlevideo.com/videoplayback?expire=1705728269&ei=rQSrZeenK8aZ0u8PopuE8A8&ip=<IP_ADDRESS>&id=o-AIR5-3CazicZeXH6Rrbwdbz_O0H0xixn_o2wukr-4L79&itag=248&source=youtube&requiressl=yes&xpc=EgVo2aDSNQ%3D%3D&mh=kp&mm=31%2C26&mn=sn-i3b7kn6s%2Csn-5goeenez&ms=au%2Conr&mv=m&mvi=2&pl=24&initcwndbps=1040000&vprv=1&svpuc=1&mime=video%2Fwebm&gir=yes&clen=24829785&dur=179.867&lmt=1699685452545601&mt=1705706225&fvip=4&keepalive=yes&fexp=24007246&c=IOS&txp=5437434&sparams=expire%2Cei%2Cip%2Cid%2Citag%2Csource%2Crequiressl%2Cxpc%2Cvprv%2Csvpuc%2Cmime%2Cgir%2Cclen%2Cdur%2Clmt&sig=AJfQdSswRQIgbuVi9MIWRjGD4opcg7gIT7tp4VUIVg5SRPbUGBVQDCACIQD2CS9uatphsFVDTLkKfCHZzXDaiquISN_JKuH4FFrZYg%3D%3D&lsparams=mh%2Cmm%2Cmn%2Cms%2Cmv%2Cmvi%2Cpl%2Cinitcwndbps&lsig=AAO5W4owRQIgN3l1xjCUKpXziJ7h_3fAz2EnwxooifH3EsYpKwvcVSYCIQCrQLqIBL_41IKX2KoU6VBkrFYV0gpzICQi4HQmka4kZw%3D%3D"
[download] Destination: 3 Minutes of Funny Animals [K5o7U1WrJXc].f248.webm
[download] 100% of 23.68MiB in 00:00:13 at 1.80MiB/s
[debug] Invoking http downloader on "https://rr4---sn-i3belne6.googlevideo.com/videoplayback?expire=1705728269&ei=rQSrZbriOZ6B0u8PgICW2A4&ip=<IP_ADDRESS>&id=o-AJnbPuIgGEgmXSEPjEvyHEu0gTSt2L-2O1lMbaBNH0dk&itag=251&source=youtube&requiressl=yes&xpc=EgVo2aDSNQ%3D%3D&mh=kp&mm=31%2C29&mn=sn-i3belne6%2Csn-i3b7kn6s&ms=au%2Crdu&mv=m&mvi=4&pl=24&initcwndbps=1040000&spc=UWF9f-Sd5lqt2O1TvyAOXnwxOejK2Q4fbxF_&vprv=1&svpuc=1&mime=audio%2Fwebm&gir=yes&clen=2021081&dur=179.881&lmt=1699681127221087&mt=1705706225&fvip=2&keepalive=yes&fexp=24007246&c=ANDROID&txp=5432434&sparams=expire%2Cei%2Cip%2Cid%2Citag%2Csource%2Crequiressl%2Cxpc%2Cspc%2Cvprv%2Csvpuc%2Cmime%2Cgir%2Cclen%2Cdur%2Clmt&sig=AJfQdSswRAIgZnTfvbIEbUVN3-e-DwRHCmILZ5Fl1TBUFESALIXZ2GwCICGdquLs-rv85ZSRo6w6vQ1zG5O6L2_XM33hjXT8Vl5Y&lsparams=mh%2Cmm%2Cmn%2Cms%2Cmv%2Cmvi%2Cpl%2Cinitcwndbps&lsig=AAO5W4owRQIgRPUJyXjn7mU6JRTjTYmTR9W-rdtH7csdSCDknmgQBy0CIQD_EPXHIPdDv7drXgrWbwO1E42_dF-Hzv4BXWDh7oZS2g%3D%3D"
[download] Destination: 3 Minutes of Funny Animals [K5o7U1WrJXc].f251.webm
[download] 100% of 1.93MiB in 00:00:03 at 646.84KiB/s
[Merger] Merging formats into "3 Minutes of Funny Animals [K5o7U1WrJXc].webm"
[debug] ffmpeg command line: ffmpeg -y -loglevel repeat+info -i 'file:3 Minutes of Funny Animals [K5o7U1WrJXc].f248.webm' -i 'file:3 Minutes of Funny Animals [K5o7U1WrJXc].f251.webm' -c copy -map 0:v:0 -map 1:a:0 -movflags +faststart 'file:3 Minutes of Funny Animals [K5o7U1WrJXc].temp.webm'
Deleting original file 3 Minutes of Funny Animals [K5o7U1WrJXc].f248.webm (pass -k to keep)
Deleting original file 3 Minutes of Funny Animals [K5o7U1WrJXc].f251.webm (pass -k to keep)
[VideoRemuxer] Remuxing video from webm to mp4; Destination: 3 Minutes of Funny Animals [K5o7U1WrJXc].mp4
[debug] ffmpeg command line: ffmpeg -y -loglevel repeat+info -i 'file:3 Minutes of Funny Animals [K5o7U1WrJXc].webm' -map 0 -dn -ignore_unknown -c copy -movflags +faststart 'file:3 Minutes of Funny Animals [K5o7U1WrJXc].mp4'
Deleting original file 3 Minutes of Funny Animals [K5o7U1WrJXc].webm (pass -k to keep)
3 Minutes of Funny Animals [K5o7U1WrJXc].webm
https://github.com/yt-dlp/yt-dlp/issues/5517#issuecomment-1312583224
... this is not the recommended way to get the filename. Assuming single videos ... this is prefered:
with yt_dlp.YoutubeDL(ydl_opts) as ydl:
info = ydl.extract_info(URL, download=True)
file_path = info['filepath']
print(file_path)
@bashonly what you mentioned doesn't work. filepath is not in the info
The code I'm running:
import yt_dlp
URL = 'https://www.youtube.com/watch?v=K5o7U1WrJXc&'
ydl_opts = {'proxy': 'http://<IP_ADDRESS>:20801',
}
def get_filepaths(info):
if info.get('filepath'):
yield info['filepath']
return
for entry in info.get('entries', []):
yield from get_filepaths(entry)
with yt_dlp.YoutubeDL(ydl_opts) as ydl:
info = ydl.extract_info(URL, download=True)
filepaths = tuple(get_filepaths(info))
print(filepaths)
It prints ()
oops it's info['requested_downloads']['filepath']
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.