added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:37:36.246888
| 2024-02-11T21:39:00
|
2129181246
|
{
"authors": [
"AbhijeetKrishnan",
"theneosloth"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2824",
"repo": "TLNBS2405/heihachi",
"url": "https://github.com/TLNBS2405/heihachi/pull/5"
}
|
gharchive/pull-request
|
Fix typing annotation
I run an instance of the bot in the containerized 3.9 environment, after rebasing with the master fork started seeing failures like
TypeError: unsupported operand type(s) for |: 'type' and 'NoneType'
Seems to do with an invalid union type annotation. I updated it to use the standard Optional types instead
| was added as a union type shorthand in python 3.10 so feel free to close if you don't intend to support earlier versions
discord.py is the main dependency and it requires Python 3.8+, so I wonder if it would be a better idea to support at least versions 3.8+.
|
2025-04-01T06:37:36.248685
| 2023-02-22T21:14:34
|
1595840014
|
{
"authors": [
"ThiloAschebrock"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2825",
"repo": "TNG/next-generation-scrum-poker",
"url": "https://github.com/TNG/next-generation-scrum-poker/pull/106"
}
|
gharchive/pull-request
|
Add button to refresh session id
I change the focus state to use outline similar to our links instead to separate from the hover state as suggested.
|
2025-04-01T06:37:36.298286
| 2015-03-12T03:14:35
|
60764859
|
{
"authors": [
"enzoz",
"rodrigomaia17"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2826",
"repo": "TWtablero/tablero",
"url": "https://github.com/TWtablero/tablero/pull/183"
}
|
gharchive/pull-request
|
Add vagrant to provide a faster/easier way to run application
In order to fulfill #153 , I wrote a script to provision Tablero with ansible. To make this even easier, I create a config.json file that can be used to setup configs without environment variables.
###How to run
fil the config.json file with your config.
Install Vagrant and Ansible
run vagrant up
open the box with vagrant ssh
run npm start
now tablero will be available on the host machine on localhost:3000
@enzoz Can you check this, please? (:
@rodrigomaia17 :+1: :smile:
|
2025-04-01T06:37:36.306133
| 2019-10-10T16:18:26
|
505375507
|
{
"authors": [
"featdd",
"sypets"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2827",
"repo": "TYPO3-Documentation/t3SphinxThemeRtd",
"url": "https://github.com/TYPO3-Documentation/t3SphinxThemeRtd/issues/127"
}
|
gharchive/issue
|
403 Error on docs.typo3.org for my documentation
Hi,
the link in the extension browser to my documentation results in a 403 (https://docs.typo3.org/typo3cms/extensions/dpn_glossary/stable/)
There hasn't been any changes in the last releases, is this the right place for this issue?
Greetings
Daniel
You can report problems with documentation rendering either
on Slack (https://typo3.slack.com) in #typo3-documentation (see https://typo3.org/community/teams/documentation/#c9886)
or in this issue: https://github.com/TYPO3-Documentation/T3DocTeam/issues/92 (this is also linked on the "Migrate Documentation" page under step 4: request redirects)
This here is not the correct repository, but it's not a problem. I will forward your request.
In any case, you should request a redirect if you have not done so already. I saw that your extension has already been rendered with the new mechanism: https://docs.typo3.org/p/featdd/dpn-glossary/3.0/en-us/
Actually, extensions.typo3.org should automatically update its information and use the new URL. This may take up to a day.
Hi @sypets,
thanks for all the information and the forwaring of my request.
Greetings
Daniel
|
2025-04-01T06:37:36.335896
| 2016-12-21T10:56:39
|
196903353
|
{
"authors": [
"TakahikoKawasaki",
"phihochzwei"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2828",
"repo": "TakahikoKawasaki/nv-websocket-client",
"url": "https://github.com/TakahikoKawasaki/nv-websocket-client/issues/90"
}
|
gharchive/issue
|
Cannot reconnect to Websocket
Hi there,
here I have something really strange for you.
We are using you library to establish a connection between an Android Wear Device (V 1.5) and a WebSocket. The device is paired with a Samsung Galaxy S5 (latest Android). When our app first starts (Watch is connected via bluetooth to Mobile), everything works fine and as expected. If then the Bluetooth connection is cut (for whatever reason), the Watch switches to WIFI and also connects as intended. When you now reconnect the Watch to the Mobile via Bluetooth, the WebSocket connection does not get established. Only solution is, to reboot the Watch (which actually is no solution :) )
One "funny" thing to mention: When you get the Network-Type on the Watch while connected via Bluetooth, it returns the type "PROXY" (constant value is 16) which actually is NOWHERE documented. And when you try to get the DefaultProxy-Information vom the ConnectivityManager, it returns a null
And here is the message thrown, when trying to connect
com.neovisionaries.ws.client.WebSocketException: Failed to connect to 'XXXX:3000': Host is unresolved: XXXX at com.neovisionaries.ws.client.SocketConnector.doConnect(SocketConnector.java:119) at com.neovisionaries.ws.client.SocketConnector.connect(SocketConnector.java:81) at com.neovisionaries.ws.client.WebSocket.connect(WebSocket.java:2022) at com.neovisionaries.ws.client.ConnectThread.run(ConnectThread.java:37) Caused by: java.net.UnknownHostException: Host is unresolved: XXXX at java.net.Socket.connect(Socket.java:867) at com.neovisionaries.ws.client.SocketConnector.doConnect(SocketConnector.java:110)
I would be really thankful for any help
After purchasing an Acer and a LG Phone, I can confirm that this is not related to the known "Samsung cannot resolve URI"-Problems
I'm sorry I don't understand your issue well enough, but it seems that the error message says XXXX is an invalid host name. Is it an actual error message, or did you replace an actual host name with XXXX before pasting the error message here?
If "Host is unresolved" is the root cause, it's not a WebSocket issue but a DNS issue.
Maybe the problem gets clearer when I explain the steps to reproduce:
Wear-Device connected to Internet via Bluetooth over Smartphone -> Host can be resolved and WebSocket connects
Turn off Bluetooth on Wear-Device
Wear-Device connected via WiFi directly -> Host can be resolved and WebSockets connects
Turn on Bluetooth on Wear-Device
Wear-Device connected to Internet via Bluetooth over Smartphone (again) -> Host cannot be resolved and WebSocket cant connect. Never again. Until you manually reboot the Wear-Device. Then it all start´s at 1.
Well, then, it sounds that the root cause exists in the network layer in Wear-Device. A WebSocket library cannot do anything for the problem...
|
2025-04-01T06:37:36.338265
| 2022-06-09T09:34:26
|
1265860452
|
{
"authors": [
"LittleUnicat",
"Takak11"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2829",
"repo": "Takak11/recruit",
"url": "https://github.com/Takak11/recruit/issues/1"
}
|
gharchive/issue
|
How does it work?
How dos it work?
Can you teach me?
😁🤣😍
turtle with babershop
LittleUnicat
turtle with babershop
💈
🐢
I've already hanked this project.
I've already hanked this project.
WTF
|
2025-04-01T06:37:36.339659
| 2020-03-11T21:33:10
|
579557811
|
{
"authors": [
"asajatovic"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2830",
"repo": "TakeLab/spacy-udpipe",
"url": "https://github.com/TakeLab/spacy-udpipe/pull/11"
}
|
gharchive/pull-request
|
Feature/Optional fine-grained POS
'Dirty' fix for issue #10
Latest update fixes #10 by adding UDPipeLanguage.tokenizer.fine_POS bool attribute and
addresses #9 with a code example in README.md
|
2025-04-01T06:37:36.344332
| 2016-05-25T17:07:01
|
156802416
|
{
"authors": [
"KonsomeJona",
"Takohi-Jona",
"jackzampolin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2831",
"repo": "Takohi-Jona/OctoMouse",
"url": "https://github.com/Takohi-Jona/OctoMouse/issues/3"
}
|
gharchive/issue
|
[feature-request] Document storage
I see visualizations in the upcoming feature list but it is not here yet. Perhaps you can document the storage for event data and then people can help themselves! This would be great for non Objective-C programmers who want to hack and create their own visualizations.
Hello,
I'm not sure to understand your request, but as suggested in this comment:
https://github.com/Takohi-Jona/OctoMouse/issues/2
Exporting the data in a CSV document would be good as any developer or non-developer could import these data into their own applications or a Excel sheet in order to create their own visualizations.
I would love that! What I'm most interested in is historical data that I could graph. Thank you!
I'm gonna work on it. I can't give to you an estimate right now as I'm working on my personal project during my spare time but I will keep you in touch.
Pull Request for this feature:
https://github.com/KonsomeJona/OctoMouse/pull/8
|
2025-04-01T06:37:36.390026
| 2023-06-05T13:10:35
|
1741754306
|
{
"authors": [
"edgar0011",
"tannerlinsley"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2832",
"repo": "TanStack/router",
"url": "https://github.com/TanStack/router/issues/609"
}
|
gharchive/issue
|
Link not working on first click
Describe the bug
Link, not working properly, on first click, only the second click will change url params
can be seen in the examples:
https://tanstack.com/router/v1/docs/examples/react/basic
Your Example Website or App
https://tanstack.com/router/v1/docs/examples/react/basic
Steps to Reproduce the Bug or Issue
go to : https://tanstack.com/router/v1/docs/examples/react/basic
click on Home, Posts, or any post
Expected behavior
Expect Link to work on the frist click
Screenshots or Videos
No response
Platform
OS: [e.g. macOS, Windows, Linux]
Browser: [e.g. Chrome, Safari, Firefox]
Version: [e.g. 91.1]
any
Additional context
No response
Fixed in latest
|
2025-04-01T06:37:36.428212
| 2016-09-26T10:32:54
|
179192868
|
{
"authors": [
"TannerRogalsky",
"bonez001"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2833",
"repo": "TannerRogalsky/love.js",
"url": "https://github.com/TannerRogalsky/love.js/issues/40"
}
|
gharchive/issue
|
Player not creating the proper game.data and game.js file
I am done publishining a game as a html 5 game using
python C:\Users\pc\love.js\emscripten\tools\file_packager.py game.data --preload C:\Users\pc\Desktop\Example_Game_1\ExampleGame.exe@/ --js-output=game.js
but for my second try, it doesn't output the correct file. What is worng?
You should be packaging either your game folder or the .love file, not the
.exe file.
On Monday, 26 September 2016, bonez001<EMAIL_ADDRESS>wrote:
I am done publishining a game as a html 5 game using
python C:\Users\pc\love.js\emscripten\tools\file_packager.py game.data
--preload C:\Users\pc\Desktop\Example_Game_1\ExampleGame.exe@/
--js-output=game.js
but for my second try, it doesn't output the correct file. What is worng?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/TannerRogalsky/love.js/issues/40, or mute the thread
https://github.com/notifications/unsubscribe-auth/AAn-U0lQgbexFZWaNUcxArkyNtyijcunks5qt59WgaJpZM4KGXpK
.
Thanks for the reply. I tried to do it but it outputs files that hvae very little data 10-500 kb. When I put the game.data and game.js to the debug folder, it always outputs the sound and the screen of one previous game. I dunno what is wrong. Thank you.
I type this one
python C:\Users\pc\love.js\emscripten\tools\file_packager.py game.data --preload C:\Users\pc\game_cards@/ --js-output=game.js.
I dunno what is wrong. Maybe it has something to do with the conf.lua of previous game? I will try it on other pc.
That looks right to me. What is the structure of the game_cards directory?
Hello. I made it again. I dunoo what is wrong last time.
|
2025-04-01T06:37:36.434234
| 2023-08-14T07:38:05
|
1849190590
|
{
"authors": [
"MenkOrd",
"Post-Mortem",
"TapiocaFox"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2835",
"repo": "TapiocaFox/Daijishou",
"url": "https://github.com/TapiocaFox/Daijishou/issues/509"
}
|
gharchive/issue
|
Retroachievements doesn't update in the widget
Hi, I can enter into my RA profile in the web, everything okay. But when I try to refresh my RA widget, it doesn't work.
Could you provide more details? Thanks.
Issue creator didn't provide any details after 2 weeks.
|
2025-04-01T06:37:36.474715
| 2024-05-19T15:12:41
|
2304656326
|
{
"authors": [
"gibman",
"ug30242"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2836",
"repo": "Tasshack/dreame-vacuum",
"url": "https://github.com/Tasshack/dreame-vacuum/issues/552"
}
|
gharchive/issue
|
One of c102cn's cleaning modes has disappeared.
Describe the bug
There are a total of 4 cleaning modes in c102cn, of which the mopping after sweeping mode has disappeared.
To Reproduce
When I select cleaning mode, I can't find the mopping after sweeping mode. (Mi home app has that mode) It seems to have probably disappeared since version 2.0.0b12 or 2.0.0b11.
Expected behavior
In cleaning mode, you should see four modes: sweeping, mopping, sweeping and mopping, and mopping after sweeping.
Screenshots
Additional Information (please complete the following information)
Model name: dreame.vacuum.c102cn
When this happens to me I can either do one of two things in order to temp. fix it.
restart the dreame integration.
start vacuum job on the dreame app. and then just force it to return to dock.
This seems to fix the available cleaning modes in home assistant.
|
2025-04-01T06:37:36.568198
| 2022-10-08T03:56:57
|
1401811243
|
{
"authors": [
"Rahul-Palamarthi",
"TaufeeqRiyaz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2837",
"repo": "TaufeeqRiyaz/Spotify-Clone",
"url": "https://github.com/TaufeeqRiyaz/Spotify-Clone/issues/41"
}
|
gharchive/issue
|
Carousel Slider
Want to add carousel slider to the "Good Morning section". Pls assign it to me..
Sure, I've assigned this to you
|
2025-04-01T06:37:36.585301
| 2023-03-30T22:15:44
|
1648397475
|
{
"authors": [
"abbaskhank",
"arcticfly"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2838",
"repo": "TaxyAI/browser-extension",
"url": "https://github.com/TaxyAI/browser-extension/issues/15"
}
|
gharchive/issue
|
Issue with back button/pdf docs in browser
I have tried a couple of websites where I'm trying to download documents from, it does pretty good job at identifying the docs but then after opening a doc it fails to go back to download more.
Starting page:
https://louisville.edu/purchasing/bids
Next page: this works as expected
https://louisville.edu/purchasing/bids/RP-026-22
Next page: this is where it should download the document but instead it click on the doc which opens a pdf in the browser and then things fail
https://louisville.edu/purchasing/bids/rp-026-22-document
Error
Invalid argument type: Expected a number for argument "elementId", but got ""B186503E38A0DCEF54EB2C9C31EE4177"".
The user requests the following task:
Download all open bids details that I need to submit my proposal
You have already taken the following actions:
I should click on each bid to access its details
click(469)
I should click on all the related documents to download the details
click(476)
Current time: 3/30/2023, 6:11:06 PM
Current page contents:
Response
I should download the PDF file
click("B186503E38A0DCEF54EB2C9C31EE4177")
Action
{
"error": "Invalid argument type: Expected a number for argument "elementId", but got ""B186503E38A0DCEF54EB2C9C31EE4177""."
}
Very weird. Could you also post the "Prompt" section of the last action?
i will try to recreate it but basically it just fails every time you click on a link that opens a pdf. Not fails per say bit it just doesn't know what to do after that, maybe it consider that pdf doc a new website?
|
2025-04-01T06:37:36.586860
| 2017-03-01T11:11:45
|
211049848
|
{
"authors": [
"Harmonicorps"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2839",
"repo": "TaylanTatli/Moon",
"url": "https://github.com/TaylanTatli/Moon/issues/104"
}
|
gharchive/issue
|
Help eliminating the annoying flash between pages
So every time I click a link on my site, the screen flashes whatever $bodycolor is currently set to. I want to eliminate this flash. I am aware of InstantClick.js which eliminates the flash, but it doesn't play well with the website. For some reason it makes the actual content disappear after clicking to new pages a few times. I also noticed it tries to load posts (posts button on home page) via http instead of https and as a result, the list of posts never loads. Any chance you could look into getting this awesome little javascript to work correctly?
https://github.com/dieulot/instantclick/issues/151
|
2025-04-01T06:37:36.587810
| 2020-07-12T22:11:28
|
655486263
|
{
"authors": [
"Tbaut"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2840",
"repo": "Tbaut/open-polkassembly",
"url": "https://github.com/Tbaut/open-polkassembly/pull/6"
}
|
gharchive/pull-request
|
Add initial DB creation
Creates a DB and collection at first run (or if the expected env var aren't set)
cc @niklabh
merging to keep it running.
|
2025-04-01T06:37:36.836248
| 2024-04-12T20:05:03
|
2240818214
|
{
"authors": [
"SFM61319",
"TechnikTobi",
"Xuf3r"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2851",
"repo": "TechnikTobi/little_exif",
"url": "https://github.com/TechnikTobi/little_exif/issues/7"
}
|
gharchive/issue
|
[Bug]: little_exif::metadata::Metadata::write_to_file is too slow and requires optimization
As the title suggests, the little_exif::metadata::Metadata::write_to_file method is too slow and needs to be optimized.
I am using little_exif as a dependency in a CLI app of mine. I love the API design and appreciate the good use of Rust's type system (especially for EXIF tags). However, there is a major drawback when it comes to the performance of the crate. Most (~99%) of the time in my core logic function (that involves writing a single EXIF tag to a JPEG image) is spent in little_exif::metadata::Metadata::write_to_file. Before this, my core logic function also reads the same file using std::fs::read and computes its hash, and yet that part of the logic seems to only take ~1% of the time.
My program without little_exif::metadata::Metadata::write_to_file takes less than 100ms (even after including logging IO) for 10 JPEG images (~20MB each). But with little_exif::metadata::Metadata::write_to_file it takes over 4s with the same data under the same conditions. That is 40x the original execution time. And this was just a sample dataset. The program's expected input specification states the JPEG image count to be >=60,000, the size of each image being >=20MB. This means a task supposed to be completed in under 10 minutes will take over a day, which is unacceptable to say the least.
A flame graph of the little_exif::metadata::Metadata::write_to_file part of my core logic function:
When I asked for help on the Rust Community Discord server, someone mentioned that little_exif::metadata::Metadata::write_to_file was likely reading the same file again and again, 4 times in total, therefore increasing the time as well as the memory consumption. I haven't gone through little_exif's source code, so I thought I should let you know about this possible bug source.
Yeah i was appalled by the execution time too.
The issue lies in the clean_metadata() function which write_to_file() calls.
It pounds the fs with syscalls byte by byte to find and remove the APP1 segment from JPEG.
I've refactored the JPEG part into in-memory mutation but since then wrote my own parser for PNG and JPEG for all the markers.
I can submit the pull request if the owner is fine with refactor of the clean_metadata() for only one format.
@SFM61319 Thanks for your feedback - glad you like the API design! :) Regarding the performance issue: Speed was never one of the main objectives of little_exif. I’m aware that some functions are primitive and have a lot of room for improvement regarding performance. So far during development it was a trade-off between speed and being able to follow what is going on, how metadata is encoded, etc. This however does not mean that I am not open for suggestions/pull requests/etc. on how to improve the current situation! Speaking of pull requests:
@Xuf3r I'd be happy to incorporate your modifications to metadata cleaning!
@SFM61319 There is now a new version that includes the optimized version for metadata clearing from @Xuf3r. Would be great to hear from you whether this fixed the performance issue you experienced!
Nice! There is a major performance boost in the new update. I appreciate the quick response from your side, and a satisfying one at that! Thank you so much!
Here is the new flame chart:
|
2025-04-01T06:37:36.876717
| 2018-07-05T09:43:47
|
338500852
|
{
"authors": [
"gdgellatly",
"pedrobaeza"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2852",
"repo": "Tecnativa/docker-odoo-base",
"url": "https://github.com/Tecnativa/docker-odoo-base/pull/152"
}
|
gharchive/pull-request
|
Set initdb args for Odoo in scaffolding common.yaml
to ensure data integrity and index performance.
Address point 3 of #151
What about the LC_COLLATE thing?
? locale implicitly sets LC_COLLATE afaik as well as LC_TYPE
|
2025-04-01T06:37:36.899453
| 2018-09-05T13:12:07
|
357222814
|
{
"authors": [
"iTetsuyaMurakami",
"ishidawataru"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2853",
"repo": "Telecominfraproject/oopt-tai",
"url": "https://github.com/Telecominfraproject/oopt-tai/pull/25"
}
|
gharchive/pull-request
|
Add taish app
Add taish app to provide the simple CLI interface to load an external TAI module, initialize optical modules via TAI, get the status of optical modules.
@scott-emery I'd like to have this code in this repo. I think oopt-tai-implementations repo is a repo for libtai.so implementation. This app is not implementing TAI.
@iTetsuyaMurakami Thanks!
|
2025-04-01T06:37:36.912598
| 2015-01-29T18:07:00
|
55934916
|
{
"authors": [
"SachaG",
"fobo66"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2854",
"repo": "TelescopeJS/Telescope",
"url": "https://github.com/TelescopeJS/Telescope/pull/733"
}
|
gharchive/pull-request
|
*Added russian translation (not full, because some functions are fixed i...
...n English);
+Added semicolons in users.js' clog callbacks. Also, i find some unuseful functions in users.js. Can i delete it?
We're switching to LingoHub to help manage translations. If you'd like to be added as a translator to the project, could you join the #i18n channel on Telescope's Slack?
|
2025-04-01T06:37:36.925121
| 2023-05-05T04:08:56
|
1696977011
|
{
"authors": [
"TechDufus",
"chrisbenincasa",
"eremem",
"luispabon",
"noahccoker"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2855",
"repo": "Telmate/terraform-provider-proxmox",
"url": "https://github.com/Telmate/terraform-provider-proxmox/issues/763"
}
|
gharchive/issue
|
All disk properties config show -> null as needed change.
Terraform v1.4.6
on linux_amd64
+ provider registry.terraform.io/telmate/proxmox v2.9.11
Small context, I'm unable to upgrade to newer provider versions. (See https://github.com/Telmate/terraform-provider-proxmox/issues/762)
When running a plan, even after applying, all VMs show every property needs to change -> null even if I set every value in the disk {} block, which isn't necessary.
- disk {
- backup = 0 -> null
- cache = "none" -> null
- file = "vm-112-cloudinit" -> null
- format = "raw" -> null
- iops = 0 -> null
- iops_max = 0 -> null
- iops_max_length = 0 -> null
- iops_rd = 0 -> null
- iops_rd_max = 0 -> null
- iops_rd_max_length = 0 -> null
- iops_wr = 0 -> null
- iops_wr_max = 0 -> null
- iops_wr_max_length = 0 -> null
- iothread = 0 -> null
- mbps = 0 -> null
- mbps_rd = 0 -> null
- mbps_rd_max = 0 -> null
- mbps_wr = 0 -> null
- mbps_wr_max = 0 -> null
- media = "cdrom" -> null
- replicate = 0 -> null
- size = "4M" -> null
- slot = 1 -> null
- ssd = 0 -> null
- storage = "lvm-thin-06" -> null
- storage_type = "lvmthin" -> null
- type = "scsi" -> null
- volume = "lvm-thin-06:vm-112-cloudinit" -> null
}
Naturally, being flooded with all VMs needing changed for each plan gets in the way of seeing real pending state changes.
Are you creating a VM via cloning? I'm hitting the same thing (and it seems like many others are too #655). I'm trying to figure out if this has to do with cloning a VM and not overriding the set parameters of the template?
Correct.. I set up a VM that I use as a template (I don't actually make it a template) and I clone from it and use cloud-init.
Yea that is a exactly what I'm doing as well. My guess is that the provider isn't ignoring all of the settings in the clone case
Exactly the same problem when cloning from a template.
Terraform v1.5.3
on windows_amd64
+ provider registry.terraform.io/telmate/proxmox v2.9.14
Having the same issue too, hope this can get prioritized soon
Ping. any updates on this?
The workaround for me is to specify a disk stanza that matches the disk set up on the templated VM, like so:
VM template:
Matching disk config:
|
2025-04-01T06:37:37.109771
| 2024-12-01T14:13:03
|
2709231771
|
{
"authors": [
"Darth-Carrotpie",
"grandmazboy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2856",
"repo": "TencentARC/InstantMesh",
"url": "https://github.com/TencentARC/InstantMesh/issues/178"
}
|
gharchive/issue
|
Does anyone check if installation works regularly?
I mean does it hurt to update the requirements.txt with exact versions.. since you wont know when a package is updated and if its compatible with other versions.
Does anyone ever compile code to work with newer versions of Cuda, Python or VS?
Seems like stable diffusion and other tools similar to this finish something then just give up on maintaining.
C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.1\include\crt/host_config.h(153): fatal error C1189: #error: -- unsupported Microsoft Visual Studio version! Only the versions between 2017 and 2022 (inclusive) are supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
texture.cu
ninja: build stopped: subcommand failed.
VS Build Tools 2022 17.12.2 / 14.42.34433
VS Professional 2022 14.29.30133
Desktop development with C++ / tried both MSVC v142/143
Cuda 12.1
Python 3.10
Windows 11 / Not using Conda
Nvidia RTX 4080
(venv) PS E:\InstantMesh-main> python run.py configs/instant-mesh-large.yaml examples/bird.jpg --save_video --no_rembg
Seed set to 42
Loading diffusion model ...
Loading pipeline components...: 12%|██████▌ | 1/8 [00:00<00:00, 7.82it/s]The config attributes {'dropout': 0.0, 'reverse_transformer_layers_per_block': None} were passed to UNet2DConditionModel, but are not expected and will be ignored. Please verify your config.json configuration file.
Loading pipeline components...: 100%|████████████████████████████████████████████████████| 8/8 [00:01<00:00, 6.42it/s]
Loading custom white-background unet ...
Loading reconstruction model ...
Some weights of ViTModel were not initialized from the model checkpoint at facebook/dino-vitb16 and are newly initialized: ['encoder.layer.7.adaLN_modulation.1.bias', 'encoder.layer.6.adaLN_modulation.1.weight', 'encoder.layer.4.adaLN_modulation.1.bias', 'encoder.layer.0.adaLN_modulation.1.weight', 'encoder.layer.6.adaLN_modulation.1.bias', 'encoder.layer.1.adaLN_modulation.1.bias', 'encoder.layer.10.adaLN_modulation.1.bias', 'encoder.layer.2.adaLN_modulation.1.weight', 'encoder.layer.7.adaLN_modulation.1.weight', 'encoder.layer.9.adaLN_modulation.1.weight', 'encoder.layer.3.adaLN_modulation.1.weight', 'encoder.layer.2.adaLN_modulation.1.bias', 'encoder.layer.8.adaLN_modulation.1.weight', 'encoder.layer.11.adaLN_modulation.1.bias', 'encoder.layer.8.adaLN_modulation.1.bias', 'encoder.layer.5.adaLN_modulation.1.bias', 'encoder.layer.3.adaLN_modulation.1.bias', 'encoder.layer.5.adaLN_modulation.1.weight', 'encoder.layer.9.adaLN_modulation.1.bias', 'encoder.layer.10.adaLN_modulation.1.weight', 'encoder.layer.4.adaLN_modulation.1.weight', 'encoder.layer.1.adaLN_modulation.1.weight', 'encoder.layer.11.adaLN_modulation.1.weight', 'encoder.layer.0.adaLN_modulation.1.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Traceback (most recent call last):
File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 2100, in _run_ninja_build
subprocess.run(
File "C:\Program Files\Python310\lib\subprocess.py", line 524, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "E:\InstantMesh-main\run.py", line 132, in
model.init_flexicubes_geometry(device, fovy=30.0)
File "E:\InstantMesh-main\src\models\lrm_mesh.py", line 79, in init_flexicubes_geometry
renderer = NeuralRender(device, camera_model=camera)
File "E:\InstantMesh-main\src\models\geometry\render\neural_render.py", line 73, in init
self.ctx = dr.RasterizeCudaContext(device=device)
File "E:\InstantMesh-main\venv\lib\site-packages\nvdiffrast\torch\ops.py", line 184, in init
self.cpp_wrapper = _get_plugin().RasterizeCRStateWrapper(cuda_device_idx)
File "E:\InstantMesh-main\venv\lib\site-packages\nvdiffrast\torch\ops.py", line 125, in _get_plugin
torch.utils.cpp_extension.load(name=plugin_name, sources=source_paths, extra_cflags=common_opts+cc_opts, extra_cuda_cflags=common_opts+['-lineinfo'], extra_ldflags=ldflags, with_cuda=True, verbose=False)
File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 1308, in load
return _jit_compile(
File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 1710, in _jit_compile
_write_ninja_file_and_build_library(
File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 1823, in _write_ninja_file_and_build_library
_run_ninja_build(
File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 2116, in _run_ninja_build
raise RuntimeError(message) from e
I have uninstalled Python and moved to 3.10.6 and separately installed StableDiffusion and its working . I am getting a shyt ton of package dependancy conflicts and one that always pops up is huggingface_hub
Check https://github.com/TencentARC/InstantMesh/issues/175#issuecomment-2495631712
If it's a problem in a Dockerfile, which should be stable in theory, it's gonna be even a bigger problem in regular installs...
In the end I did run it via docker, but when generating a mesh it errored out without any message. Not sure where to go on after that :/
I tried on a different computer today, with just the bare minimum of 3.10, visual studio community 2022, build tools, cuda 12.1 etc and that is getting very similar errors to what I'm getting on this PC. I suspect mass package dependency version issues just don't know the right combo..
|
2025-04-01T06:37:37.135435
| 2017-05-26T08:55:55
|
231565552
|
{
"authors": [
"anapanadero",
"kyleledbetter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2857",
"repo": "Teradata/covalent",
"url": "https://github.com/Teradata/covalent/issues/629"
}
|
gharchive/issue
|
Design for autocomplete with custom inputs or custom chips
The placeholders and labels for autocomplete with custom inputs or custom chips doesn't follow the official Material Design spec guidelines.
That spec if for the standalone autocomplete this is chips autocomplete so slightly different usage and we're using a non float label that's a feature of angular material, and we're also following the pattern of chips autocomplete from angular-material v1
|
2025-04-01T06:37:37.145967
| 2024-09-19T23:12:30
|
2537506747
|
{
"authors": [
"ACGaming",
"Mysticpasta1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2859",
"repo": "TerraFirmaCraft-The-Final-Frontier/FileDirector",
"url": "https://github.com/TerraFirmaCraft-The-Final-Frontier/FileDirector/pull/30"
}
|
gharchive/pull-request
|
For The Director on 1.20.1
all credit to juan
While this is highly appreciated, we're going to revert the rebranding, StopModReposts integration since its unmaintained and other aspects we deem unfitting. Because of my other projects, this is going to take a while.
|
2025-04-01T06:37:37.149038
| 2016-09-15T15:25:57
|
177208224
|
{
"authors": [
"ljacomet",
"mathieucarbou",
"myronkscott"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2860",
"repo": "Terracotta-OSS/terracotta-apis",
"url": "https://github.com/Terracotta-OSS/terracotta-apis/issues/152"
}
|
gharchive/issue
|
Being able to configure a @Builtin Service
Currently, a @Builtin service cannot be configured through XML.
This request is to be able to optionally configure a @Builtin service if an XML config is present.
The main use case I see is if we need, one day, to make our service that was builtin optionally configurable because some options are added.
So to keep backward compat', the service will need to support both @Builtin AND if provided, be configured by the XML config.
notify: @chrisdennis @ChrisGreenaway @myronkscott @jd0-sag @ljacomet
Might be something we may need indeed. But that's also something that can be added while being completely compatible IMO. So I do not see the point of adding it now unless really needed.
Makes sense?
Yes, for M&M, but I know @ChrisGreenaway might need it also.
This has been implemented. If a configuration is in tc-config, that will be used instead of the builtin.
|
2025-04-01T06:37:37.170386
| 2018-12-05T18:19:11
|
387883795
|
{
"authors": [
"Bootjack",
"kalimar",
"mirandawang",
"rudokemper"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2861",
"repo": "Terrastories/terrastories",
"url": "https://github.com/Terrastories/terrastories/pull/186"
}
|
gharchive/pull-request
|
[176] Allow for more granular permissions based on user access control groups
This adds a new concept in the Demographic model which links users and stories through an access control grouping. I'm open to suggestions for the name, but it seems like these will mostly be based on real-world social groupings, such as elders from a certain village or descendants of one individual. The term "demographics" felt like a decent fit. This is a different concept from both user roles and story permissions. The three complementary concepts we'd have after this PR are:
Story permissions - an individual story is marked as either:
:anonymous meaning public, or
:user_only meaning any signed-in user may see it
User roles - an individual user has a single role that's one of:
user basic signed-in user, may view some restricted content
editor same as user, and also may edit story and location content
admin same as editor, and also may edit other users (just added as an option, not yet implemented/enforced)
Demographics - a group of users and stories that defines restricted access levels using the following data (basically describing two new HABTM relations)
name a human-readable plain-language name for the group
users a list of users included in this group
stories a list of stories this group has access to
Punch List
[x] Create the Demographic model
[x] Add relationships between Demographic, User, and Story models
[x] Customize the demographic dashboard in administrate
[x] Filter stories based on Demographics
[x] Remove the direct User-Story access control association
[x] Replace the Story.permission_level field with Story.is_public
[x] Add pundit policies so that editors:
[x] may not modify User models
[x] may not view Demographics they don't belong to
[x] may not add stories to Demographics they don't belong to
[x] can only see appropriate Users and Stories when modifying Demographics
[x] may not add Demographics outside of their permissions to Stories they have access to
[ ] Automatically add the user who creates a Demographic to that Demographic (otherwise only
admins can access it)
[x] add the current user to the demographic
[ ] prevent duplicate records
[x] Do not list Demographics under Stories that when an editor doesn't belong to them
[ ] Prevent limited Demographics scope of editor from deleting existing Demographics when updating Stories
For Discussion
I think we can streamline these concepts a bit, but we probably do want some form of all three. Check out this branch, go to the new Demographics dashboard in the admin and see if creating demographic groups and associating users and stories seems logical. Here are some specific questions I have for the team.
Story permissions is currently a text field at Story.permission_level but I think at this point it only needs to be a boolean indicating whether a story should be shown to the public, something like Story.is_public.
Similarly, should we expect that the combination of user role plus demographics will always result in the correct editing permission for stories. In other words, if a user is an editor, should they have editing access to all stories their demographics allow them to see? Or should there be a distinction between view-only demographics and editor-specific demographics?
Does the admin user role make sense? My expectation would be that members of the ACT who are setting things up (plus maybe a few tech-savvy locals) would be the only admins. Everyone else would go to that group to get elevated editing permissions or to modify demographics for access control.
I also modified the models to allow for direct user-to-story access levels, but I'm curious if this is really needed or if it just adds extra complexity and confusion. The same could be accomplished by creating a demographic group with only one user. If that's a rarity it should be fine; if it's common the number of single-user demographics could become a problem. Note, there's no dashboard UI for setting up these direct user-story relationships yet.
AWESOME!!! :100: :tada:
Thanks for getting this started!
Discussion points:
I think updating that filed to a boolean is_public? makes total sense here, since we only really have that distinction now for Stories.
Editing and viewing to me are distinct. Viewing stories is on a user level, but some users also have the ability to add/update/delete stories. That also means they have access to the /admin backend. There could be an "editor-only" demographic for stories which aren't ready to be published yet, or something.
admin makes sense to me! In the future these would be the users that could add and update custom fields for the models too.
Maybe @rudokemper can help us answer that question, is it more common that stories would be visible to certain groups or usually just individuals?
Thanks for the feedback, @mirandawang! To clarify my second question, I really just mean should an editor be able to edit every story they can view? I think we're agreeing that they should, but I wanted to make sure.
Brilliant stuff Jason!! 👍👍👍
I would agree here.
Yeah, I mean, theoretically there could be a scenario where a community member gets to view stories X, Y, and Z, but only have permission to edit X. Let's say that they are a hunter and story A is a hunting story, whereas Y and Z are stories told by elders which they should be able to hear, but not modify. But I really can't imagine this playing out in reality, certainly not anywhere where we work for instance. Let's just keep it simple.
Fully agreed here and also with @mirandawang's suggestion about adding and custom fields for the models in the future.
My feeling is that the demographic groups (with one-user groups if necessary) are perfectly sufficient, and that having the additional direct-to-user setup could likely result in more confusion.
Cool so, to sum up:
Editors can edit or delete any story that they can view / they have created.
Story permission levels will be based on Demographic groups. For the rare case of individual level permissions, we can model this as a one person group.
@rudokemper @Bootjack @kalimar does that all sound right?
@kalimar In my defense, jQuery was already a part of administrate 😜
Also, some general explanation: Those last few commits (including the jQuery stuff) are in support of honoring pundit policies in the multiple-select widget for Stories/Demographics/Users in the administrate forms. To do that I created a custom field HasManyScopedField that subclasses the normal Field::HasMany from administrate. That felt a little hacky because I had to copypasta two methods from the base class and tweak them.
Even more hacky is that in order to pick up the styles and js from the base field I had to run their rails generator that copied all the assets for all fields and dumped them into our app. I'm sure a more clever dev could have come up with a cleaner way to do that. Eventually, though, I think the in-map editing will supplant this entire administrate interface so I'm not super worried.
lol, nvm. @kalimar 's nerd snipe was successful
You know me too well @bootjack. You know me too well
:eyes:
closing for now; thank you for this work!
|
2025-04-01T06:37:37.185546
| 2015-02-17T16:38:41
|
57947558
|
{
"authors": [
"TerryMooreII",
"ThomasPe"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2862",
"repo": "TerryMooreII/angular-azure-mobile-service",
"url": "https://github.com/TerryMooreII/angular-azure-mobile-service/pull/19"
}
|
gharchive/pull-request
|
Exposing .read() method
In order to send parameters to the javascript backend that can be read
through request.parameters the .read() method of the azure sdk is now
exposed with angular promise wrapped around. See also
http://stackoverflow.com/questions/17415257/passing-query-parameters-from-html-js-app-to-azure-server-script
updated the parameter name, fixed some issues in the readme and added the new method documentation
Nice, thanks. I will push this to bower and the cdn soon.
|
2025-04-01T06:37:37.189453
| 2021-03-27T21:41:33
|
842607673
|
{
"authors": [
"ritvij14"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2863",
"repo": "TesseractCoding/NeoAlgo",
"url": "https://github.com/TesseractCoding/NeoAlgo/issues/4296"
}
|
gharchive/issue
|
Armstrong number using recursive approach using Dart
🚀 Feature
Add code for Armstrong number using the recursive approach in dart.
Have you read the Contributing Guidelines on Pull Requests?
Yes
/assign
|
2025-04-01T06:37:37.191455
| 2021-04-22T16:38:14
|
865164858
|
{
"authors": [
"hemant2705"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2864",
"repo": "TesseractCoding/NeoAlgo",
"url": "https://github.com/TesseractCoding/NeoAlgo/issues/6433"
}
|
gharchive/issue
|
Floyd's Triangle
💥 Proposal
(A clear and concise description of what the proposal is.)
floyd's triangle program in kotlin
1
2 3
4 5 6
7 8 9 10
Have you read the Contributing Guidelines on Pull Requests?
yes
(Write your answer here.)
/assign
|
2025-04-01T06:37:37.194420
| 2021-04-15T15:02:22
|
858973482
|
{
"authors": [
"Amit366"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2865",
"repo": "TesseractCoding/NeoAlgo",
"url": "https://github.com/TesseractCoding/NeoAlgo/pull/5913"
}
|
gharchive/pull-request
|
Trapping rain water in python
Have you read the Contributing Guidelines on Pull Requests?
yes
Description
Added code for trapping rain water
Checklist
[ ] I've read the contribution guidelines.
[ ] I've checked the issue list before deciding what to submit.
[ ] I've edited the README.md and link to my code.
Related Issues or Pull Requests
Fixes: #5851
@HarshCasper please review
@ankitaggarwal23 please review
|
2025-04-01T06:37:37.204167
| 2024-05-17T04:00:14
|
2301764741
|
{
"authors": [
"CharliePoole"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2866",
"repo": "TestCentric/TestCentric.Cake.Recipe",
"url": "https://github.com/TestCentric/TestCentric.Cake.Recipe/issues/75"
}
|
gharchive/issue
|
Restore capability to create draft release
This was temporarily blocked in issue #67. By upgrading GitReleaseManager to version 0.17, it can be restored.
:tada: This issue has been resolved in version 1.2.1 :tada:
The release is available on:
GitHub Release
NuGet Package
|
2025-04-01T06:37:37.249952
| 2015-05-08T01:18:22
|
74184428
|
{
"authors": [
"arteme",
"davidlgj",
"nicklasb",
"subhendupsingh"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2872",
"repo": "Textalk/angular-schema-form",
"url": "https://github.com/Textalk/angular-schema-form/issues/383"
}
|
gharchive/issue
|
Divide array in 3 columns
Can we divide "type":"array" in 2 or 3 columns and have a tabarray with horizontal tabs? Please suggest a workaround.
I don't think I understand, do you want the same array flowing over three columns?
Columns can be achieved by using standard bootstrap classes, see http://schemaform.io/examples/bootstrap-example.html#/4fa8967ae5596fe8b0c0
tabarray type has support for horizontal tabs, just set tabType to "top", here are the docs https://github.com/Textalk/angular-schema-form/blob/development/docs/index.md#tabarray
In my project I achieve columns by splitting the items array in half using a custom fieldset decorator like:
<fieldset ng-disabled="form.readonly" class="schema-form-fieldset schema-form-fieldset-columns {{form.htmlClass}}">
<legend ng-show="showTitle()">{{ form.title }}</legend>
<div ng-show="form.description" ng-bind-html="form.description"></div>
<div class="column">
<div class="row">
<sf-decorator ng-repeat="item in form.items|arrayHalf" form="item"></sf-decorator>
</div>
</div>
<div class="column">
<div class="row">
<sf-decorator ng-repeat="item in form.items|arrayHalf:1" form="item"></sf-decorator>
</div>
</div>
</fieldset>
The arrayHalf filter is really easy to implement...
Please re-open if you do not feel the question has been answered sufficiently.
|
2025-04-01T06:37:37.251593
| 2023-01-24T04:29:00
|
1554299021
|
{
"authors": [
"pionxzh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2873",
"repo": "TexteaInc/json-viewer",
"url": "https://github.com/TexteaInc/json-viewer/pull/216"
}
|
gharchive/pull-request
|
chore: bump version and delay dependabot
related: #151 #184
This PR bumps the deps and tries to slow down dependabot to reduce the noise.
We will try to adopt renovate bot next week.
And I will manually update the dependencies temporarily.
cc. @rtritto
|
2025-04-01T06:37:37.255604
| 2024-06-11T19:13:24
|
2347111128
|
{
"authors": [
"ketan96-m",
"willmcgugan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2874",
"repo": "Textualize/rich",
"url": "https://github.com/Textualize/rich/issues/3379"
}
|
gharchive/issue
|
[REQUEST] Rich progress bar changes color based on the number steps/files elapsed.
Consider posting in https://github.com/textualize/rich/discussions for feedback before raising a feature request.
Have you checked the issues for a similar suggestions?
Yes
How would you improve Rich?
My application submits a bunch of transformation to a server and the rich bar gives the status of the transformation and eventual downloads. However, these transformation are prone to failures and I would like to change the color of my progress bar to indicate that some failure has occurred while other transformations continue.
I would like to change the bar color to red because it stopped just before finishing.
The update function does provide ways to make changes but I couldn't find a way to change the color of the bar using BarColumn() when passed as a keyword argument to the update method.
e.g
progress.update(
progress_task,
progress_bar_title,
completed=self.current_status.files_completed,
bar = BarColumn(complete_style="rgb(0,0,255)",
style="rgb(255,0,0)")
)
What problem does it solve for you?
This will add more style to the progress bar when there are failures or if we can't reach 100%.
You can override Progress.get_renderables() to display the progress however you like.
|
2025-04-01T06:37:37.264672
| 2021-06-04T15:17:33
|
911604909
|
{
"authors": [
"DrAkashic",
"MrTubzy456",
"Tfarcenim",
"aaronhowser1",
"epicyeeto",
"j4rw15",
"tankcr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2875",
"repo": "Tfarcenim/AE2WirelessTerminalLibrary",
"url": "https://github.com/Tfarcenim/AE2WirelessTerminalLibrary/issues/16"
}
|
gharchive/issue
|
AE2 Wireless Terminals (ae2wtlib) encountered an error during the load_registries event phase
AE2 Wireless Terminals (ae2wtlib) encountered an error during the load_registries event phase
java.lang.RuntimeException: Cannot find class appeng/container/implementations/MEMonitorableContainer
crash-2021-06-04_18.12.47-fml.txt
very undescriptive report, but this looks like your ae2 version is too new
Wouldn't that be more like, this mod is outdated? If it requires an older version of AE2
Came here with same issue btw, crash log here: https://pastebin.com/H3zFJNmK
Bump.
Running into the same issue now. Here is a link to my crash log as well: https://pastebin.com/HUKwVjKR
Same issue, looks like this addon hasn't been updated since May
same issue here
Repeating an already known issue won't fix it any faster.
I just updated my server and ran into this issue. In the crash report it lead me here so I've the crash report to pastebin and if that helps let me know.
https://pastebin.com/KrVtgELd
I've got to figure out how to get my server to run. If you need anything, ask. I have access to all of the files for the server and pack.
|
2025-04-01T06:37:37.270956
| 2021-11-06T04:19:35
|
1046391194
|
{
"authors": [
"Th3M4ttman",
"scott91e1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2876",
"repo": "Th3M4ttman/ezcolors",
"url": "https://github.com/Th3M4ttman/ezcolors/issues/1"
}
|
gharchive/issue
|
Advantages of ezcolors over the de facto color library?
Hello,
I am in the market for a python color manipulation library for a new project. Any chance of a paragraph or two of why ezcolors was created and how it compares to the other libraries out there?
Many thanks,
Scott.
Sure thing. EzColors was created when I was trying to find a multi platform terminal colour solution that was easy to use but capable of a wide array of different colour functionalities like gradients, rainbows and various generateable colour palettes and I couldn't find anything that quite fit the bill of what I was looking for.
So I started work on a few functions and classes to make ANSI escape coloured strings in the form of an ordinary str or a ColorStr class with various methods to fiddle with the colours etc I got it to where it was usable for making coloured text relatively easily with nothing but ColorStr("text", "foreground colour in RGB, integer or hex", "background colour") and sort of left it for a while but then later returned and added some extra functionality like simple coloured splash screens and separators and a coloured yes/no/whatever option you want prompt. And colored exceptions for easier reading
The whole thing is somewhat of a mess and in need of cleanup but I use it in many of my projects and for quick prototyping. Overall though I'm not sure I'd recommend it other some alternatives I haven't really looked for too long for exactly what I needed.
If you wanna take a look and suggest any improvements I'd be happy to hear. There's a whole suite of eztools I've got in the works
EzFiles (contains a file class for much easier file manipulation)
EzConfig (fully featured configs from 1 line)
EzPack (dynamically generate, package and build python scripts, modules and packages)
EzColors
Ezcmd (decorator based system to very easily turn any script into a command line utility and a class to easily create a command prompt from a list of functions)
EzTest (easy decorator based unit tests with coloured output and reports with a Test class and a Case class that can be extended easily to automatically generate and cache testcases)
EzValid (a bunch of regedit powered varlidators)
With all being available either separately or as 1 package called eztools with additional utilities such as 1 line simple input function output tkinter GUI, and Timers
Sorry to just dump a huge wall of text.
In short I'm an intermediate-advanced solo python dev I wouldn't use my EzColors for anything super important but please do mess about with it and see how you like it. I will update the documentation soon as some stuff has probably slightly changed since I did it
Also to note I was high both when I wrote EzColors and when I wrote this comment. And both my comments and the entirety of EzColors development were written on a Samsung galaxy a6. So when I say it's multi platform it works everywhere. Mac, windows, Linux, Android
Also only just realised how long ago you commented this haha. Sorry for such a slow reply
|
2025-04-01T06:37:37.274777
| 2023-11-06T00:23:56
|
1978049382
|
{
"authors": [
"ThaUnknown",
"zoriya"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2877",
"repo": "ThaUnknown/jassub",
"url": "https://github.com/ThaUnknown/jassub/issues/33"
}
|
gharchive/issue
|
Chrome android does not work anymore
Subtitles simply don't appear on chrome android anymore (could not tell from which version, it was already the case in 1.7.9 but maybe before too). This can be reproduced from this page, for example: https://thaunknown.github.io/jassub/jassub/simple/index.html
There aren't any errors, and debug messages in the console appears when the subtitle should be shown. It feels like everything is working as it should, but subtitles are not shown.
can't reproduce
Interesting, even with brave, it does not work for me.
this is likely due to offscreen render, it's a chromium bug
The issue for the sandbox was #32 (which was fixed) but my codesandbox does indeed show another issue (I have not faced in production yet)
as I said, disabling offscreen render fixes this on android, I don't know why this happens, I assume it occurs when the bitmap given to skia is too big?
closing as no reproduction is available
|
2025-04-01T06:37:37.297869
| 2023-06-22T11:43:32
|
1769507602
|
{
"authors": [
"frappeerpnext",
"nikkothari22"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2879",
"repo": "The-Commit-Company/frappe-js-sdk",
"url": "https://github.com/The-Commit-Company/frappe-js-sdk/issues/31"
}
|
gharchive/issue
|
Add server message to error object
When we make request via db or call object, if error occur from backend frappe send friendly message to client via _server_messages key.
I cannot extract it in .catch((error) { ... }). It seems that you did not include it in error object.
Here is server message when I check in network response.
Could you please help include this message in .catch error object?
Thanks,
@frappeerpnext - yes this is something that needs to be added. I'll try adding it in.
Yes, I hope this will be added soon. These error information is very useful and we save a lot time to customize friendly message to end user.
Thanks,
Hi Guys,
I have update and test it. Working perfect.
Big Thanks
|
2025-04-01T06:37:37.304790
| 2023-06-21T06:04:12
|
1766815269
|
{
"authors": [
"karthikbhandary2",
"khusheekapoor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2880",
"repo": "The-Data-Alchemists-Manipal/MindWave",
"url": "https://github.com/The-Data-Alchemists-Manipal/MindWave/issues/454"
}
|
gharchive/issue
|
Implementation of ML model on Happiness index data.
💥 Proposal
I would like to apply a machine learning model to Happiness index data. By using the data I would like to predict the overall rank. Please assign me this.
@karthikbhandary2 - please provide the dataset.
This is present in this repo only. Link: https://github.com/The-Data-Alchemists-Manipal/MindWave/tree/main/Data Analytics/Happines_index_Data_analysis_visualization
@karthikbhandary2 - you can go ahead! We are assigning you 21 days for this project, after which it will be assigned to someone else if not completed. All the best!
Name the file as: algorithm_dataset.ipynb and link it in the readme of the labeled directory as algorithm - dataset.
ok @khusheekapoor
|
2025-04-01T06:37:37.322059
| 2022-08-31T16:49:24
|
1357605921
|
{
"authors": [
"WebKingdom",
"donn",
"vijayank88"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2881",
"repo": "The-OpenROAD-Project/OpenLane",
"url": "https://github.com/The-OpenROAD-Project/OpenLane/issues/1299"
}
|
gharchive/issue
|
Flow consistently fails regardless of configurations
Description
When hardening a SHA1 design inside the caravel user project, flow fails with FP_SIZING=relative and FP_CORE_UTIL=30%. Other configurations were also tried but hardening failed for all. Configuration trials can be seen in openlane/sha1_top/config.tcl (https://github.com/WebKingdom/bitcoin_asic/blob/055aed80953a4996cf8b931a32cf2a102ebe4ca6/openlane/sha1_top/config.tcl).
Expected behavior
The design should be able to harden and go through layout and placement.
Environment
Kernel: Darwin v21.6.0
Distribution: macOS 10.16
Python: v3.9.0 (OK)
Container Engine: docker v20.10.17 (OK)
OpenLane Git Version: f9b5781f5ef0bbdf39ab1c2bbd78be8db11b27f2
pip: INSTALLED
pip:venv: INSTALLED
---
PDK Version Verification Status: OK
---
Git Log (Last 3 Commits)
f9b5781 2022-07-01T16:04:31+02:00 Fix a bug with `-overwrite` (#1171) - Anton Blanchard - (grafted, HEAD, tag: 2022.07.02_01.38.08)
Reproduction Material
issue_reproducible attached as zip.
issue_reproducible.zip
Logs
Console logs:
[STEP 18]
[INFO]: Running Detailed Placement...
[ERROR]: during executing openroad script /openlane/scripts/openroad/opendp.tcl
[ERROR]: Exit code: 1
...
[ERROR]: Last 10 lines:
[INFO DPL-0035] ANTENNA__14155__A1
[INFO DPL-0035] ANTENNA__14155__A1
[INFO DPL-0035] ANTENNA__20474__A2
[INFO DPL-0035] ANTENNA__20930__A
[INFO DPL-0035] ANTENNA__20930__A
[INFO DPL-0035] ANTENNA__16103__A0
[INFO DPL-0035] message limit reached, this message will no longer print
[ERROR DPL-0036] Detailed placement failed.
Error: opendp.tcl, 32 DPL-0036
@WebKingdom
Update your config.tcl with following configuration resolve the issue:
set ::env(CELL_PAD) 2
Also you can use set ::env(PL_TARGET_DENSITY) 40 to utilize more core area.
I set those 2 variables but set PL_TARGET_DENSITY = 0.4 (not 40). I was able to get further to [STEP 39] with the following openlane.log:
[INFO]: Writing Verilog...
[INFO]: Running LEF LVS...
[INFO]: Running Magic DRC...
[INFO]: Converting Magic DRC Violations to Magic Readable Format...
[INFO]: Converting Magic DRC Violations to Klayout XML Database...
[ERROR]: There are violations in the design after Magic DRC.
[ERROR]: Total Number of violations is 5
[INFO]: Saving current set of views in '../Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/results/final'...
[INFO]: Generating final set of reports...
[INFO]: Created manufacturability report at '../Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/reports/manufacturability.rpt'.
[INFO]: Created metrics report at '../Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/reports/metrics.csv'.
[INFO]: Saving runtime environment...
[ERROR]: Flow failed.
The manufacturability.rpt contained:
Design Name: sha1_top
Run Directory: /Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54
Magic DRC Summary:
Source: /Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/reports/signoff/drc.rpt
Violation Message "Min area of metal1 holes > 0.14um^2 (met1.7) "found 5 Times.
Total Magic DRC violations is 5
LVS Summary:
Source: /Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/logs/signoff/38-sha1_top.lvs.lef.log
LVS reports no net, device, pin, or property mismatches.
Total errors = 0
Antenna Summary:
No antenna report found.
If you update open_pdks to latest it may resolve this error.
I do have the latest version of OpenLane and still get the error. Any suggestions? Thanks!
OpenLane Git Version: f9b5781f5ef0bbdf39ab1c2bbd78be8db11b27f2
While creating this issue, you shared above OpenLane version.
Can you post latest OpenLane version do you tried?
sha1_top design tested at my end and am not seening magic drc error.
[INFO]: Running LEF LVS...
[STEP 43]
[INFO]: Running Magic DRC (log: logs/signoff/43-drc.log)...
[INFO]: Converting Magic DRC Violations to Magic Readable Format...
[INFO]: Converting Magic DRC Violations to Klayout XML Database...
[INFO]: No DRC violations after GDS streaming out.
[STEP 44]
[INFO]: Running OpenROAD Antenna Rule Checker (log: logs/signoff/44-antenna.log)...
[STEP 45]
[INFO]: Running CVC (log: logs/signoff/45-erc_screen.log)...
[INFO]: Saving current set of views in 'results/final'...
[INFO]: Saving runtime environment...
[INFO]: Generating final set of reports...
[INFO]: Created manufacturability report at 'reports/manufacturability.rpt'.
[INFO]: Created metrics report at 'reports/metrics.csv'.
[WARNING]: There are max fanout violations in the design at the typical corner. Please refer to 'reports/signoff/33-rcx_sta.slew.rpt'.
[INFO]: There are no hold violations in the design at the typical corner.
[INFO]: There are no setup violations in the design at the typical corner.
[SUCCESS]: Flow complete.
[INFO]: Note that the following warnings have been generated:
[WARNING]: There are max fanout violations in the design at the typical corner. Please refer to 'reports/signoff/33-rcx_sta.slew.rpt'.
I think this is fixed now?
|
2025-04-01T06:37:37.326790
| 2023-07-18T22:25:40
|
1810809629
|
{
"authors": [
"annapetrosyan26",
"maliberty",
"oharboe"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2882",
"repo": "The-OpenROAD-Project/OpenROAD",
"url": "https://github.com/The-OpenROAD-Project/OpenROAD/issues/3658"
}
|
gharchive/issue
|
make write_lef and others write out files atomically
Description
A half-written .lef file can be lying around after a crash which complicates e.g. deltaDebug.py.
Suggestion:
modify all write_* commands to write to a temp file
when the write completes, rename temp file to final name
This way there are no partial files lying around
Suggested Solution
No response
Additional Context
No response
Hello @maliberty @oharboe , I'd like to work on this issue. Could you please assign it to me ?
@annapetrosyan26 You can simply create a pull request without being assigned this task, I don't have the role to assign github issues in this project.
I haven't seen that @maliberty assigns github issues to new contributors, but perhaps that's something that would be approperiate?
Thank you @oharboe for the clarification. Then I'll work on the issue and once the code is ready I will create a pull request.
@maliberty Fixed, no?
Fixed in https://github.com/The-OpenROAD-Project/OpenROAD/pull/5109
|
2025-04-01T06:37:37.332806
| 2024-09-26T18:21:12
|
2551259012
|
{
"authors": [
"jeffng-or",
"oharboe"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2883",
"repo": "The-OpenROAD-Project/megaboom",
"url": "https://github.com/The-OpenROAD-Project/megaboom/pull/118"
}
|
gharchive/pull-request
|
updated SDCs to define both read and write clocks
PR to update SRAM SDC file to define both clocks
Isnt this also used by BoomTile? In which case we need an additional .sdc file.
Also, the BUILD bazel need to be updated to use the file.
If I read the BUILD.bazel correctly, the current constraints.sdc is used by the SRAM, regfile, and L1MetadataArray abstract generation flows. It is not used by BoomTile (it uses constraints-boomtile.sdc).
But, yes, the proposed change wouldn't work for dataArrayB, tag_array*, or L1MetadataArray, since their clocks are named either RW0_clk, R*_clk, W*_clk or just clock.
So, it seems like we might be able to:
Modify the existing constraints.sdc to use *_clk to find the clock ports and use it for all SRAMs and regfiles
Have L1MetadataArray continue to use the existing constraints-boomtile.sdc since they both have just "clock"
Please correct me if I'm wrong, but I think that will hook up the appropriate clocks for the abstract generation flow.
Yes, this should work.
lgtm.
should this have gone through jenkins CI?
Interesting, the Jenkins CI jobs got kicked off and passed:
https://jenkins.openroad.tools/job/megaboom-Public/view/change-requests/job/PR-118-head/
https://jenkins.openroad.tools/job/megaboom-Public/view/change-requests/job/PR-118-merge/
Maybe there needs to be some hook up in GitHub so that it shows up in the Rules section? @vvbandeira , how do we register the CI as a check?
|
2025-04-01T06:37:37.338989
| 2021-06-29T18:04:43
|
932972008
|
{
"authors": [
"Swastyy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2884",
"repo": "TheAlgorithms/C-Plus-Plus",
"url": "https://github.com/TheAlgorithms/C-Plus-Plus/pull/1518"
}
|
gharchive/pull-request
|
Create n_bonacci.cpp
Description of Change
Checklist
[x] Added description of change
[x] Added file name matches File name guidelines
[x] Added tests and example, test must pass
[x] Added documentation so that the program is self-explanatory and educational - Doxygen guidelines
[x] Relevant documentation/comments is changed or added
[x] PR title follows semantic commit guidelines
[x] Search previous suggestions before making a new one, as yours may be a duplicate.
[x] I acknowledge that all my contributions will be made under the project's license.
Notes:
@Panquesito7 See now this PR, I think you will be able to see it clearly now as I have recreated it from the correct fork.
Thanks in advance 😄
|
2025-04-01T06:37:37.369491
| 2022-09-01T15:19:13
|
1359021652
|
{
"authors": [
"TheArkive",
"gcailly"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2885",
"repo": "TheArkive/GuiCtlExt_ahk2",
"url": "https://github.com/TheArkive/GuiCtlExt_ahk2/issues/5"
}
|
gharchive/issue
|
No property named GuiControl
Hi,
(Thank you for all your AHK libraries !)
Here is an error message with AutoHotKey 2.0 beta 7.
Oh crud, I may not have tested that as thoroughly as I should have. Thanks for the report.
fixed
|
2025-04-01T06:37:37.412466
| 2021-03-22T13:57:02
|
837743819
|
{
"authors": [
"TheDiscordian"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2886",
"repo": "TheDiscordian/ipfs-sync",
"url": "https://github.com/TheDiscordian/ipfs-sync/issues/29"
}
|
gharchive/issue
|
Do less files/mkdir calls
Related to #28, currently we call files/mkdir for every single file updated/added. It'd speed up these operations a lot if we only did this once per directory.
For context I just attempted to update 18k files totalling ~1.1GiB, it's taking over 12 hours on a powerful system with an NVME SSD. This is obviously not good enough.
Each call can take 1-3s!!
Testing a possible patch for this right now. I think my set took 13 or 14 hours to process. I'll see how long this one takes.
Fixed in 02a37e96cf6e118647b0da899ca7573332f19fb8
|
2025-04-01T06:37:37.459292
| 2022-05-13T04:10:04
|
1234715041
|
{
"authors": [
"Spappz",
"TheGiddyLimit",
"jbowensii",
"revilowaldow"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2887",
"repo": "TheGiddyLimit/plutonium-addon-automation",
"url": "https://github.com/TheGiddyLimit/plutonium-addon-automation/issues/26"
}
|
gharchive/issue
|
Complex items and optionality
Many things in 5e contain options, whereby the AE to activate depends on a choice made. This leads to two options for data:
Split each option out into a different Foundry item (and therefore a different button on the sheet)
Use a macro to create a pop-up querying which option the player would like to trigger
The latter is neater and more intuitive imo, but it might be slow if a player never uses a certain option. We need to agree how to handle this—and whether we are happy to accept 'half-complete' items until a general solution is found—because there's probably a fucctonne of items that it applies to
Only one item please imo. You're only importing one thing from plutonium, you only expect one thing. So needs to be a macro. But also how would you implement that macro, is there a template you'd need to use for every item like that, or can you structure the data such that the module/plutonium generates a standard macro from an array of options?
is there a template you'd need to use for every item like that
Midi QoL has a (non-AE) option to trigger a macro when an item is activated (i.e. clicked). Using Item Macro, we can bundle this macro with the item itself rather than import it into the normal macros directory.
I agree it would be nice to have all the popups be 'consistent' with each other in styling, but the issue is that the macro will invariably have more programming than just the popup. That is, once the popup appears, you need to write code to handle what each button does, and that has to be included in the same macro! I suppose in principle the popup could be triggered by a core Plutonium function with reliable behaviour, so the PAD-only, item-specific macro only needs to call the initial 'create modal' function and provide instructions for each result.
await plutonium.optionsPopup.create(
"Body text <b>with HTML</b> and whatnot like.",
"Button 1", "function1",
"Button 2", "function 2"
);
function function1 {
executeCommands();
}
function function2 {
executeCommands();
}
function closePopup {
execute Commands();
}
I agree, one item/feature in the game equals one item on the character sheet, so please use macros. I personally do not see a reason to dictate code structure, format, or styling.
The reason to at least have a style guide is so other people can review the code easily, if it ever needs updates! This could be fixing a bug, an API change, Midi overhauling itself, etc.
The only requirements I'd like to 'enforce', to be clear, are:
Write comments (obvious)
No minified code unless the human-readable one is available somewhere else (see my post above)
Use the same pop-up formatting as everything else (which we can draft whether or not Giddy makes some Plutonium magic to make it easier; this is to provide a 'consistent' experience so users don't have to constantly reread every single pop-up because they're all phrased differently)
Whatever solution I end up cobbling together for jamming macros into proceedings, I'll make it eslint-able, so a minimum bar for code styling will be in place
Writing good code is a whole different kettle of fish, and (un)fortunately not machine-enforceable (yet 😓)
kinda made irrelevant by CPR/etc. managing it better than anything any of us clearly have the effort to maintain 😏
|
2025-04-01T06:37:37.469558
| 2024-12-06T21:32:17
|
2723964282
|
{
"authors": [
"fercer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2888",
"repo": "TheJacksonLaboratory/activelearning",
"url": "https://github.com/TheJacksonLaboratory/activelearning/issues/28"
}
|
gharchive/issue
|
Generated layers do not follow OME standard
The generated layers, when saved as zarr files, do not follow the OME standard and cannot be opened back again correctly in napari.
Solved in PR #29
|
2025-04-01T06:37:37.475736
| 2023-09-14T11:59:25
|
1896417738
|
{
"authors": [
"TheLartians",
"ldeng-ustc"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2889",
"repo": "TheLartians/ModernCppStarter",
"url": "https://github.com/TheLartians/ModernCppStarter/issues/178"
}
|
gharchive/issue
|
Suggestion to define main project name in a separate file
The main project name (Greeter) is currently used across multiple files. So it requires modifications in all these files to use this template, which can be time-consuming and error-prone.
I suggest defining the main project name in a separate file. This way, any changes to the name would only need to be made in one place, reducing potential errors and increasing efficiency.
for example, we can create a new file info.cmake:
# Note: update this to your new project's name and version
set(MAIN_PROJECT_NAME Greeter)
And then, include this file in other files:
# ./CMakeLists.txt
cmake_minimum_required(VERSION 3.14...3.22)
# ---- Project ----
include(info.cmake)
project(
${MAIN_PROJECT_NAME}
VERSION 1.0
LANGUAGES CXX
)
Then, one can start its own projects with changing only one Greeter.
I write a demo in this branch, but I haven’t updated the documentation. If this suggestion is accepted, I can submit a PR.
Hey thanks for the input! I've also been bugged by this, but decided to not bother as it would introduce extra complexity without not adding long term benefits. Also in an actual project, you would expect the project name to be hardcoded in multiple places instead of being encoded in a generic variable name.
|
2025-04-01T06:37:37.480357
| 2023-03-08T18:08:10
|
1615696830
|
{
"authors": [
"The-Great-Nothing",
"TheLastBen",
"blazing",
"cxyzdroid90"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2890",
"repo": "TheLastBen/fast-stable-diffusion",
"url": "https://github.com/TheLastBen/fast-stable-diffusion/issues/1705"
}
|
gharchive/issue
|
module not found error
after entering
Path_to_HuggingFace:XpucT/Deliberate
I'm encountering this error
ModuleNotFoundError Traceback (most recent call last)
in
3 from IPython.utils import capture
4 from IPython.display import clear_output
----> 5 import wget
6
7 #@markdown - Skip this cell if you are loading a previous session that contains a trained model.
ModuleNotFoundError: No module named 'wget'
NOTE: If your import is failing due to a missing package, you can
manually install dependencies using either !pip or !apt.
To view examples of installing some common dependencies, click the
"Open Examples" button below.
Same here.
Me too but yesterday it worked.
Solution: put !pip install wget before import wget.
Thank You! Still, later in the training cell an avalanche of errors happen so I guess we wait for the updated colab.
Oh yes I see, hope they will fix it ASAP.
fixed
|
2025-04-01T06:37:37.483560
| 2022-12-04T04:55:33
|
1474454438
|
{
"authors": [
"AlternativelyMaybe",
"TheLastBen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2891",
"repo": "TheLastBen/fast-stable-diffusion",
"url": "https://github.com/TheLastBen/fast-stable-diffusion/issues/859"
}
|
gharchive/issue
|
"LayerNormKernelImpl" not implemented for 'Half'
No flames please, be gentle, it's my first time posting. :)
I got this error message after successfully (so I thought) installing and then running a request: "LayerNormKernelImpl" not implemented for 'Half'.
Any words of wisdom? What else can I tell you to help someone help me out?
Thanks so much!
Where exactly do you get this error ?
After putting in search terms and hitting the Generate button. I just went back to the main screen and saw the code output that had been generated:
https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb#scrollTo=PjzwxTkPSPHf
Did this work? I copied and pasted but it wasn't getting the entire output as for some reason the paste results were truncated. I can also try copying to a .txt file if you need.
That phrase shows up a few times in the text at the end of each block.
Thanks!
For clear instructions how to use it, look for AUTOMATIC1111 colab tutorial on YouTube.
Thanks much, I'll look into that!
|
2025-04-01T06:37:37.494879
| 2019-06-19T08:59:56
|
457890301
|
{
"authors": [
"TheM4hd1",
"canaksoy",
"sbertix"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2892",
"repo": "TheM4hd1/SwiftyInsta",
"url": "https://github.com/TheM4hd1/SwiftyInsta/issues/25"
}
|
gharchive/issue
|
/feed/reels_media/
Great repo.
Is there any way we could have story highlights added as well?
make sure update to version 1.0.5
Example:
https://github.com/TheM4hd1/SwiftyInsta/blob/10db0b97b7508a5c492c9f178c35b63768b2811b/SwiftyInstaTests/SwiftyInstaTests.swift#L1330
Great!
Unfortunately it does not seem to populate items for every Higlight, only the first one, and not reliably either.
Quickly googling it seems to suggest that to retrieve the actual items (i.e. stories) we need to pass the id to the "reels_media" endpoint. Any plans to add that as well?
Thanks @TheM4hd1
did you try passing ID like this? highlight:123882132324123
I see, I'll check it for you.
Thanks man
There is an endpoint: /feed/reels_media/
it accepts an array of user_ids like highlight:123882132324123
I've tried to implement it but I'm receiving an error from server
{"message": "Invalid reel id list", "status": "fail"}
try it yourself, see if you can fix it, maybe I'm doing wrong somewhere.
{"message": "Invalid reel id list", "status": "fail"}
I have to admit that I've been trying to implement this for a while, without success. I was hoping you would fare better, but we seem to have stumbled on the same error unfortunately.
I'l let you know if I find anything, but as I see similar libraries used this endpoint but I didn't test them to see if they works or not.
maybe we're missing something... idk
I try with "supported_capabilities_new" but same response. anyone fixed?
sample request
POST: https://i.instagram.com/api/v1/feed/reels_media/
DATA: signed_body=bca1cf35fe840d851d3d37f488665e955fa517e06a521843b619fd44183e28e3.{"supported_capabilities_new":"[{"name":"SUPPORTED_SDK_VERSIONS","value":"9.0,10.0,11.0,12.0,13.0,14.0,15.0,16.0,17.0,18.0,19.0,20.0,21.0,22.0,23.0,24.0,25.0,26.0,27.0,28.0,29.0,30.0,31.0,32.0,33.0,34.0,35.0,36.0,37.0,38.0,39.0,40.0,41.0,42.0,43.0"},{"name":"FACE_TRACKER_VERSION","value":"10"},{"name":"segmentation","value":"segmentation_enabled"},{"name":"WORLD_TRACKER","value":"WORLD_TRACKER_ENABLED"}]","source":"feed_timeline","_csrftoken":"lQElejwdXHJToUUryVZWghOEN2X8GFn0","user_ids":["archiveDay:17960651299222776"],"_uid":"5889897609","_uuid":"1df4e0f8-fc98-4250-a6a4-56b455e75699"}&ig_sig_key_version=4
@TheM4hd1 how can we create signed_body with array inside. how to know which part will be sign?
@canaksoy
Here is an example of how to sign body.
https://github.com/TheM4hd1/SwiftyInsta/blob/c2beaad164a49b84fe8254af25ac75e36e192f72/SwiftyInsta/API/Services/UserHandler.swift#L164
The sample request you attached above, is it a working version?
I tried with signed_body samples but its hard to handle [String] with encoder. Yes its a working php sample. @TheM4hd1
@canaksoy
Okay, I'll check it.
Thanks.
I tested function with 2 id, if you see there are some missing data in returned model, you can decode the returned data to access full info about request.
|
2025-04-01T06:37:37.545882
| 2022-12-18T02:35:22
|
1501710358
|
{
"authors": [
"P4ranoidAndroid"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2893",
"repo": "TheRealArthurDent/renovate",
"url": "https://github.com/TheRealArthurDent/renovate/pull/32"
}
|
gharchive/pull-request
|
Update renovatebot/github-action action to v34.62.1
This PR contains the following updates:
Package
Type
Update
Change
renovatebot/github-action
action
minor
v34.60.0 -> v34.62.1
Release Notes
renovatebot/github-action
v34.62.1
Compare Source
See the the changelog for changes in all releases.
34.62.1 (2022-12-17)
Bug Fixes
deps: update renovate/renovate docker tag to v34.62.1 (afc5d3c)
v34.61.0
Compare Source
See the the changelog for changes in all releases.
34.61.0 (2022-12-17)
Bug Fixes
deps: update renovate/renovate docker tag to v34.61.0 (57b8e85)
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Enabled.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
[ ] If you want to rebase/retry this PR, check this box
This PR has been generated by Renovate Bot.
Branch automerge failure
This PR was configured for branch automerge. However, this is not possible, so it has been raised as a PR instead.
|
2025-04-01T06:37:37.851045
| 2024-07-11T15:36:57
|
2403502904
|
{
"authors": [
"TheRedHatter"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2894",
"repo": "TheRedHatter/CE",
"url": "https://github.com/TheRedHatter/CE/pull/501"
}
|
gharchive/pull-request
|
[Snyk] Security upgrade com.amazonaws:aws-java-sdk-core from 1.12.239 to 1.12.760
This PR was automatically created by Snyk using the credentials of a real user.
Snyk has created this PR to fix 61 vulnerabilities in the maven dependencies of this project.
Snyk changed the following file(s):
api/pacman-api-admin/pom.xml
Vulnerabilities that will be fixed with an upgrade:
Issue
Score
Upgrade
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-608664
780
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 Reachable Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-450917
705
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Mature
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-467015
675
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Mature
Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONCORE-3038426
670
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 Reachable Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1054588
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056416
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056418
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056420
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056421
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056426
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056427
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-174736
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-548451
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-559094
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-559106
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-560762
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561585
630
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1009829
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1047324
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056414
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056417
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056419
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056424
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056425
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-467016
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-560766
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561362
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561373
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561586
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561587
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-564887
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-564888
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-570625
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-572300
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-572314
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-572316
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72448
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72449
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72450
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72451
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72882
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72883
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72884
563
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
XML External Entity (XXE) Injection SNYK-JAVA-COMFASTERXMLJACKSONCORE-1048302
560
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1052449
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1052450
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1061931
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-455617
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-467014
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-469674
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-469676
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-471943
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-472980
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-540500
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-6056407
555
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONCORE-2421244
525
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONDATAFORMAT-1047329
525
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONCORE-3038424
520
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-450207
520
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept
Information Exposure SNYK-JAVA-COMMONSCODEC-561518
485
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 Reachable No Known Exploit
Improper Input Validation SNYK-JAVA-ORGAPACHEHTTPCOMPONENTS-1048058
415
com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit
[!IMPORTANT]
Check the changes in this PR to ensure they won't cause issues with your project.
Max score is 1000. Note that the real score may have changed since the PR was raised.
This PR was automatically created by Snyk using the credentials of a real user.
Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs.
For more information:
🧐 View latest project report
📜 Customise PR templates
🛠 Adjust project settings
📚 Read about Snyk's upgrade logic
Learn how to fix vulnerabilities with free interactive lessons:
🦉 Deserialization of Untrusted Data
🦉 XML External Entity (XXE) Injection
🦉 Denial of Service (DoS)
🦉 More lessons are available in Snyk Learn
Checkmarx One – Scan Summary & Details – f1ea51e7-8150-4369-b139-d654d1e8728c
New Issues
Severity
Issue
Source File / Package
Checkmarx Insight
Cleartext_Submission_of_Sensitive_Information
/lambda-functions/notification-es-logging-service/src/main/java/com/paladincloud/notification_log/config/AuthManager.java: 44
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/lambda-functions/notification-es-logging-service/src/main/java/com/paladincloud/notification_log/config/AuthManager.java: 44
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/commons/autofix/manager/AuthManager.java: 41
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-aqua-enricher/src/main/java/com/tmobile/cso/pacman/aqua/jobs/AquaDataImporter.java: 37
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/commons/autofix/manager/AuthManager.java: 41
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/commons/autofix/manager/AuthManager.java: 41
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-asset/src/main/java/com/tmobile/pacman/api/asset/service/AssetServiceImpl.java: 856
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/executor/PolicyExecutor.java: 141
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/executor/PolicyExecutor.java: 141
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 281
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 72
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 67
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 281
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 72
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 67
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/service/AmazonCognitoConnector.java: 121
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-vulnerability/src/main/java/com/tmobile/pacman/api/vulnerability/service/VulnerabilityService.java: 973
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 281
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 72
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 67
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 143
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 144
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 391
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 373
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 391
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 373
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 391
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 373
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/lambda-functions/notification-template-formatter-service/src/main/java/com/paladincloud/HttpUtil.java: 154
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/lambda-functions/notification-send-email-service/src/main/java/com/paladincloud/utils/HttpUtil.java: 152
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 132
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 125
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AuthManager.java: 22
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 125
Attack Vector
Cleartext_Submission_of_Sensitive_Information
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/TenableAccountServiceImpl.java: 119
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540
Attack Vector
Client_Potential_XSS
/webapp/src/app/shared/searchable-dropdown/searchable-dropdown.component.ts: 142
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21841
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21838
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21831
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21828
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21825
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21822
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger.js: 63
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20550
Attack Vector
Client_Potential_XSS
/commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20547
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 89
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AwsAccountServiceImpl.java: 213
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AwsAccountServiceImpl.java: 216
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 155
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 158
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154
Attack Vector
Privacy_Violation
/api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111
Attack Vector
More results are available on AST platform
|
2025-04-01T06:37:37.867481
| 2024-05-18T08:44:00
|
2303935190
|
{
"authors": [
"TheRedHatter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2895",
"repo": "TheRedHatter/vulhub",
"url": "https://github.com/TheRedHatter/vulhub/pull/61"
}
|
gharchive/pull-request
|
[Snyk] Security upgrade php from 7.1-apache to 7.4.33-apache
This PR was automatically created by Snyk using the credentials of a real user.
Keeping your Docker base image up-to-date means you’ll benefit from security fixes in the latest version of your chosen image.
Changes included in this PR
base/gitlist/0.6.0/Dockerfile
We recommend upgrading to php:7.4.33-apache, as this image has only 225 known vulnerabilities. To do this, merge this pull request, then verify your application still works as expected.
Some of the most important vulnerabilities in your base image include:
Severity
Priority Score / 1000
Issue
Exploit Maturity
929
Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740
Mature
929
Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740
Mature
929
Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740
Mature
929
Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740
Mature
886
Out-of-bounds Write SNYK-DEBIAN10-APACHE2-2322058
Mature
Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs.
For more information:
🧐 View latest project report
🛠 Adjust project settings
Note: This is a default PR template raised by Snyk. Find out more about how you can customise Snyk PRs in our documentation.
Learn how to fix vulnerabilities with free interactive lessons:
🦉 Server-Side Request Forgery (SSRF)
Checkmarx One – Scan Summary & Details – 7b6409b7-8ce8-417a-8cec-254bc00144e0
No New Or Fixed Issues Found
|
2025-04-01T06:37:37.905061
| 2016-06-08T19:08:29
|
159247251
|
{
"authors": [
"Cat-McCatface",
"Rumpelstiltskinny",
"TheTemportalist"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2896",
"repo": "TheTemportalist/EsoTeriCraft",
"url": "https://github.com/TheTemportalist/EsoTeriCraft/issues/8"
}
|
gharchive/issue
|
<EMAIL_ADDRESS>
Server crash on killing blow to EnderZoo Wither Witch while using Tconstruct broadsword.
fml-server-latest.log.txt
crash-2016-06-08_11.57.10-server.txt
Can you provide more ways and steps to reproduce?
I got a similar crash. Happened when I entered menu to pause the game for an unrelated issue of hostile mobs ignoring players. Last entity killed was the only actual hostile entity, ironically a bat (part of rough mobs).
forge: 1954
mc: 1.9.4
EsoTeriCraft-1.9.4-<IP_ADDRESS>
Origin-1.9.4-9.1.6
log.txt
|
2025-04-01T06:37:37.941625
| 2017-09-26T09:42:49
|
260558830
|
{
"authors": [
"bassosimone",
"coveralls"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2897",
"repo": "TheTorProject/ooni-backend",
"url": "https://github.com/TheTorProject/ooni-backend/pull/111"
}
|
gharchive/pull-request
|
fix(report/handlers): accept more semver versions
The current problem we have, specifically, is that since v1.2.0 the
mobile apps emit full semver including also the build.
This requires several parts of our infrastructure to improve the
regexp being used so to validate also this kind of input.
See also: measurement-kit/measurement-kit#1388
An earlier version of this diff was blessed by @hellais on Slack and
since then I just changed comments.
RET=$?
+RET=0
if [[ "$(ps -ef | awk -v pid=$prog '$2==pid{print}{}')" != "" ]]; then
kill -ALRM $prog
wait $prog
fi
+kill -ALRM 6046
ps -ef | awk -v pid=$prog '$2==pid{print}{}'
killitwithfire
++killitwithfire
++trap - ALRM
++kill -ALRM 6047
++kill -9 6049
++ps -ef
++awk -v pid=6047 '$2==pid{print}{}'
/home/travis/.travis/job_stages: line 57: 6046 Segmentation fault (core dumped) ./.travis.test.sh 30 ./bin/oonib
The command "chmod +x .travis.test.sh && ./.travis.test.sh 30 ./bin/oonib" exited with 139.
I don't understand very well what it could be, and it seems like this is something related to the way in which travis is handling the test. I am going to restart it to see if it's deterministic (I actually hope so).
I am experiencing more errors with the build, like that the keyserver doesn't know the key. 😡
Alright, it seems it's time to play golf.
Coverage increased (+0.04%) to 78.469% when pulling 9e0eca426cacf7ec4e301fd612ed25d9c824e22a on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master.
Coverage increased (+0.09%) to 78.527% when pulling ec91e00d53ea57413c7cb5301fc2d747433ee7d9 on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master.
Coverage increased (+0.04%) to 78.469% when pulling 9bb6db5e2a453911873e6fa069f12a85466473fe on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master.
I really hate these coveralls annoying comments that don't serve any purpose.
Coverage increased (+0.09%) to 78.527% when pulling d09c1c59e09743a3986ef27ca4d31f6878125b80 on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master.
In the end, I decided to rewrite the test that caused a segfault.
I believe the problem appeared now that travis has upgraded its infrastructure to 14.04: the previous travis build was for 03d738b and occurred on Apr, 6 when travis was still using 12.04.
The original script was brilliant: it did two nested waits to make sure the running process was either killed gracefully or with fire. But probably was too brilliant and triggered some edge case.
I did not want to wrestle too much with travis. Also, reading the script, it seems to me it's fine to rewrite it such that, if we cannot kill the background process, the build will hang and then fail (on travis).
I guess having the local build hang and the travis build fail is good enough for our purposes.
Alright, I have read the diff once more. I am going to self bless this as good, given that @hellais already blessed the diff improving the regexp on Slack and that the test changes "look good to me".
|
2025-04-01T06:37:37.955958
| 2022-10-04T04:52:48
|
1395664400
|
{
"authors": [
"NordicGamerFE",
"Speiger"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2899",
"repo": "TheUsefulLists/UsefulMods",
"url": "https://github.com/TheUsefulLists/UsefulMods/issues/137"
}
|
gharchive/issue
|
[Addition] World Pregenerators
Checklist
[X] I've checked that my mod suggestion hasn't already been added to one of the lists.
[X] I've checked the latest issues to make sure that this mod hasn't already been suggested.
What is your mod's name?
Chunk Pregenerator
What is your mod's main version?
Other (Forge)
What is your mod's other versions?
1.8.X or earlier, 1.12.X, 1.15.X (Forge), 1.16.X (Forge), 1.18.X (Forge), 1.19.X (Forge)
What is your mod's type?
Enhancement
What side dose your mod need to run on?
Worldgen lag is a real issue. Even in Singleplayer, and worldpregeneration is actually a really important thing you should do.
This can turn it from "unplayable" to "playable" especially if your core count (including threads) is stuck at 4 or less or the CPU isn't the newest anymore... And even then its suggested to do so.
That's why I am suggestion ChunkPregenerator.
Which supports 1.4.7-1.7.x-1.8.9-1.10.x-1.11.x-1.12.x-1.14.x-1.15x-1.16x-1.18x-1.19.x
Also another mod that I would suggest is Chunky
Which is another pregenerator that provides a lot of performance gains and covers
forge-fabric-customservers
while chunkpregenerator is dedicated for forge.
(OPTIONAL) Give some extra information about the mod.
Chunk Pregenerator is a tool for generating your world before you actually play it.
It also includes maintenance tools such as:
Chunk Deletion/Trimming
Performance Tracking (up to 1.12)
World Maintenance tools (up to 1.12)
Retro-generation that isn't relying on the Player Itself.
Harddrive protection (1.14 or newer where it becomes nessesary)
Memory Leak fixes that become apparent when pregeneration.
And a few other things.
I've added your mod to performance mods as i think that is a better fit then Enhancement. Love your mods, Keep up the good work :D
@NordicGamerFE
Small extra note since i saw this.
Know issues + Fixes are actually tracked by myself.
Here you find both lists.
1.12 or older: https://github.com/TinyModularThings/Chunk-Pregenerator-Issue-Tracker/issues/1
1.14 or newer: https://github.com/TinyModularThings/Chunk-Pregenerator-Issue-Tracker/issues/2
|
2025-04-01T06:37:37.962842
| 2024-11-29T12:42:40
|
2705012586
|
{
"authors": [
"AakashJaiswal-beta",
"freeboub"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2900",
"repo": "TheWidlarzGroup/react-native-video",
"url": "https://github.com/TheWidlarzGroup/react-native-video/issues/4313"
}
|
gharchive/issue
|
[BUG]: Mute/Volume 0.0 not working while screen recording
Version
6.8.2
What platforms are you having the problem on?
Android
System Version
All
On what device are you experiencing the issue?
Real device
Architecture
Old architecture
What happened?
I encountered an issue, when I screen record my application using, react native video,
Screen Recorded video have the sound of the muted videos.
In some phone the file is corrupted or is a lot laggy
The saved video in the gallery have alot of sounds all over the components where the video is muted and volume is 0;
DO NOTE: the sound is not being played in the app itself, muting and unmuting is working in the app,
BUT on screen recording it, the sound of all those unpaused and muted video is there in the recorded video.
The code snippet is:
video config= Platform.select({
ios: {
automaticallyWaitsToMinimizeStalling: false,
bufferConfig: {
minBufferMs: 2000,
maxBufferMs: 5000,
bufferForPlaybackMs: 1000,
bufferForPlaybackAfterRebufferMs: 2000,
},
},
android: {
bufferConfig: {
minBufferMs: 2000,
maxBufferMs: 5000,
bufferForPlaybackMs: 1000,
bufferForPlaybackAfterRebufferMs: 2000,
},
},
});
<Video
source={{
uri: videoUri,
}}
ref={ref => {
this.player = ref;
}}
onBuffer={this.onBuffer}
onError={this.videoError}
style={styles.thumbnailImage}
paused={isPaused}
repeat={true}
resizeMode="stretch"
muted={true}
{...videoConfig}
/>
I even tried adding volume: 0, but it disnot work as well
Reproduction Link
Reproduction
Step to reproduce this bug are:
Initially when my video is paused nothing is happening,
But when video is changed from paused to played, the video sound is coming in the screen recording of my app,
DO NOTE: the sound is not being played in the app itself, muting and unmuting is working in the app,
BUT on screen recording it, the sound of all those unpaused and muted video is there in the recorded video.
look strange, but I don't know how we can fix it...
I usually use another package to control device volume instead of player volume :/
maybe switching to texture view can fix the issue. you can have a try I think.
Tried doing it on adding a viewType prop and changing it to textured view, still I am facing the same issue and sometimes the video is not openable by the phone, it says corrupted file or screen recording cannot be played, even on showing it on meet or any other screen sharing calls, It is creating an issue
Ok, another solution can be to unselect audio tracks I think
|
2025-04-01T06:37:37.999878
| 2023-07-08T11:00:43
|
1794891640
|
{
"authors": [
"DerpDerpling",
"Therkelsen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2902",
"repo": "Therkelsen/echoes_of_the_wilderness",
"url": "https://github.com/Therkelsen/echoes_of_the_wilderness/issues/8"
}
|
gharchive/issue
|
Update Fadeless to 1.20.1
Currently, the mod is made for 1.20-Snapshot, and not 1.20.1.
Update to 1.20.1, when the update comes out.
Curse Forge Page
It works fine on 1.20.1, no update is required. Another mod it works well together with is remove reloading screen , which removes the resource pack reloading screen.
Great, I just noticed that it wasn't specifically for 1.20.1, so I figured that might cause issues.
Also yeah, that mod is also part of the pack :D
|
2025-04-01T06:37:38.039593
| 2022-03-07T17:59:16
|
1161745137
|
{
"authors": [
"Thiritin",
"mowny"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2904",
"repo": "Thiritin/identity",
"url": "https://github.com/Thiritin/identity/issues/10"
}
|
gharchive/issue
|
It should log you out of all devices instead.
Session handling is a bit weird anyway. After the session times out, a reload redirects to https://identity.eurofurence.org/auth/choose. Clicking "Login with existing account" gets me directly to the dashboard without having to actually log in.
Hey @mowny
Thanks for your comment. Yes that is wanted as the openid connect Server saves a cookie depending on if you habe set the remember me.
The clients all got a limited session time. Although that session time does not end the cookie session at the idp.
E.x. the idp may have a cookie lifetime of 180 days meanwhile the apps only got an hour.
The right solution should be to redirect when the session times out.
Fixed by implementing #11, sessions are not concern of the idp. So session management can only be done by the idp for the IDP Apps not for the Reg as example. A fix for this could be backchannel logout.
|
2025-04-01T06:37:38.041979
| 2015-03-27T01:51:32
|
64672965
|
{
"authors": [
"ThirteenAG",
"jm10087"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2905",
"repo": "ThirteenAG/Widescreen_Fixes_Pack",
"url": "https://github.com/ThirteenAG/Widescreen_Fixes_Pack/issues/28"
}
|
gharchive/issue
|
[GTA:SA] IVRadarScaling
Would it be possible to add the IVRadarScaling option to the GTA: San Andreas fix like in the Vice City and GTA 3 fixes?
I guess, but maybe later.
|
2025-04-01T06:37:38.111505
| 2023-07-29T00:09:54
|
1827229336
|
{
"authors": [
"scala-steward"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2907",
"repo": "ThoughtWorksInc/enableIf.scala",
"url": "https://github.com/ThoughtWorksInc/enableIf.scala/pull/127"
}
|
gharchive/pull-request
|
Update scalafmt-core to 3.7.11
About this PR
📦 Updates org.scalameta:scalafmt-core from 3.7.5 to 3.7.11
📜 GitHub Release Notes - Version Diff
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scalameta", artifactId = "scalafmt-core" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "org.scalameta", artifactId = "scalafmt-core" }
}]
labels: library-update, early-semver-patch, semver-spec-patch, commit-count:1
Superseded by #128.
|
2025-04-01T06:37:38.126839
| 2021-07-28T02:16:25
|
954407753
|
{
"authors": [
"Tienisto",
"ffshy1214"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2908",
"repo": "Tienisto/flutter-fast-i18n",
"url": "https://github.com/Tienisto/flutter-fast-i18n/issues/32"
}
|
gharchive/issue
|
'in' can't be used as an identifier because it's a keyword.
iso 3166 country code of India have problem, as below:
Thanks for this issue. I will probably add a rename from in to indian
By the way, it is more common to use language codes. Or language code + country code. Using country code only is pretty rare
Fixed in 5.0.3.
By the way, it is more common to use language codes. Or language code + country code. Using country code only is pretty rare.
Also be aware that country codes must be in uppercase, e.g. CN not cn. Otherwise it is interpreted as language code like zh of zh-Hant-TW
Thanks a lot for the quick response, and We'll consider your suggestion. Thanks
|
2025-04-01T06:37:38.127981
| 2018-09-25T03:52:26
|
363400142
|
{
"authors": [
"eteters"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2909",
"repo": "TigerHacks/app-ios",
"url": "https://github.com/TigerHacks/app-ios/pull/49"
}
|
gharchive/pull-request
|
Api/sponsors implementation
I did it I think... I may have missed some good error handling and may have left a todo in there somewhere but I'll let you guys tell me if things still need to change
Unless I'm reading this wrong (which I definitely could be) the only travis errors I'm getting are for naming enum cases with a lowercase letter, which I do to match up with the variable name in the structs and the name of the api fields for decodable to work. Can we add a thing to ignore the rule for enums? I'm hoping that would be the easiest fix.
|
2025-04-01T06:37:38.141527
| 2022-03-22T14:33:10
|
1176879889
|
{
"authors": [
"Shelnutt2",
"thetorpedodog"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2910",
"repo": "TileDB-Inc/TileDB-Cloud-Py",
"url": "https://github.com/TileDB-Inc/TileDB-Cloud-Py/pull/242"
}
|
gharchive/pull-request
|
Allow raw installation outside of the context of PyPI.
If the package is installed outside of PyPI or an active checkout
(e.g. via pip install [some-path]), a version.py file might not be
created. Use a local version number in that case.
@thetorpedodog I've used python setup.py develop --user as my go-to for a while. Are you trying to get it working from the top level repo folder without any install?
Figured out what spurred me to do this. It would cause an error if you pip-installed directly from git:
pip install --upgrade git+https://github.com/TileDB-Inc/TileDB-Cloud-Py.git@some-version-hash
|
2025-04-01T06:37:38.145059
| 2024-11-07T02:06:46
|
2639644963
|
{
"authors": [
"jdblischak"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2911",
"repo": "TileDB-Inc/cellxgene-census-feedstock",
"url": "https://github.com/TileDB-Inc/cellxgene-census-feedstock/pull/17"
}
|
gharchive/pull-request
|
Bump to 1.16.2 and update requirements
This PR does a few related things:
Bumps version to 1.16.2 (and resets build number to 0)
Builds for Python 3.12 (replaces #15 because 1.15.0 couldn't be built against 3.12)
Use the new conda-forge syntax to control and test the minimum supported Python version. Upstream supports >=3.10
Also note that this version 1.16.2 will not be able to be installed in the TileDB Cloud py39 environment since it requires py>=310. The goal is to install it in the updated py312 environment that is in progress
Confirmed that the version number is still working as expected (#12):
import: 'cellxgene_census'
+ python -c 'import cellxgene_census; print(cellxgene_census.__version__)'
1.16.2
+ pip check
No broken requirements found.
|
2025-04-01T06:37:38.147026
| 2018-12-17T11:28:16
|
391669454
|
{
"authors": [
"TimBarham",
"ruslan-bikkinin"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2912",
"repo": "TimBarham/send",
"url": "https://github.com/TimBarham/send/pull/1"
}
|
gharchive/pull-request
|
Fix security vulnerabilities
This PR fixes security vulnerabilities for current version of send-transform. All tests are passed.
If you're going to go the route of updating my fork, I'd rather keep it in sync with the original as much as possible.
That would mean applying my changes on top of the latest release of send. However, I notice you've updated some of the dependencies here to newer versions than those used by the latest release of send - does that mean the versions currently used by send have security vulnerabilities?
@TimBarham just a friendly ping on this.
|
2025-04-01T06:37:38.149785
| 2024-07-04T20:51:27
|
2391452582
|
{
"authors": [
"Christopher-Chianelli",
"triceo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2913",
"repo": "TimefoldAI/timefold-solver",
"url": "https://github.com/TimefoldAI/timefold-solver/pull/934"
}
|
gharchive/pull-request
|
fix: Do not throw an exception when the solution or entity classes are interfaces
The superclass of an interface is null. This caused this line of code to throw an exception for solution classes that are interface:
var superclass = bottomClass.getSuperclass();
lineageClassList.addAll(getAllAnnotatedLineageClasses(superclass, annotation));
Since getAllAnnotatedLineageClasses expected superclass to not be null. getAllAnnotatedLineageClasses now returns an empty list for null arguments.
I'm wondering if solutions as interfaces should be supported. What would be the use case? What would be the downsides?
Considering that in all the years we have not seen anyone ask us for this, maybe we don't need this.
|
2025-04-01T06:37:38.191346
| 2023-01-25T10:08:54
|
1556378592
|
{
"authors": [
"AurelVU",
"Tkko",
"bugrevealingbme"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2914",
"repo": "Tkko/Flutter_Pinput",
"url": "https://github.com/Tkko/Flutter_Pinput/pull/118"
}
|
gharchive/pull-request
|
replaced FocusTrap to TapRegion
I replaced FocusTrap to TapRegion because FocusTrap is not supporten now in Flutter 3.7.0 (https://github.com/flutter/flutter/pull/107262)
..
Hey @AurelVU, I'm sorry for closing your PR without merging it but I was already working on it.
The update is live on pub
Version 2.2.22
|
2025-04-01T06:37:38.192257
| 2023-02-23T10:38:02
|
1596608685
|
{
"authors": [
"StunninglyWrong"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2915",
"repo": "Tlesis/SquakePlusPlus",
"url": "https://github.com/Tlesis/SquakePlusPlus/pull/4"
}
|
gharchive/pull-request
|
Update to 1.19.3
Update required libraries.
By the way, do you ever plan on uploading this to either or both Curseforge and Modrinth? Also, what about cutting off sharking from the mod if it keeps being unusable?
|
2025-04-01T06:37:38.200529
| 2024-07-09T10:34:07
|
2397818486
|
{
"authors": [
"Chaotixu",
"TobyAdd",
"exploitle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2916",
"repo": "TobyAdd/GDH",
"url": "https://github.com/TobyAdd/GDH/issues/243"
}
|
gharchive/issue
|
frame stepper doesnt work in geode
annoying
frame stepper isn't even a feature in the mod yet.
https://github.com/TobyAdd/GDH/issues/198
the feature is back in 4.6.4
|
2025-04-01T06:37:38.247602
| 2019-10-10T20:23:34
|
505491054
|
{
"authors": [
"FutureFutureTo",
"TomLingham"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2917",
"repo": "TomLingham/Laravel-Searchy",
"url": "https://github.com/TomLingham/Laravel-Searchy/issues/103"
}
|
gharchive/issue
|
SQLite Testing Environment
Does this package work w/ a sqlite testing environment?
I'm getting the following error when using 'TEST' as the query string:
+errorInfo: array:3 [
0 => "HY000"
1 => 1
2 => "near "'T%E%S%T%'": syntax error"
]
The query in my controller is as follows:
$query = Searchy::search('products')
->fields('title', 'producer')
->query($q)
->getQuery()
->having('relevance', '>', 20)
->limit(20)
->pluck('id')
->toArray();
Unfortunately, this package only works with a MySQL database. It uses MySQL specific features to calculate match relevance.
|
2025-04-01T06:37:38.281488
| 2018-08-09T14:42:40
|
349159039
|
{
"authors": [
"LabShareLouie",
"TonyGermaneri"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2919",
"repo": "TonyGermaneri/canvas-datagrid",
"url": "https://github.com/TonyGermaneri/canvas-datagrid/issues/154"
}
|
gharchive/issue
|
Questions on columnOrder API
Hi @TonyGermaneri,
Thanks for the quick feature release for columns sorting.
I've been playing around with it, and am able to retrieve the order using grid.columnOrder but cannot set the order. Looking over the docs, I assumed grid.columnOrder([1, 0 , 2, 3]) would work but got the error:
grid.columnOrder is not a function
May I get an example of how to set column order?
Thanks,
Louis
PS: Theres a duplicate entry for columnOrder in the API documentation https://tonygermaneri.github.io/canvas-datagrid/docs/#canvasDatagrid.columnOrder
All fixed!
https://tonygermaneri.github.io/canvas-datagrid/docs/#canvasDatagrid.columnOrder
Thanks again for pointing this out.
|
2025-04-01T06:37:38.292460
| 2014-07-10T22:18:55
|
37610798
|
{
"authors": [
"TooTallNate",
"anprogrammer",
"bdunlay"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2920",
"repo": "TooTallNate/node-gyp",
"url": "https://github.com/TooTallNate/node-gyp/pull/473"
}
|
gharchive/pull-request
|
Adds support for custom GYP include
When developers have libraries in non-standard locations (or don't wish
to dump them into System32 on Windows), they need to run "node-gyp
configure", edit the generated files to add include paths, and then run
"node-gyp install"
This is doubly painful when you wanted to "npm install" something, but
it depends on a library being in a default path on your system.
I've added the ability for users to include an additional ".gyp" file in
all node-gyp builds by setting an environment variable
(NODE_GYP_ADDITIONAL_CONFIG), which they can use to configure
include/library paths. No longer must I manually install packages or
dump files into global include directories.
Has this been resolved in subsequent releases? I'm currently stuck trying to install packages that require libraries I've installed in custom locations.
The preferred way to do this is to add a common.gypi file to the root of your module. For example: https://github.com/TooTallNate/node-vorbis/blob/master/common.gypi
|
2025-04-01T06:37:38.333590
| 2023-01-22T11:12:17
|
1552080517
|
{
"authors": [
"TousstNicolas",
"Xyntexx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2921",
"repo": "TousstNicolas/JLC2KiCad_lib",
"url": "https://github.com/TousstNicolas/JLC2KiCad_lib/issues/33"
}
|
gharchive/issue
|
Proper Arc support for schematic and footprint
The arcs for the schematic and footprint are rarely properly placed. It needs to be correctly handled.
I pushed some work on the Fix_ARC branch, but I cannot figure out how to draw them, there are always components for which it does not work.
For now, I tried to use the GetCenterParam function, which is reversed from https://easyeda.com/editor/6.5.5/js/editorPCB.min.js
This function seems to return the coordinates of the center (or possibly the midpoint in some occasions ? ), and two angles, but when using theses to calculate the start point, end point and center, the result is inconsistent, sometimes it works, sometimes it doesn't.
The schematic equivalent is more consistent and seems to have less issues.
The following components have an arc in their footprint and could be used to test:
C55684 C185659 C86002 C312983 C1341701 C307522 C689358 C403695 C602208 C152951 C688068 C163798 C661330
This concern the h_ARC function in the footprint :
https://github.com/TousstNicolas/JLC2KiCad_lib/blob/b5c38c2beff6f710eb8ac427622384717539ce84/JLC2KiCadLib/footprint/footprint_handlers.py#L174-L278
I made my version of the code and it seems to be working pretty well so far:
https://github.com/Xyntexx/JLC2KiCad_lib/tree/my_ARC_Fixes
Thanks a lot for your work, merged in 12c6860aecd43dd8b4262347f74a3bb06de85959
|
2025-04-01T06:37:38.338065
| 2022-08-05T02:34:20
|
1329369236
|
{
"authors": [
"BCoyler",
"Disctanger",
"Narann"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2922",
"repo": "Tow-Boot/Tow-Boot",
"url": "https://github.com/Tow-Boot/Tow-Boot/issues/169"
}
|
gharchive/issue
|
USB Mass Storage mode - eMMC unrecognized with pinephoneA64
Successful Tow-Boot Install
- Tow-Boot 2021.10-004 was successfully installed on a pinephoneA64.
- I first verified booting pOS from eMMC.
- I then verified the Mass Storage mode successfully using a Windows10 host.
- USB device was recognized as a "PinePhone (A64)", and the eMMC partitions are viewable.
Mass Storage mode issue
- After install; rebooted phone with USB plug still plugged in, but battery out.
- Entered into Mass Storage mode, (blue LED).
- Windows host now reporting, "Unknown USB Device (Device Descriptor Request Failed)".
Mass Storage mode issue reoccurring across phone reboots
- Tried again by disconnecting USB cable, inserting battery, and rebooting phone.
- Entered into Mass Storage mode, (blue LED).
- Plugged in USB cable.
- Windows host still reporting, "Unknown USB Device (Device Descriptor Request Failed)".
@BCoyler
I was having very similar issue.
Turns out i connected my pinephone to USB 2.0 port of the host device.
Tried with different port(USB 3.0) port and device was recognized.
If you haven't already found the solution. Give above method a try maybe it will resolve your issue as well.
Had the same problem. For me, the problem cames from the provided (red) cable. Using another USB-C cable works at first try.
|
2025-04-01T06:37:38.341605
| 2023-11-27T20:00:02
|
2012999629
|
{
"authors": [
"louisfd"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2923",
"repo": "Tracel-AI/burn",
"url": "https://github.com/Tracel-AI/burn/pull/1006"
}
|
gharchive/pull-request
|
fix/docs/chunk
@dcvz
I was a bit late to review your chunk operation. It's very nice but I found something was wrong with the doc. I made a test that explains why I'm changing it. Is this the behaviour you want?
@dcvz now it should have the right behaviour
|
2025-04-01T06:37:38.383426
| 2023-08-29T13:19:43
|
1871662043
|
{
"authors": [
"CLAassistant",
"Fluder-Paradyne",
"andreacasarin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2924",
"repo": "TransformerOptimus/SuperAGI",
"url": "https://github.com/TransformerOptimus/SuperAGI/pull/1149"
}
|
gharchive/pull-request
|
Using a standard .env file instead of config.yaml for docker deployment.
Refactoring docker image example docker deployments to take a standard .env file with variables defined in the docker-compose itself to simplify docker custom deployments (eg. on Portainer).
Description
You will customize the installation via .env file instead of YAML.
Related Issues
No direct issue, just a Discord conversation.
Solution and Design
This setup helps deployments via docker for users which have a production-like environment. It's focused on the docker installation, it will not impact non-docker usage.
Test Plan
Type of change
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[x] Breaking change (fix or feature that would cause existing functionality to change)
[ ] Docs update
Checklist
[x] My pull request is atomic and focuses on a single change.
[x] I have read the contributing guide and my code conforms to the guidelines.
[x] I have documented my changes clearly and comprehensively.
[ ] I have added the required tests.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
With the lastest release I'm having another issue here:
Traceback (most recent call last):
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 145, in __init__
self._dbapi_connection = engine.raw_connection()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 3288, in raw_connection
return self.pool.connect()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 452, in connect
return _ConnectionFairy._checkout(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 1267, in _checkout
fairy = _ConnectionRecord.checkout(pool)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 716, in checkout
rec = pool._do_get()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/impl.py", line 284, in _do_get
return self._create_connection()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 393, in _create_connection
return _ConnectionRecord(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 678, in __init__
self.__connect()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 902, in __connect
with util.safe_reraise():
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/util/langhelpers.py", line 147, in __exit__
raise exc_value.with_traceback(exc_tb)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 898, in __connect
self.dbapi_connection = connection = pool._invoke_creator(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/create.py", line 637, in connect
return dialect.connect(*cargs, **cparams)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/default.py", line 615, in connect
return self.loaded_dbapi.connect(*cargs, **cparams)
File "/opt/venv/lib/python3.10/site-packages/psycopg2/__init__.py", line 122, in connect
conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
psycopg2.OperationalError: could not translate host name "super__postgres" to address: Name or service not known
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/opt/venv/bin/alembic", line 8, in <module>
sys.exit(main())
File "/opt/venv/lib/python3.10/site-packages/alembic/config.py", line 632, in main
CommandLine(prog=prog).main(argv=argv)
File "/opt/venv/lib/python3.10/site-packages/alembic/config.py", line 626, in main
self.run_cmd(cfg, options)
File "/opt/venv/lib/python3.10/site-packages/alembic/config.py", line 603, in run_cmd
fn(
File "/opt/venv/lib/python3.10/site-packages/alembic/command.py", line 385, in upgrade
script.run_env()
File "/opt/venv/lib/python3.10/site-packages/alembic/script/base.py", line 582, in run_env
util.load_python_file(self.dir, "env.py")
File "/opt/venv/lib/python3.10/site-packages/alembic/util/pyfiles.py", line 94, in load_python_file
module = load_module_py(module_id, path)
File "/opt/venv/lib/python3.10/site-packages/alembic/util/pyfiles.py", line 110, in load_module_py
spec.loader.exec_module(module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/app/migrations/env.py", line 94, in <module>
run_migrations_online()
File "/app/migrations/env.py", line 82, in run_migrations_online
with connectable.connect() as connection:
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 3264, in connect
return self._connection_cls(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 147, in __init__
Connection._handle_dbapi_exception_noconnection(
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 2426, in _handle_dbapi_exception_noconnection
raise sqlalchemy_exception.with_traceback(exc_info[2]) from e
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 145, in __init__
self._dbapi_connection = engine.raw_connection()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 3288, in raw_connection
return self.pool.connect()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 452, in connect
return _ConnectionFairy._checkout(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 1267, in _checkout
fairy = _ConnectionRecord.checkout(pool)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 716, in checkout
rec = pool._do_get()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/impl.py", line 284, in _do_get
return self._create_connection()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 393, in _create_connection
return _ConnectionRecord(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 678, in __init__
self.__connect()
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 902, in __connect
with util.safe_reraise():
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/util/langhelpers.py", line 147, in __exit__
raise exc_value.with_traceback(exc_tb)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 898, in __connect
self.dbapi_connection = connection = pool._invoke_creator(self)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/create.py", line 637, in connect
return dialect.connect(*cargs, **cparams)
File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/default.py", line 615, in connect
return self.loaded_dbapi.connect(*cargs, **cparams)
File "/opt/venv/lib/python3.10/site-packages/psycopg2/__init__.py", line 122, in connect
conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
sqlalchemy.exc.OperationalError: (psycopg2.OperationalError) could not translate host name "super__postgres" to address: Name or service not known
Looks like alembic.ini has hardcoded db credentials.
We have a pr for this,
https://github.com/TransformerOptimus/SuperAGI/pull/1136
waiting on its review @andreacasarin
What if I make a new docker compose definition, something like: docker-compose.prod.yaml which has the preferred vector store and explicit env vars for production like deployments? I can let the example there and add a new one.
I think it would also be great to have a nginx image built so that you can actually deploy without pulling the whole repo.
Perfect then, I'll restore the example and go with docker-compose.prod.yaml, just a couple more questions:
which would be the preferred vector store?
i'd keep the env file minimal, just the required env vars to start the project, then if I got this right, the rest can be configured via gui, right?
Also, we still need https://github.com/TransformerOptimus/SuperAGI/pull/1149#issuecomment-1698503784 done to make it viable.
which would be the preferred vector store?
By default we are using redis vector store
keep .env.dist the same, because we can set variables from the gui, but the vector stores
in docker hub we have setup autobuilds so every push to main and dev gets build
which would be the preferred vector store?
By default we are using redis vector store
Ok, but does it work adding knowledge to that? Because in my installation it asks to configure a new vector store to add knowledge.
keep .env.dist the same, because we can set most variables from the gui, but the vector stores
I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications.
in docker hub we have setup autobuilds so every push to main and dev gets build
Ok!
Pushed, it requires nginx/DockerfileNginx to be added to the autobuild as superagi/superagi-proxy.
Pushed, it requires nginx/DockerfileNginx to be added to the autobuild as superagi/superagi-proxy.
i'll add it
keep .env.dist the same, because we can set most variables from the gui, but the vector stores
I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications.
To explain a bit more my point-of-view:
I'd go with 1 base docker compose with the bare minimum and then maybe a couple more which will be used to override the base one for different applications (like local dev, local llm, prod) they could only override the base one in specific keys;
we'll discuss more towards this, there are some issues with gui hmr
I'd add a .env.dist with the minimum variables needed to start the project and point out that you can actually configure all the rest from the gui.
The reason why I wanted to keep env.dist with all the keys is to let users know all the keys in the project in one place
we'll comment everything out other than the bare minimum
Just my 2 cents, I'll stop arguing :)
No issues
which would be the preferred vector store?
By default we are using redis vector store
Ok, but does it work adding knowledge to that? Because in my installation it asks to configure a new vector store to add knowledge.
sorry forgot about knowledge, asked @Tarraann ( main contributor to knowledges )
to use knowledge for now only pinecone, weviate and qdrant are supported
so keeping the weviate commented out should be good for users to quickly set it up
keep .env.dist the same, because we can set most variables from the gui, but the vector stores
I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications.
in docker hub we have setup autobuilds so every push to main and dev gets build
Ok!
keep .env.dist the same, because we can set most variables from the gui, but the vector stores
I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications.
To explain a bit more my point-of-view:
I'd go with 1 base docker compose with the bare minimum and then maybe a couple more which will be used to override the base one for different applications (like local dev, local llm, prod) they could only override the base one in specific keys;
we'll discuss more towards this, there are some issues with gui hmr
👍
I'd add a .env.dist with the minimum variables needed to start the project and point out that you can actually configure all the rest from the gui.
The reason why I wanted to keep env.dist with all the keys is to let users know all the keys in the project in one place we'll comment everything out other than the bare minimum
I see, looks fine to me, I took the time to rearrange the variables there and put them under some comment headings, that might help this process.
|
2025-04-01T06:37:38.418044
| 2024-03-21T09:02:21
|
2199584151
|
{
"authors": [
"oscorops"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2926",
"repo": "Tresjs/tres",
"url": "https://github.com/Tresjs/tres/issues/594"
}
|
gharchive/issue
|
如何设置背景颜色为透明?
Description
如何设置背景颜色为透明?
Suggested solution
如何设置背景颜色为透明?
Alternative
No response
Additional context
No response
Validations
[X] I agree to follow this project's Code of Conduct
[X] Read the Contributing Guidelines.
[X] Read the docs.
[X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate.
Ok its <TresCanvas clearColor: ' '>
|
2025-04-01T06:37:38.551890
| 2022-08-20T12:59:02
|
1345159555
|
{
"authors": [
"akinaritsugo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2927",
"repo": "TripInsurance/devopsoh78828",
"url": "https://github.com/TripInsurance/devopsoh78828/pull/3"
}
|
gharchive/pull-request
|
Features/challenge4/userprofile
PR Template
Purpose
...
Does this introduce a breaking change?
[ ] Yes
[ ] No
Pull Request Type
What kind of change does this Pull Request introduce?
[ ] Bugfix
[ ] Feature
[ ] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Documentation content changes
[ ] Other... Please describe:
How to Test
Get the code
git clone [repo-address]
cd [repo-name]
git checkout [branch-name]
npm install
Test the code
What to Check
Verify that the following are valid
...
Other Information
NG
|
2025-04-01T06:37:38.556185
| 2023-07-22T03:42:16
|
1816594901
|
{
"authors": [
"TripitakaBC",
"spilegi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2928",
"repo": "TripitakaBC/ampinvt_esphome",
"url": "https://github.com/TripitakaBC/ampinvt_esphome/issues/4"
}
|
gharchive/issue
|
Multiple Charge Controllers
Sorry for raising an issue for this, but I have multiple Charge controllers. You can assign them numbers. I don't see anywhere in the code where it's referencing the unit number (It's 1 by default)
Is it possible to specify that in the protocol?
Not a problem!
The charge controller is referenced in the hex code string sent over UART by the ESP32. Look for this code block in your YAML file:
time: - platform: homeassistant id: esptime on_time: - seconds: 0,30 then: - uart.write: id: uart_bus data: [ 0x01, 0xB3, 0x01, 0x00, 0x00, 0x00, 0x00, 0xB5 ] # Reads only real-time data
I think the controller address is the first hex block so you would need to change that and recalculate the checksum. That part is pretty easy. The bigger part of the work is that you would need to duplicate the ampinvt.h file to create separate sensors for each charge controller. If you don't need or want all the sensors then you could cut a lot of it out and just keep the sensors you did want but essentially, you have two parts to this; 1) sending the command to each controller (easy and quick) and 2) processing the response from each controller (harder and more time).
Thanks, I'll go ahead an give it a shot. I currently have 6 of these, and was hoping to not have to purchase 6 ESP32's. Lot's of cabling and mess.
Two questions
I'm a Network Engineer who only dabbles in Computer Science. How would I calculate the checksum?
If I were to duplicate only specific sensors, how would I differentiate between which charge controller it was coming from? I do not see a reference to the controller address in the ampinvt.h file.
Thanks,
PS. I'm ok with being told it's too far over my head, in that case I'll just purchase more ESP32's and do it that way. ESP32 I have setup connect to one of my charge controllers has been rock solid this entire time during testing.
One ESP32 will do the job and even 1 ampinvt.h file but what you will need is to setup the ampinvt.h to read 6 x sensor sets and replicate all those custom sensors in your YAML file.
If you go the 6 x ESP32 setup, you will still need to recode the sensors because they all come together in HA and HA will not know how to process 6 instances of the same sensors. When you figure out the renaming for those, just work backwards to the ampinvt sensor naming.
|
2025-04-01T06:37:38.567278
| 2015-11-04T17:20:43
|
115102875
|
{
"authors": [
"ItzDan",
"TheSpaceArmy"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2929",
"repo": "Tromino/PolyExtend",
"url": "https://github.com/Tromino/PolyExtend/issues/12"
}
|
gharchive/issue
|
Team groups going off edge of page.
Most likely known of but here: http://i.imgur.com/c4aJxd0.png
its a Known Bug <B
Would love This Fixed https://surl.im/i/8qvwj
|
2025-04-01T06:37:38.601633
| 2017-05-11T16:17:47
|
228047273
|
{
"authors": [
"coveralls",
"kevinansfield"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2930",
"repo": "TryGhost/Ghost-Admin",
"url": "https://github.com/TryGhost/Ghost-Admin/pull/690"
}
|
gharchive/pull-request
|
[WIP] ✨ use markdown-it for markdown previews
refs https://github.com/TryGhost/Ghost/pull/8451
replaces SimpleMDE's default marked rendering with markdown-it
adds markdown-it plugins to more closely match legacy Showdown behaviour
footnotes
highlight/mark
named headers
don't require a space after the # for headers
adds ember-browserify so that markdown plugins that only provide CommonJS modules can be imported
Coverage decreased (-0.2%) to 71.556% when pulling f190198f49cb38b32133abd721f528564f89d1d1 on kevinansfield:markdown-it into fbb46dc72c4380a632e73db37c4ae8f2370b2087 on TryGhost:master.
Coverage decreased (-0.2%) to 71.556% when pulling 0f7dd8a26025fd7a05a389190746bebcc077cee1 on kevinansfield:markdown-it into 627a71e1a4ea95a9d0b61dd92b90c0d822c2f907 on TryGhost:master.
|
2025-04-01T06:37:38.630936
| 2020-04-09T16:40:23
|
597410275
|
{
"authors": [
"gargol",
"matthanley",
"naz",
"vikaspotluri123"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2931",
"repo": "TryGhost/Ghost",
"url": "https://github.com/TryGhost/Ghost/issues/11729"
}
|
gharchive/issue
|
[API] Ordering by meta_title is not possible
Issue Summary
It's not possible to make an API request (internally or externally) that sorts by meta_title. Moreover, sorting by a related table (e.g. posts_meta) is allowed, but the request fails. I think this is a regression from 2.x, after the post metadata got split out into a new table.
Technical details:
Ghost Version: 3.13.1
Forum Ref: https://forum.ghost.org/t/order-by-meta-title-doesnt-seem-to-work/13349
Possible related issue:
The model base parseOrderOptions pulls in a list of permitted attributes from the child modal, including allowed relations. This allows you to set the order to one of those relations, which kind-of breaks the API request. For example, ordering by authors asc will give you a 400, but title asc works perfectly fine, and potato asc succeeds because you can’t sort by potatoes, and therefore the order is ignored.
Might be related to this issue - https://github.com/TryGhost/Ghost/issues/11572, should be checked when investigated deeper.
After deeper digging into the issue, confirming that problem is how we create the order query here and the way Post model's permittedAttributes function returns only posts table specific fields ignoring posts_meta fields.
A closest candidate to explore a solution is tinkering bookshelf-relations plugin, which should be able to abstract the table split in a way that permittedAttributes gives back correct fields from posts_meta .
This issue needs appetite do dig into possible solution through bookshelf-relations. Guesstimating 1 day.
Note, while loading up on the context have observed another place which will need an update when the ordering is implemented - pagination plugin. Specifically, options.order should most likely come with resolved tableName because pagination plugin should not know about relations and not do the tableName + '.' + property concatenation.
Braindump after playing extensively with ordering
There are 2 separate problems that this issue raises:
Main one - cannot order post resource by a legitimate field, which comes from a 1:1 relation (posts:posts_meta tables).
Ordering behaves inconsistently for fields that are not meant to be ordered by: ignores unknown properties and returns 400 when trying to order by property like authors, which comes from a relation name. In case of any field that can't be ordered by, it should consistently ignore that field (throwing a validation error would break API compatibility, so this is a no-go)
The key to both problems is how parseOrderOption function relies on permittedAttributes to calculate all the fields that could be ordered upon.
(pt.1) Ordering by fields which come from 1:1 relation
The missing piece here is a mechanism in model layer which would recognize a 1:1 relation and extract "orderable" fields out of related table. Additionally, the fields would have to come in a format that includes related table name so that pagination plugin could correctly form the query, e.g: ['posts_meta.meta_title ASC'] instead of ['meta_title ASC'].
The solution I'm thinking of here is adding a orderAttributes method which would be used instead of permittedAttributes in parseOrderOption. orderAttributes would have the same values as permittedAttributes for tables with no additional relations logic, and would have special overrides on each model with "expanded" field names coming from related table.
To make above solution more maintainable would need to create a declarative way of describing which fields from related model the orderAttributes could be taken from. One of the possible directions to explore is expanding hasOne relation in bookshelf-relations - plan to timebox this direction to half a day tomorrow (cc @matthanley).
(pt. 2) Treating all non-orderable fields the same
The solution here somewhat relates to the solution from pt. 1. When parsing order options the function should not rely on permittedAttributes because those are not the same fields that could be "ordered" upon. Adding a orderAttributes method to the base model would possibly solve the problem.
Thoughts on holistically solving posts<>posts_meta ordering/filtering/change detection problems
I've dug through bookshelf-relations aiming to figure out an abstraction which would allow solving current ordering, related filtering, and change-detection(1) problems(2). I have not found a way that would solve all of them in a holistic way.
These are 3 distinct areas that have some or no relation to bookshelf-relations:
Ordering - has to do with findPage method from pagination plugin and depends on buggy/outdated handling in parseOrderOption function. Nothing to do with bookshelf-relations because the ordering fields should be calculated independently IMO not just for Post model but all other models that have relations.
Filtering - problems here stem from lack of explicit NQL configuration for posts_meta table in combination with mapping posts_meta filter fields to correct ones (using posts_meta. prefix). This kind of mapping should be done in serialization layer same way other field mappings are done there.
Change detection - to fix this problem on a deeper level (the fix that was done through override to wasChanged() was a "patch") we could look into fixing bookshelf-relations change tracking. This might come through some special parameter passed with hasOne relation that posts_meta is declared with or invent a whole new relation for this specific situation. New relation could also help with elimination of patchwork that has to be done in Post model. This needs to be researched.
Additional note, bookshelf-relations inherently deals with create/update/delete operations only and doesn't have anything to do with read operations that are needed in case of ordering and filtering.
Conclusion with regards to current issue
Two problems that were described in the comment above should be solved outside of bookshelf-relations. This proposed solution:
adding a orderAttributes method which would be used instead of permittedAttributes in parseOrderOption. orderAttributes would have the same values as permittedAttributes for tables with no additional relations logic, and would have special overrides on each model with "expanded" field names coming from related table.
seems like the most viable solution for now and would solve both problems.
@matthanley would love to know if you have any feedback on this? I estimate implementing the proposed solution would take about a day.
adding a orderAttributes method which would be used instead of permittedAttributes in parseOrderOption. orderAttributes would have the same values as permittedAttributes for tables with no additional relations logic, and would have special overrides on each model with "expanded" field names coming from related table
@naz this looks like a reasonable approach to me 👍
After implementing the ordering idea I had in mind, got faced with a Bookshelf limitation I completely forgot existed. Have my experimentation available on this spike branch - https://github.com/TryGhost/Ghost/compare/master...naz:ordering-for-posts-meta-fields.
To sum up the problem. In pagination plugin, when self.fetchAll is done it doesn't have post_meta table loaded into the query and it fails with "no such column" error (when trying to order by posts_meta.meta_title for example). The problem of not having loaded relations when fetching records in Bookshelf is summed up with references here - https://github.com/bookshelf/bookshelf/issues/1707#issuecomment-351026830.
There are two possible ways to get around this problem which I've been thinking of:
Extend pagination with detection of relations and build in a join to related table when the ordering query is built. We extend the query in similar way in filtering plugin through NQL (it adds joins and filters records at the same stage). The downside for this method, is possibly keeping yet another configuration that might be similar to one in filter. This could end up hard to maintain longterm.
Explore fixing/extending Bookshelf itself and load up relations like posts_meta into the queryBuilder automatically. This potentially, would be a more maintainable approach. The downside is, I have no clue how hard this would be to achieve, would need to dive into bookshelf codebase to understand more. The problem seems to have been around in Bookshelf since 2014.
@matthanley the point Hannah made earlier about approaching the problem through bookshelf-relations is now clearer to me. Probably the limitation I've rediscovered was the reason. This issue could become a scope creep, so I think approaching it by timeboxing and researching possible directions would be best approach forward. I'm thinking of timeboxing first approach above to half a day and see what comes out and then if we really end up with need for configs would research the second approach. Let me know what you think or have any questions about the issue itself!
Summary of the discussion around future plan for this issue
Short term plan (to be done now) would be implementing a solution which expands query builder object inside of pagination (or ordering) plugin based on additional configurations similar to ones done with NQL/filtering.
Long term, aiming to develop a maintainable solution which is not based on configuration but rather on special relation type coming from Bookshelf or some other fix which might also address https://github.com/bookshelf/bookshelf/issues/202. Will timebox 1 day for research to figure out right questions and maybe possible solutions. A solution that is in mind right now is expansion of hasOne relation with an "alwaysFetched" parameter - this might be done through a plugin or from within bookshelf itself (needs digging around).
Extend pagination with detection of relations and build in a join to related table when the ordering query is built. We extend the query in similar way in filtering plugin through NQL (it adds joins and filters records at the same stage). The downside for this method, is possibly keeping yet another configuration that might be similar to one in filter. This could end up hard to maintain longterm.
This solution has landed in master through https://github.com/TryGhost/Ghost/pull/12226.
Next up will be pushing pt. 2 and figuring out more generic way to handle relation inclusion in query builder object through Bookshelf/special plugin. This should help solve filtering/ordering and attribute change tracking all together.
|
2025-04-01T06:37:38.637726
| 2022-04-18T22:17:12
|
1207504489
|
{
"authors": [
"ErisDS",
"guidefox"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2932",
"repo": "TryGhost/Ghost",
"url": "https://github.com/TryGhost/Ghost/issues/14508"
}
|
gharchive/issue
|
Impersonation tokens do signup action if no account is found
Issue Summary
When a user is deleted it is expected that all tokens associated with that account are revoked. This is currently not the case.
Creating a token before deleting a user creates the account again when pasted into the browser.
Steps to Reproduce
Create an access token for an account.
Delete that account.
Paste that token into your browser.
User is created again.
The user should NOT be created again and all tokens should be revoked when a user is deleted.
Ghost Version
4.44.0
Node.js Version
16.14.2
How did you install Ghost?
OS - Debian 11 with MariaDB 10.5.15
Database type
MySQL 8
Browser & OS version
No response
Relevant log / error output
No response
Code of Conduct
[X] I agree to be friendly and polite to people in this repository
Hey there @guidefox. Ghost's magic links are based on JWTs, the tokens aren't stored and there's not really a concept of revocation here. What's happening is that the magic link has a fall back behaviour of creating a new account if no matching account is found.
I realise that's a little jarring, but it's a brand new account that is created, not an old one being restored.
I think it would make sense to pin the impersonation links to only be allowed to do signin, rather than falling back to signup, to make this a little less weird.
Ah, that makes more sense. I think that the current behavior can be improved because it is a little bit jarring right now.
Perhaps a dedicated button for having impersonation tokens re-create the account instead of doing it automatically would be a better solution.
And maybe make the tokens one use only? or at least provide the option to have it expire after one use.
This has cropped up in other forms recently, and is something we want to prioritise fixing.
|
2025-04-01T06:37:38.665761
| 2023-07-27T02:22:40
|
1823498846
|
{
"authors": [
"ErisDS",
"Grasume",
"TheLaurenBarger",
"daniellockyer",
"davedub",
"hussainb",
"joe-blocher",
"kilmarnock",
"vikaspotluri123"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2933",
"repo": "TryGhost/Ghost",
"url": "https://github.com/TryGhost/Ghost/issues/17514"
}
|
gharchive/issue
|
Make this site private - not working in Chrome and Opera
Issue Summary
Change or delete the row 58:
versions/5.54.4/core/frontend/apps/private-blogging/lib/middleware.js
return session({
name: 'ghost-private',
maxAge: constants.ONE_MONTH_MS,
signed: false,
sameSite: 'lax' <----- row 58: instead of 'none' or delete row
})(req, res, next);
Or you can delete the row 58 because sameSite: 'Lax' is the default value.
You can't code 'secure' within an object - secure: true will not work.
Works now in Chrome and Opera.
See https://web.dev/i18n/en/samesite-cookies-explained
Steps to Reproduce
See https://forum.ghost.org/t/make-this-site-private-not-working/39938/1
Ghost Version
5.54.4
Node.js Version
v18.15.0
How did you install Ghost?
local, macos
Database type
SQLite3
Browser & OS version
No response
Relevant log / error output
No response
Code of Conduct
[X] I agree to be friendly and polite to people in this repository
In core/frontend/apps/private-blogging/lib/middleware.js:
´´´
const privateBlogging = {
…
return session({
name: 'ghost-private',
maxAge: constants.ONE_MONTH_MS,
signed: false,
// sameSite: 'none' <——— replace this with 2 lines below
sameSite: urlUtils.isSSL(config.get('url')) ? 'none' : 'lax',
secure: urlUtils.isSSL(config.get('url'))
})(req, res, next);
},
´´´
and all is fine!
Implement in the same way as you did in
core/server/services/auth/session/express-session.js
Hey there, thank you so much for the detailed bug report.
That does look like something that shouldn't happen! A PR to fix this issue would be very welcome 🙂
I have made the PR
Hi, I am facing the same issue in the latest version of Ghost, unable to login to the private site using Chrome based browsers.
It's nearly three month later ... an nothing happened. But for me it closed, because I'm working locally :-)
https://forum.ghost.org/t/make-this-site-private-not-working/39938
It's nearly three month later ... an nothing happened. But for me it closed, because I'm working locally :-) https://forum.ghost.org/t/make-this-site-private-not-working/39938
yeah, they didn't care to merge it. but I appreciate you for your troubleshooting and the fix.
I think just a rerun would be required to pass the build, otherwise the PR is already approved:
https://github.com/TryGhost/Ghost/actions/runs/6057836235/job/16821076886?pr=17938
I have made my first PR.
I think the 2. one failed: Merge branch 'main' into joe-blocher-patch-1 https://github.com/TryGhost/Ghost/pull/17938/commits/ae0f64eb578f4aa43248da1b3e807a1f0c3b9bef
I don't really know what is for and how I can delete this PR...
Am 16.10.2023 um 08:45 schrieb Hussain @.***>:
It's nearly three month later ... an nothing happened. But for me it closed, because I'm working locally :-) https://forum.ghost.org/t/make-this-site-private-not-working/39938 https://forum.ghost.org/t/make-this-site-private-not-working/39938
yeah, they didn't care to merge it. but I appreciate you for your troubleshooting and the fix.
I think just a rerun would be required to pass the build, otherwise the PR is already approved:
https://github.com/TryGhost/Ghost/actions/runs/6057836235/job/16821076886?pr=17938 https://github.com/TryGhost/Ghost/actions/runs/6057836235/job/16821076886?pr=17938
https://user-images.githubusercontent.com/4962633/275414297-2abc6c33-ed0c-442d-9e81-9cdbd0e405cd.png
—
Reply to this email directly, view it on GitHub https://github.com/TryGhost/Ghost/issues/17514#issuecomment-1763832578, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWHNLAX5ONSU2FFFUG6MRN3X7TJZDANCNFSM6AAAAAA2ZM7QIU.
You are receiving this because you modified the open/close state.
Maybe @daniellockyer can help
You did't fix the error: Make this site private - not working not working in Chrome and Opera
SOLUTION - it told you in August 2023 and I have made the PR!
versions/5.82.2/core/frontend/apps/private-blogging/lib/middleware.js
`const privateBlogging = {
....
return session({
name: 'ghost-private',
maxAge: constants.ONE_MONTH_MS,
signed: false,
sameSite: urlUtils.isSSL(config.get('url')) ? 'none' : 'lax', <------------ insert this
secure: urlUtils.isSSL(config.get('url')) <------------------------------ insert this
//sameSite: 'none' <------------------------------------------ row 58: remove
})(req, res, next);
},`
The pull request still not merged in version 5.82.2:
Fixed private mode cookie for local development #17938
What makes you say the PR wasn't merged? The commit shows that it's been in releases starting from 5.70.0.
I've downloaded the code:
versions/5.82.2/core/frontend/apps/private-blogging/lib/middleware.js
But the code is still the same:
` return session({
name: 'ghost-private',
maxAge: constants.ONE_MONTH_MS,
signed: false,
sameSite: 'none' <------------------------------------------ why this?
})(req, res, next);
},``
The code being the same does not mean your PR was not merged. In this case it looks like this change ended up possibly breaking something else so it was reverted:
https://github.com/TryGhost/Ghost/pull/19298
The code being the same does not mean your PR was not merged. In this case it looks like this change ended up possibly breaking something else so it was reverted:
#19298
OK so that means it is still a problem. I am running 5.79.6 (released Feb 26) and cannot make the site private because of this bug. What's the ETA on solving this?
The code being the same does not mean your PR was not merged. In this case it looks like this change ended up possibly breaking something else so it was reverted:
#19298
OK so that means it is still a problem. I am running 5.79.6 (released Feb 26) and cannot make the site private because of this bug. What's the ETA on solving this?
My solution:
I change always the code by myself, when I install an update. You have to change only 2 lines.
The first time I reported the solution in August 2023.
Maybe they will fix the bug sometimes ...
Hey guys, any update on this one? I couldn't access the links behind why the commit was reverted. So not sure on the details or complexity of the bug, is there any progress towards figuring it out? Thanks, and I'm a huge Ghost fan 😊
Downloaded version 5.89.1 - this bug still not fixed
You have to change only 2 lines in your code.
The first time I reported the solution in August 2023.
Why this is impossible?
@daniellockyer this is disappointing that this is still an issue exp with docker involved.
I can confirm that I have no access via chromium, chrome and edge. Works with firefox. Will tell that to my customers not.
Hey, I'm sorry that it wasn't made clear when the related PR was reverted. Unfortunately the fix broke the theme preview in Ghost admin for private sites, when the admin and site URLs are configured differently (the recommended configuration).
We are clearly missing some test coverage there, as the PR looked good to merge.
In the meantime, whilst trying to understand this issue I wasn't able to reproduce it in Arc, Chrome or Chromium.
There's something really janky going on here, because whilst there are clearly a couple of people here on this issue experiencing the problem, there's not a lot of wider noise despite private sites being used widely very successfully.
Meanwhile, when we merged the PR and broke the theme preview for private sites, we heard about it instantly from many people.
So there has to be a caveat that's not being covered here in the reproduction steps. I have a feeling that is something to do with SSL, which shouldn't be impacting production sites.
I'm going to close this bug as it stands. If anyone has the detailed reproduction case, feel free to open a new issue and we can work through what cases should and shouldn't work & making sure that fixing this issue doesn't cause a more widespread issue for private site users.
I wasn't able to reproduce it in Arc, Chrome or Chromium.
I've updated Chrome and the bug disappeared.
It was a bug in Chrome with the error-message:
OK, many thanks
|
2025-04-01T06:37:38.672269
| 2018-07-30T09:22:11
|
345677414
|
{
"authors": [
"c0derabbit",
"kevinansfield"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2934",
"repo": "TryGhost/Ghost",
"url": "https://github.com/TryGhost/Ghost/issues/9760"
}
|
gharchive/issue
|
Ajax call from external website - Response 200 but no posts
Issue Summary
I’d like to fetch blog posts into my website client-side.
When I fetch posts, I get this response, and no response body:
ok: true
redirected: false
status: 200
statusText: "OK"
type: "cors"
url: "http://my_website/ghost/api/v0.1/posts/?limit=3&client_id=ghost-frontend&client_secret=MY_SECRET
To Reproduce
Follow this guide: https://api.ghost.org/docs/ajax-calls-from-an-external-website
Try to fetch posts from an external website.
Expected behaviour is to get a list of posts. When I curl it works fine.
Technical details:
Ghost Version: 1.25.1
Node Version: 6.14.3
Browser/OS: Chrome 67.0.3396.99 on the client, backend running on Ubuntu 16.04.5 (aws ec2 t2.micro)
Database: mysql
Thanks for looking into it :)
Hey @c0derabbit 👋 We ask that you please do not use GitHub for help or support, the default issue template pointed you to our forum for this type of question 😄 We use GitHub solely for bug-tracking and on-going feature development so we try to keep it noise free.
Many questions can be answered by reviewing our docs for self-hosters, our theme API, or our public API. If you can't find an answer then our forum is a great place to get community support, plus it helps create a central location for searching problems/solutions.
FYI: Many projects have their own support guidelines and GitHub will highlight them for you, or the project owners will use issue templates to point you in the right direction, please read them before opening issues
Hi @kevinansfield, I started there but no response, also it seems to me more like a bug than a question, as I did follow the instructions, and I suspect it's either not working as it should, or there's a missing step in the docs.
|
2025-04-01T06:37:38.738097
| 2018-03-08T11:16:29
|
303444258
|
{
"authors": [
"daniellockyer",
"scidran"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2935",
"repo": "TryGhost/node-sqlite3",
"url": "https://github.com/TryGhost/node-sqlite3/issues/949"
}
|
gharchive/issue
|
Sqlite3 doen't install into dory-node android app
Hy.
I'm testing a proget with dory node (app into google play from tempage.io).
I've installed into a Samsung Tab A10.1 with android 7.0 and kernel 3.18.14, the gekko project from git hub.
Then i used:
npm install --only=production
But there was error to install sqlite3.
I tryed to install itmanually with
npm install sqlite3
But i have error
I attach this log
npm-debug (npm bugs sqlite3).log
npm-debug (npm install --build-from-source).log
npm-debug (npm install sqlite3 --loglevel=info).log
Before try to root my Tablet, i would like to know if there is an other solution...
Thanks
I'm trying to find different solution for the problem... I tried also to root my tablet but I had issue also with this procedure because my tablet is unsupported yet.
I read carefully the wiki but I'm not able to understand how to do it. I'm a very newbie with github and npm so I can't understand some things.
I have a package to clone from github (gekko from askmike);
Before putting into terminal npm install --only=production after the cloning, I have understand that I must:
download the package of sqlite3 (I downloaded the zip file from github: node-sqlite3-master.zip);
then, where should I put the sqlite3 unzipped files? Into what directory of the cloned package (there isn't a node-modules directory after the cloning process... the directory is created after the install process)?
After this, I have to open the file node.gyp but info the cloned directory, there isn't any *.gyp file...
only after the install process a node-pre-gyp file exist and isn't a node-gyp file like the wiki.
So I'm very confused!
Would you be able to try on the latest version v5.0.3? 🙂
|
2025-04-01T06:37:38.744587
| 2023-08-04T08:17:07
|
1836288865
|
{
"authors": [
"Tsingularity",
"ZY123-GOOD"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2936",
"repo": "Tsingularity/dift",
"url": "https://github.com/Tsingularity/dift/issues/6"
}
|
gharchive/issue
|
How to choose features from OpenCLIP?
Hello! Thanks for your great work.
I don't know how to use OpenCLIP to find correspondences. Could you please share these codes?
Thanks again.
Hi, thanks for your interest in our work!
The original OpenCLIP codebase didn't support input image size larger than the training resolution (at least it was still the case when we wrote our paper), so we follow the common practice and manually interpolate position encoding to support larger input resolution. This pull request could be very helpful as a reference.
Feel free to let us know if you have more questions.
@ZY123-GOOD I attached a cleaned-up version of our implementation here. Hope this helps!
|
2025-04-01T06:37:38.754745
| 2020-11-27T05:01:49
|
751982563
|
{
"authors": [
"RheingoldRiver",
"turtleworks"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2937",
"repo": "TsubakiBotPad/pad-cogs",
"url": "https://github.com/TsubakiBotPad/pad-cogs/issues/272"
}
|
gharchive/issue
|
transforminfo command
Shows an overview of a card including the awakenings of both base & transformed version, as well as the cooldown it takes to transform
Overview tab
Description
Show types of transformed version
Show both pre-transform and transform (bc we wanna count total SB, SBR, etc)
Show latents but don't bother showing caption that mentions pre-xform cos we know that
Stats etc
Leave as-is for now
Skills
Active skill (2cd) (Base: 30 -> 30)
Leave description as-is for now
Leave leader skill as-is for now
Additional tabs
Overview tab of base
Overview tab of transformed card
(Similar to how ^ls works)
Now I see why you've asked me to comment: https://github.com/isaacs/github/issues/100
Now I see why you've asked me to comment: https://github.com/isaacs/github/issues/100
|
2025-04-01T06:37:38.764707
| 2024-05-07T02:52:05
|
2282163518
|
{
"authors": [
"joobus"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2938",
"repo": "Tsuzat/NeoSolarized.nvim",
"url": "https://github.com/Tsuzat/NeoSolarized.nvim/issues/21"
}
|
gharchive/issue
|
How do I change the background color of a highlighted variable?
In the screenshot below, what is the name of the highlight group which is making "lunarvim/colorschemes" appear with a yellow background? I would like to change the color.
Figured it out: CurrentWord. Add this to your config to change the color:
-- Add specific highlight groups
on_highlights = function(highlights, colors)
highlights.CurrentWord.bg = colors.blue
end,
|
2025-04-01T06:37:38.772845
| 2016-04-12T15:01:48
|
147779523
|
{
"authors": [
"BoBeR182",
"Tudmotu"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2940",
"repo": "Tudmotu/gnome-shell-extension-bettervolume",
"url": "https://github.com/Tudmotu/gnome-shell-extension-bettervolume/issues/5"
}
|
gharchive/issue
|
Update for Gnome 3.20
Please update the extension to be compatible with Gnome version 3.20
Thanks :)
Uploaded a new version.
|
2025-04-01T06:37:38.785817
| 2023-11-08T10:21:44
|
1983262428
|
{
"authors": [
"clizbe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2941",
"repo": "TulipaEnergy/TulipaEnergyModel.jl",
"url": "https://github.com/TulipaEnergy/TulipaEnergyModel.jl/issues/237"
}
|
gharchive/issue
|
Define the pipeline & decide ecosystem
Description
What do we need from Toolbox/ESDL and what's our pipeline going to be?
[x] What are all the pieces of the pipeline?
[ ] How much of the pipeline can SpineToolbox cover?
[ ] How much of the pipeline can ESDL/EDR cover?
[ ] What is left uncovered?
[ ] Do we need to change the input of the model to "match" SpineToolbox?
Related Issues
Blocking #94, #106, #105, #118, #88, #89, #115, #36
The more I look at this, the more I think we should use SpineToolbox to integrate everything and the EDR as just one of several data sources. And maybe also be able to put results back into ESDL to use the MapEditor etc for analysis (depending on its capability).
What I'm a bit concerned about is the whole system still working in 5 years. Seems a bit complex.
Assigned myself although this is a group effort.
WHAT WE WANT
Build the network once (in a while)
Use draft networks to build new networks
Sufficient flexibility for ad-hoc code for experimentation
Definition of temporal stuff
Definition of solver specifications
Be able to mix data sources (ESDL + ENTSO-E for example)
|
2025-04-01T06:37:38.851811
| 2016-01-25T22:20:03
|
128656772
|
{
"authors": [
"ArthurHoaro",
"ksamuel",
"rugk",
"sametmax",
"wankbank"
],
"license": "WTFPL",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2942",
"repo": "Tygs/0bin",
"url": "https://github.com/Tygs/0bin/issues/101"
}
|
gharchive/issue
|
zerobin.py runserver: Unknown option '--host'
Using the command:
python zerobin.py --host <IP_ADDRESS> --port 80 --compressed-static
returns the error:
zerobin.py runserver: Unknown option '--host'
Changing it to:
python zerobin.py host <IP_ADDRESS> port 8001 compressed-static
returns:
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 3099, in run
server = server(host=host, port=port, **kargs)
File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 2723, in __init__
self.port = int(port)
ValueError: invalid literal for int() with base 10: '<IP_ADDRESS>'
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 3099, in run
server = server(host=host, port=port, **kargs)
File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 2723, in __init__
...
Non-stop...
Either the docs are incorrect, there's a bug, or I'm just stupid?
I have the same issue. It looks like the way 0bin handles parameters changed.
I've found this digging a bit in the code:
def runserver(host='', port='', debug=None, user='', group='',
settings_file='', compressed_static=None,
version=False, paste_id_length=None, server="cherrypy"):
So I managed to run 0bin on a custom port with:
python zerobin.py <IP_ADDRESS> 8006
But it doesn't really make any sense:
How am I supposed to enable compressed-static without setting previous parameters?
I have to put dummy parameters to set the settings files?
I'm probably missing something, but yes, the docs are outdated.
All of this happen because I merged some PR and trusted the content instead of reading it all. I'm guilty of lazyness. I will have to roll back everything, and fix all the bugs one by one.
I'm sorry for the mess, especially since I'm so slow at fixing it.
Any news on this issue?
BTW I also get issues with the command line suggested by @ArthurHoaro:
$ python zerobin.py --host <IP_ADDRESS> --port 80 --compressed-static
Traceback (most recent call last):
File "zerobin.py", line 4, in <module>
from zerobin.cmd import main
File "/home/zerobin-python/0bin/zerobin/cmd.py", line 12, in <module>
from sigtools.modifiers import annotate, autokwoargs
ImportError: No module named sigtools.modifiers
$ python zerobin.py <IP_ADDRESS> 8006
Traceback (most recent call last):
File "zerobin.py", line 4, in <module>
from zerobin.cmd import main
File "/home/zerobin-python/0bin/zerobin/cmd.py", line 12, in <module>
from sigtools.modifiers import annotate, autokwoargs
ImportError: No module named sigtools.modifiers
Can/Do I need to install sigtools.modifiers somehow?
This has been fixed in the V2 branch. It will be merged in master and pushed to pypi soon.
|
2025-04-01T06:37:38.868609
| 2019-11-12T03:26:19
|
521305936
|
{
"authors": [
"Perryvw",
"Sanjo",
"ark120202"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2943",
"repo": "TypeScriptToLua/TypeScriptToLua",
"url": "https://github.com/TypeScriptToLua/TypeScriptToLua/issues/746"
}
|
gharchive/issue
|
Transpiling Math.atan2 to math.atan2
Currently Math.atan2(y, x) is transpiled to math.atan(y / x).
math.atan2 is available in Lua 5.
math.atan2 (y, x)
Returns the arc tangent of y/x (in radians), but uses the signs of both parameters to find the quadrant of the result. (It also handles correctly the case of x being zero.)
Therefore Math.atan2(y, x) can be transpiled to math.atan2(y, x) for the extra functionality of quadrant checking and handling the case of x being zero.
It is a little more complicated: Lua 5.1, 5.2 and I'm assuming lua JIT (should be checked) do support math.atan2. Lua 5.3 however does not have math.atan2, but instead takes an optional second argument to math.atan, which allows it to function as atan2 https://www.lua.org/manual/5.3/manual.html#pdf-math.atan
The math.atan2 function is labelled as deprecated but still available in Lua 5.3. See https://www.lua.org/manual/5.3/manual.html#8.2 and I have tested if the math.atan2 function exists in Lua 5.3.5. So until the function is removed in a future version of Lua, it could be just transpiled directly to math.atan2. I have also verified that math.atan2 is available in LuaJIT 2.1.0-beta3.
I have tested if the math.atan2 function exists in Lua 5.3.5
Are you sure about that? I don't have it in a standard 5.3.5 build, maybe you had some compatibility flags enabled?
|
2025-04-01T06:37:38.879747
| 2020-01-09T10:48:21
|
547400000
|
{
"authors": [
"Gerrit0",
"jeremyrea",
"kobezzza",
"nknapp",
"socsieng"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2944",
"repo": "TypeStrong/typedoc",
"url": "https://github.com/TypeStrong/typedoc/issues/1159"
}
|
gharchive/issue
|
TypeDoc doesn't render comments with Handlebars 4.6.0
[x] I have checked issues with bug label and found no duplicates
Expected Behavior
Comments should be rendered when generating documents using globally installed typedoc. (i.e. npm install typedoc --global)
Note that it works as expected when installed locally.
# doesn't work - documentation rendered without comments
typedoc --out docs
# works - documentation rendered with comments
node_modules/.bin/typedoc --out docs
Expected:
Actual Behavior
Comments aren't included when run using typedoc global install.
The comment section is missing from the html output:
<section class="tsd-panel tsd-comment">
<div class="tsd-comment tsd-typography">
<div class="lead">
<p>Base class for animals</p>
</div>
</div>
</section>
Steps to reproduce the bug
Command:
npm install typedoc --global
git clone https://github.com/socsieng/typedoc-plugin-typescript-declaration.git
cd typedoc-plugin-typescript-declaration/example
typedoc --out docs
open docs/classes/_index_.example.animal.html
Environment
Typedoc version: 0.15.6
Node.js version: 8.16.0
npm version: 6.4.1
nvm version: 0.34.0
OS: macOS Catalina 10.15.1 (19B88)
I'm experiencing a similar issue, although it's the local installation that's not rendering the comments in the HTML output.
When I add the --json flag, I can see that my tags and shortText are picked-up in the object's comment property.
+1
This is.... really weird. I can confirm the global/local issue, no idea what's causing it yet. Looking into it.
@jeremyrea could you provide a repo with a repro for the issue when run locally?
It seems like the global install has been broken for a long time<EMAIL_ADDRESS>also has this issue.
@Gerrit0 https://github.com/jeremyrea/typedoc-comment-repro
Looks like Handlebars is the cause of this break (global + local I'm guessing, I bet the local install that works has a lower version of handlebars pinned in package-lock.json) - https://github.com/wycats/handlebars.js/pull/1633...
I'm not exactly sure how we should go about fixing this... listing out all of the prototype methods that we expect a template (as suggested in the handlebars PR) to be able to access isn't feasible and is very likely to break in the future whenever a new method is added.
For now, I'll pin handlebars to a lower version and release a patch with that change.
Fixed in v0.15.7, thanks for the report @socsieng + @jeremyrea!
Leaving this open to track finding a better solution. I don't want to be stuck on an old version of handlebars forever.
Thanks @Gerrit0, can confirm that it works for me.
Handlebars 4.7.0 has been release with options to disable prototype restrictions:
https://handlebarsjs.com/api-reference/runtime-options.html#options-to-control-prototype-access
Thanks @nknapp!
I'll release 0.15.8 with a handlebars version bump to 4.7.0 later today :)
v0.15.8 is released
|
2025-04-01T06:37:38.882644
| 2022-06-01T08:40:25
|
1255373926
|
{
"authors": [
"Gerrit0",
"Kivylius"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2945",
"repo": "TypeStrong/typedoc",
"url": "https://github.com/TypeStrong/typedoc/issues/1947"
}
|
gharchive/issue
|
Use typedoc as remote type
Search terms
typedoc, typescript, remote, micro frontends
Question
I love typedoc and its generated docs, but is it possible to use the output of typedoc or even write some plugin to be able to use typedoc output as remote type (comment or import) e.g.
/** typedoc:url website.com/typedoc/MyComponentType **/
const MyComponent = import('website.com/components/MyComponent');
- or -
import MyComponentType from `website.com/typedoc/MyComponentType`;
const MyComponent: MyComponentType = import('website.com/components/MyComponent');
This pattern would allow remote types, when using patterns like micro frontend and there no direct importing of the different repos. This could also aid in the generation of better docs, where the type is remote, we could say something like:
import MyComponentType from `website.com/typedoc/MyComponentType`;
type Something {
MyComponent: MyComponentType
}
Let me know if this is possible, if not maybe point me in the right direction.
You're probably after renderer.addUnknownSymbolResolver - https://github.com/TypeStrong/typedoc/blob/master/internal-docs/third-party-symbols.md
|
2025-04-01T06:37:38.901348
| 2021-07-05T17:39:27
|
937273081
|
{
"authors": [
"EremitaDelle6Vie",
"KavyaKinjalk",
"M123-dev",
"berkayyildi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2946",
"repo": "Tyrrrz/YoutubeDownloader",
"url": "https://github.com/Tyrrrz/YoutubeDownloader/issues/222"
}
|
gharchive/issue
|
MP3 downloaded music missing background image
Version
1.8.2
Details
When i download a song in MP3 version I don't get the image when I play it with VLC Media Player, there is only a black screen instead of the background (which is the thumbnail of the video I guess)
Is that bug or a problem with my media player or you intended to remove this feature in this version?
Steps to reproduce
Download any video in MP3 version
Open the file with a media player
This only extracts audio from the video from YouTube, so there won't be any "tags" for the audio file, so this is not a bug, I suppose, as you are getting the audio only
@KavyaKinjalk yeah but with the previous version I could do it, so I don't know if they removed this feature
@EremitaDelle6Vie, I have the suffer from issue, I tried around with some older versions but it did not work there either, maybe its caused by some changes on the serverside.
Same problem ++ Please add thumbnail images to mp3 files.
I can add on my previous comments.
In the Audi entertainment system the thumbnails were shown, but not in Windows.
But I used a older version in the process of getting it to work again.
|
2025-04-01T06:37:38.952143
| 2021-07-28T15:48:40
|
954999569
|
{
"authors": [
"astrochun",
"yhan818"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2948",
"repo": "UAL-RE/ReQUIAM",
"url": "https://github.com/UAL-RE/ReQUIAM/pull/161"
}
|
gharchive/pull-request
|
Merge develop into master - Include Grouper 2.5 migration
This is the primary PR now for testing Grouper 2.5 and implementation.
Note: That the original feature was so outdated that required a git merge --no-ff master.
Before this is merged, here are a few things to do:
[ ] Bump the version from v1.0.0 to v1.0.1
[ ] setup.py
[ ] requiam/__init__.py
[ ] docs/source/conf.py
[ ] Update CHANGELOG.md
Note that we have two config files for v2.2 and v2.5 deployment. It's likely that when Grouper 2.5 goes public, that the grouper25.iam.arizona.edu will become grouper.iam.arizona.edu
Closes #116
To test this PR, we do:
./scripts/script_run --config config/figshare_grouper25.ini --persistent_path /mnt/block3_sfo2/ --ldap_password $eds_pass --grouper_password $eds_pass --portal --quota --sync
A dry run tested 99% of the cases except it does not update the quota etc. So after the grouper 2.5 upgrade, partial failures exists for automatic run.
A manual run was carried out. A dry run was successful, but a "--sync" run failed in timeout. Updated "timeout =100" (originally timeout =60 in figshare.ini) solved this issue. The first batch of 100 took 73s. (that is the reason to timeout =60 failed).
17:41:48 - INFO: batch size = 100, batch timeout = 100 seconds, batch delay = 0 seconds
17:41:48 - INFO: processing drops:
17:41:48 - INFO: processing adds:
17:43:02 - INFO: added batch 1, 100 entries, 73.567426 seconds
17:43:12 - INFO: added batch 2, 100 entries, 10.436354 seconds
17:43:22 - INFO: added batch 3, 100 entries, 10.194059 seconds
17:43:32 - INFO: added batch 4, 100 entries, 9.949762 seconds
17:43:42 - INFO: added batch 5, 100 entries, 10.029692 seconds
17:43:45 - INFO: added batch 6, 33 entries, 2.717927 seconds
17:43:45 - INFO: QUOTA : Total time: 0 hours 7 minutes 17.11 seconds
17:43:45 - INFO: Total time: 0 hours 15 minutes 39.70 seconds
17:43:45 - INFO: ******************************
17:43:45 - INFO: SUMMARY DATA
num_EDS
num_Grouper
adds
drops
total
fine_arts
258
258
0
0
0
performing_arts
426
426
0
0
0
architecture
178
178
0
0
0
arts_design
41
41
0
0
0
business_econ
424
424
0
0
0
management
272
272
0
0
0
economics
110
110
0
0
0
education
1301
1301
0
0
0
english
296
296
0
0
0
lang_culture
392
392
0
0
0
humanities
271
271
0
0
0
law
402
402
0
0
0
nursing
369
369
0
0
0
med_health
1949
1949
0
0
0
clinical
857
857
0
0
0
ped_reprod
197
197
0
0
0
neurology
104
104
0
0
0
oncology
280
280
0
0
0
pharmacology
723
723
0
0
0
physiology
134
134
0
0
0
public_health
492
492
0
0
0
astro
646
646
0
0
0
cognitive_sci
360
360
0
0
0
life_sci
1846
1846
0
0
0
sci_math
1898
1898
0
0
0
earth_sci
1047
1047
0
0
0
physics
200
200
0
0
0
lpl
310
310
0
0
0
agriculture
143
143
0
0
0
anthropology
207
207
0
0
0
social_sci
870
870
0
0
0
cultural_studies
169
169
0
0
0
history
72
72
0
0
0
journalism
245
245
0
0
0
engineering
1268
1268
0
0
0
technology
144
144
0
0
0
libraries
277
277
0
0
0
536870912
15097
15097
0
0
0
2147483648
24393
23860
533
0
533
17:43:45 - INFO: ******************************
17:43:45 - INFO: Exit 0
|
2025-04-01T06:37:38.966050
| 2020-11-19T07:27:03
|
746322230
|
{
"authors": [
"chuangw46",
"huan-ds"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:2949",
"repo": "UBC-MDS/Abalone_Age_Prediction",
"url": "https://github.com/UBC-MDS/Abalone_Age_Prediction/pull/1"
}
|
gharchive/pull-request
|
create code-of-conduct and contribution file
I also created a .gitignore file which ignores unnecessary files.
Reviewed by Huan
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.