added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:37:36.246888
2024-02-11T21:39:00
2129181246
{ "authors": [ "AbhijeetKrishnan", "theneosloth" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2824", "repo": "TLNBS2405/heihachi", "url": "https://github.com/TLNBS2405/heihachi/pull/5" }
gharchive/pull-request
Fix typing annotation I run an instance of the bot in the containerized 3.9 environment, after rebasing with the master fork started seeing failures like TypeError: unsupported operand type(s) for |: 'type' and 'NoneType' Seems to do with an invalid union type annotation. I updated it to use the standard Optional types instead | was added as a union type shorthand in python 3.10 so feel free to close if you don't intend to support earlier versions discord.py is the main dependency and it requires Python 3.8+, so I wonder if it would be a better idea to support at least versions 3.8+.
2025-04-01T06:37:36.248685
2023-02-22T21:14:34
1595840014
{ "authors": [ "ThiloAschebrock" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2825", "repo": "TNG/next-generation-scrum-poker", "url": "https://github.com/TNG/next-generation-scrum-poker/pull/106" }
gharchive/pull-request
Add button to refresh session id I change the focus state to use outline similar to our links instead to separate from the hover state as suggested.
2025-04-01T06:37:36.298286
2015-03-12T03:14:35
60764859
{ "authors": [ "enzoz", "rodrigomaia17" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2826", "repo": "TWtablero/tablero", "url": "https://github.com/TWtablero/tablero/pull/183" }
gharchive/pull-request
Add vagrant to provide a faster/easier way to run application In order to fulfill #153 , I wrote a script to provision Tablero with ansible. To make this even easier, I create a config.json file that can be used to setup configs without environment variables. ###How to run fil the config.json file with your config. Install Vagrant and Ansible run vagrant up open the box with vagrant ssh run npm start now tablero will be available on the host machine on localhost:3000 @enzoz Can you check this, please? (: @rodrigomaia17 :+1: :smile:
2025-04-01T06:37:36.306133
2019-10-10T16:18:26
505375507
{ "authors": [ "featdd", "sypets" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2827", "repo": "TYPO3-Documentation/t3SphinxThemeRtd", "url": "https://github.com/TYPO3-Documentation/t3SphinxThemeRtd/issues/127" }
gharchive/issue
403 Error on docs.typo3.org for my documentation Hi, the link in the extension browser to my documentation results in a 403 (https://docs.typo3.org/typo3cms/extensions/dpn_glossary/stable/) There hasn't been any changes in the last releases, is this the right place for this issue? Greetings Daniel You can report problems with documentation rendering either on Slack (https://typo3.slack.com) in #typo3-documentation (see https://typo3.org/community/teams/documentation/#c9886) or in this issue: https://github.com/TYPO3-Documentation/T3DocTeam/issues/92 (this is also linked on the "Migrate Documentation" page under step 4: request redirects) This here is not the correct repository, but it's not a problem. I will forward your request. In any case, you should request a redirect if you have not done so already. I saw that your extension has already been rendered with the new mechanism: https://docs.typo3.org/p/featdd/dpn-glossary/3.0/en-us/ Actually, extensions.typo3.org should automatically update its information and use the new URL. This may take up to a day. Hi @sypets, thanks for all the information and the forwaring of my request. Greetings Daniel
2025-04-01T06:37:36.335896
2016-12-21T10:56:39
196903353
{ "authors": [ "TakahikoKawasaki", "phihochzwei" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2828", "repo": "TakahikoKawasaki/nv-websocket-client", "url": "https://github.com/TakahikoKawasaki/nv-websocket-client/issues/90" }
gharchive/issue
Cannot reconnect to Websocket Hi there, here I have something really strange for you. We are using you library to establish a connection between an Android Wear Device (V 1.5) and a WebSocket. The device is paired with a Samsung Galaxy S5 (latest Android). When our app first starts (Watch is connected via bluetooth to Mobile), everything works fine and as expected. If then the Bluetooth connection is cut (for whatever reason), the Watch switches to WIFI and also connects as intended. When you now reconnect the Watch to the Mobile via Bluetooth, the WebSocket connection does not get established. Only solution is, to reboot the Watch (which actually is no solution :) ) One "funny" thing to mention: When you get the Network-Type on the Watch while connected via Bluetooth, it returns the type "PROXY" (constant value is 16) which actually is NOWHERE documented. And when you try to get the DefaultProxy-Information vom the ConnectivityManager, it returns a null And here is the message thrown, when trying to connect com.neovisionaries.ws.client.WebSocketException: Failed to connect to 'XXXX:3000': Host is unresolved: XXXX at com.neovisionaries.ws.client.SocketConnector.doConnect(SocketConnector.java:119) at com.neovisionaries.ws.client.SocketConnector.connect(SocketConnector.java:81) at com.neovisionaries.ws.client.WebSocket.connect(WebSocket.java:2022) at com.neovisionaries.ws.client.ConnectThread.run(ConnectThread.java:37) Caused by: java.net.UnknownHostException: Host is unresolved: XXXX at java.net.Socket.connect(Socket.java:867) at com.neovisionaries.ws.client.SocketConnector.doConnect(SocketConnector.java:110) I would be really thankful for any help After purchasing an Acer and a LG Phone, I can confirm that this is not related to the known "Samsung cannot resolve URI"-Problems I'm sorry I don't understand your issue well enough, but it seems that the error message says XXXX is an invalid host name. Is it an actual error message, or did you replace an actual host name with XXXX before pasting the error message here? If "Host is unresolved" is the root cause, it's not a WebSocket issue but a DNS issue. Maybe the problem gets clearer when I explain the steps to reproduce: Wear-Device connected to Internet via Bluetooth over Smartphone -> Host can be resolved and WebSocket connects Turn off Bluetooth on Wear-Device Wear-Device connected via WiFi directly -> Host can be resolved and WebSockets connects Turn on Bluetooth on Wear-Device Wear-Device connected to Internet via Bluetooth over Smartphone (again) -> Host cannot be resolved and WebSocket cant connect. Never again. Until you manually reboot the Wear-Device. Then it all start´s at 1. Well, then, it sounds that the root cause exists in the network layer in Wear-Device. A WebSocket library cannot do anything for the problem...
2025-04-01T06:37:36.338265
2022-06-09T09:34:26
1265860452
{ "authors": [ "LittleUnicat", "Takak11" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2829", "repo": "Takak11/recruit", "url": "https://github.com/Takak11/recruit/issues/1" }
gharchive/issue
How does it work? How dos it work? Can you teach me? 😁🤣😍 turtle with babershop LittleUnicat turtle with babershop 💈 🐢 I've already hanked this project. I've already hanked this project. WTF
2025-04-01T06:37:36.339659
2020-03-11T21:33:10
579557811
{ "authors": [ "asajatovic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2830", "repo": "TakeLab/spacy-udpipe", "url": "https://github.com/TakeLab/spacy-udpipe/pull/11" }
gharchive/pull-request
Feature/Optional fine-grained POS 'Dirty' fix for issue #10 Latest update fixes #10 by adding UDPipeLanguage.tokenizer.fine_POS bool attribute and addresses #9 with a code example in README.md
2025-04-01T06:37:36.344332
2016-05-25T17:07:01
156802416
{ "authors": [ "KonsomeJona", "Takohi-Jona", "jackzampolin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2831", "repo": "Takohi-Jona/OctoMouse", "url": "https://github.com/Takohi-Jona/OctoMouse/issues/3" }
gharchive/issue
[feature-request] Document storage I see visualizations in the upcoming feature list but it is not here yet. Perhaps you can document the storage for event data and then people can help themselves! This would be great for non Objective-C programmers who want to hack and create their own visualizations. Hello, I'm not sure to understand your request, but as suggested in this comment: https://github.com/Takohi-Jona/OctoMouse/issues/2 Exporting the data in a CSV document would be good as any developer or non-developer could import these data into their own applications or a Excel sheet in order to create their own visualizations. I would love that! What I'm most interested in is historical data that I could graph. Thank you! I'm gonna work on it. I can't give to you an estimate right now as I'm working on my personal project during my spare time but I will keep you in touch. Pull Request for this feature: https://github.com/KonsomeJona/OctoMouse/pull/8
2025-04-01T06:37:36.390026
2023-06-05T13:10:35
1741754306
{ "authors": [ "edgar0011", "tannerlinsley" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2832", "repo": "TanStack/router", "url": "https://github.com/TanStack/router/issues/609" }
gharchive/issue
Link not working on first click Describe the bug Link, not working properly, on first click, only the second click will change url params can be seen in the examples: https://tanstack.com/router/v1/docs/examples/react/basic Your Example Website or App https://tanstack.com/router/v1/docs/examples/react/basic Steps to Reproduce the Bug or Issue go to : https://tanstack.com/router/v1/docs/examples/react/basic click on Home, Posts, or any post Expected behavior Expect Link to work on the frist click Screenshots or Videos No response Platform OS: [e.g. macOS, Windows, Linux] Browser: [e.g. Chrome, Safari, Firefox] Version: [e.g. 91.1] any Additional context No response Fixed in latest
2025-04-01T06:37:36.428212
2016-09-26T10:32:54
179192868
{ "authors": [ "TannerRogalsky", "bonez001" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2833", "repo": "TannerRogalsky/love.js", "url": "https://github.com/TannerRogalsky/love.js/issues/40" }
gharchive/issue
Player not creating the proper game.data and game.js file I am done publishining a game as a html 5 game using python C:\Users\pc\love.js\emscripten\tools\file_packager.py game.data --preload C:\Users\pc\Desktop\Example_Game_1\ExampleGame.exe@/ --js-output=game.js but for my second try, it doesn't output the correct file. What is worng? You should be packaging either your game folder or the .love file, not the .exe file. On Monday, 26 September 2016, bonez001<EMAIL_ADDRESS>wrote: I am done publishining a game as a html 5 game using python C:\Users\pc\love.js\emscripten\tools\file_packager.py game.data --preload C:\Users\pc\Desktop\Example_Game_1\ExampleGame.exe@/ --js-output=game.js but for my second try, it doesn't output the correct file. What is worng? — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/TannerRogalsky/love.js/issues/40, or mute the thread https://github.com/notifications/unsubscribe-auth/AAn-U0lQgbexFZWaNUcxArkyNtyijcunks5qt59WgaJpZM4KGXpK . Thanks for the reply. I tried to do it but it outputs files that hvae very little data 10-500 kb. When I put the game.data and game.js to the debug folder, it always outputs the sound and the screen of one previous game. I dunno what is wrong. Thank you. I type this one python C:\Users\pc\love.js\emscripten\tools\file_packager.py game.data --preload C:\Users\pc\game_cards@/ --js-output=game.js. I dunno what is wrong. Maybe it has something to do with the conf.lua of previous game? I will try it on other pc. That looks right to me. What is the structure of the game_cards directory? Hello. I made it again. I dunoo what is wrong last time.
2025-04-01T06:37:36.434234
2023-08-14T07:38:05
1849190590
{ "authors": [ "MenkOrd", "Post-Mortem", "TapiocaFox" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2835", "repo": "TapiocaFox/Daijishou", "url": "https://github.com/TapiocaFox/Daijishou/issues/509" }
gharchive/issue
Retroachievements doesn't update in the widget Hi, I can enter into my RA profile in the web, everything okay. But when I try to refresh my RA widget, it doesn't work. Could you provide more details? Thanks. Issue creator didn't provide any details after 2 weeks.
2025-04-01T06:37:36.474715
2024-05-19T15:12:41
2304656326
{ "authors": [ "gibman", "ug30242" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2836", "repo": "Tasshack/dreame-vacuum", "url": "https://github.com/Tasshack/dreame-vacuum/issues/552" }
gharchive/issue
One of c102cn's cleaning modes has disappeared. Describe the bug There are a total of 4 cleaning modes in c102cn, of which the mopping after sweeping mode has disappeared. To Reproduce When I select cleaning mode, I can't find the mopping after sweeping mode. (Mi home app has that mode) It seems to have probably disappeared since version 2.0.0b12 or 2.0.0b11. Expected behavior In cleaning mode, you should see four modes: sweeping, mopping, sweeping and mopping, and mopping after sweeping. Screenshots Additional Information (please complete the following information) Model name: dreame.vacuum.c102cn When this happens to me I can either do one of two things in order to temp. fix it. restart the dreame integration. start vacuum job on the dreame app. and then just force it to return to dock. This seems to fix the available cleaning modes in home assistant.
2025-04-01T06:37:36.568198
2022-10-08T03:56:57
1401811243
{ "authors": [ "Rahul-Palamarthi", "TaufeeqRiyaz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2837", "repo": "TaufeeqRiyaz/Spotify-Clone", "url": "https://github.com/TaufeeqRiyaz/Spotify-Clone/issues/41" }
gharchive/issue
Carousel Slider Want to add carousel slider to the "Good Morning section". Pls assign it to me.. Sure, I've assigned this to you
2025-04-01T06:37:36.585301
2023-03-30T22:15:44
1648397475
{ "authors": [ "abbaskhank", "arcticfly" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2838", "repo": "TaxyAI/browser-extension", "url": "https://github.com/TaxyAI/browser-extension/issues/15" }
gharchive/issue
Issue with back button/pdf docs in browser I have tried a couple of websites where I'm trying to download documents from, it does pretty good job at identifying the docs but then after opening a doc it fails to go back to download more. Starting page: https://louisville.edu/purchasing/bids Next page: this works as expected https://louisville.edu/purchasing/bids/RP-026-22 Next page: this is where it should download the document but instead it click on the doc which opens a pdf in the browser and then things fail https://louisville.edu/purchasing/bids/rp-026-22-document Error Invalid argument type: Expected a number for argument "elementId", but got ""B186503E38A0DCEF54EB2C9C31EE4177"". The user requests the following task: Download all open bids details that I need to submit my proposal You have already taken the following actions: I should click on each bid to access its details click(469) I should click on all the related documents to download the details click(476) Current time: 3/30/2023, 6:11:06 PM Current page contents: Response I should download the PDF file click("B186503E38A0DCEF54EB2C9C31EE4177") Action { "error": "Invalid argument type: Expected a number for argument "elementId", but got ""B186503E38A0DCEF54EB2C9C31EE4177""." } Very weird. Could you also post the "Prompt" section of the last action? i will try to recreate it but basically it just fails every time you click on a link that opens a pdf. Not fails per say bit it just doesn't know what to do after that, maybe it consider that pdf doc a new website?
2025-04-01T06:37:36.586860
2017-03-01T11:11:45
211049848
{ "authors": [ "Harmonicorps" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2839", "repo": "TaylanTatli/Moon", "url": "https://github.com/TaylanTatli/Moon/issues/104" }
gharchive/issue
Help eliminating the annoying flash between pages So every time I click a link on my site, the screen flashes whatever $bodycolor is currently set to. I want to eliminate this flash. I am aware of InstantClick.js which eliminates the flash, but it doesn't play well with the website. For some reason it makes the actual content disappear after clicking to new pages a few times. I also noticed it tries to load posts (posts button on home page) via http instead of https and as a result, the list of posts never loads. Any chance you could look into getting this awesome little javascript to work correctly? https://github.com/dieulot/instantclick/issues/151
2025-04-01T06:37:36.587810
2020-07-12T22:11:28
655486263
{ "authors": [ "Tbaut" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2840", "repo": "Tbaut/open-polkassembly", "url": "https://github.com/Tbaut/open-polkassembly/pull/6" }
gharchive/pull-request
Add initial DB creation Creates a DB and collection at first run (or if the expected env var aren't set) cc @niklabh merging to keep it running.
2025-04-01T06:37:36.836248
2024-04-12T20:05:03
2240818214
{ "authors": [ "SFM61319", "TechnikTobi", "Xuf3r" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2851", "repo": "TechnikTobi/little_exif", "url": "https://github.com/TechnikTobi/little_exif/issues/7" }
gharchive/issue
[Bug]: little_exif::metadata::Metadata::write_to_file is too slow and requires optimization As the title suggests, the little_exif::metadata::Metadata::write_to_file method is too slow and needs to be optimized. I am using little_exif as a dependency in a CLI app of mine. I love the API design and appreciate the good use of Rust's type system (especially for EXIF tags). However, there is a major drawback when it comes to the performance of the crate. Most (~99%) of the time in my core logic function (that involves writing a single EXIF tag to a JPEG image) is spent in little_exif::metadata::Metadata::write_to_file. Before this, my core logic function also reads the same file using std::fs::read and computes its hash, and yet that part of the logic seems to only take ~1% of the time. My program without little_exif::metadata::Metadata::write_to_file takes less than 100ms (even after including logging IO) for 10 JPEG images (~20MB each). But with little_exif::metadata::Metadata::write_to_file it takes over 4s with the same data under the same conditions. That is 40x the original execution time. And this was just a sample dataset. The program's expected input specification states the JPEG image count to be >=60,000, the size of each image being >=20MB. This means a task supposed to be completed in under 10 minutes will take over a day, which is unacceptable to say the least. A flame graph of the little_exif::metadata::Metadata::write_to_file part of my core logic function: When I asked for help on the Rust Community Discord server, someone mentioned that little_exif::metadata::Metadata::write_to_file was likely reading the same file again and again, 4 times in total, therefore increasing the time as well as the memory consumption. I haven't gone through little_exif's source code, so I thought I should let you know about this possible bug source. Yeah i was appalled by the execution time too. The issue lies in the clean_metadata() function which write_to_file() calls. It pounds the fs with syscalls byte by byte to find and remove the APP1 segment from JPEG. I've refactored the JPEG part into in-memory mutation but since then wrote my own parser for PNG and JPEG for all the markers. I can submit the pull request if the owner is fine with refactor of the clean_metadata() for only one format. @SFM61319 Thanks for your feedback - glad you like the API design! :) Regarding the performance issue: Speed was never one of the main objectives of little_exif. I’m aware that some functions are primitive and have a lot of room for improvement regarding performance. So far during development it was a trade-off between speed and being able to follow what is going on, how metadata is encoded, etc. This however does not mean that I am not open for suggestions/pull requests/etc. on how to improve the current situation! Speaking of pull requests: @Xuf3r I'd be happy to incorporate your modifications to metadata cleaning! @SFM61319 There is now a new version that includes the optimized version for metadata clearing from @Xuf3r. Would be great to hear from you whether this fixed the performance issue you experienced! Nice! There is a major performance boost in the new update. I appreciate the quick response from your side, and a satisfying one at that! Thank you so much! Here is the new flame chart:
2025-04-01T06:37:36.876717
2018-07-05T09:43:47
338500852
{ "authors": [ "gdgellatly", "pedrobaeza" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2852", "repo": "Tecnativa/docker-odoo-base", "url": "https://github.com/Tecnativa/docker-odoo-base/pull/152" }
gharchive/pull-request
Set initdb args for Odoo in scaffolding common.yaml to ensure data integrity and index performance. Address point 3 of #151 What about the LC_COLLATE thing? ? locale implicitly sets LC_COLLATE afaik as well as LC_TYPE
2025-04-01T06:37:36.899453
2018-09-05T13:12:07
357222814
{ "authors": [ "iTetsuyaMurakami", "ishidawataru" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2853", "repo": "Telecominfraproject/oopt-tai", "url": "https://github.com/Telecominfraproject/oopt-tai/pull/25" }
gharchive/pull-request
Add taish app Add taish app to provide the simple CLI interface to load an external TAI module, initialize optical modules via TAI, get the status of optical modules. @scott-emery I'd like to have this code in this repo. I think oopt-tai-implementations repo is a repo for libtai.so implementation. This app is not implementing TAI. @iTetsuyaMurakami Thanks!
2025-04-01T06:37:36.912598
2015-01-29T18:07:00
55934916
{ "authors": [ "SachaG", "fobo66" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2854", "repo": "TelescopeJS/Telescope", "url": "https://github.com/TelescopeJS/Telescope/pull/733" }
gharchive/pull-request
*Added russian translation (not full, because some functions are fixed i... ...n English); +Added semicolons in users.js' clog callbacks. Also, i find some unuseful functions in users.js. Can i delete it? We're switching to LingoHub to help manage translations. If you'd like to be added as a translator to the project, could you join the #i18n channel on Telescope's Slack?
2025-04-01T06:37:36.925121
2023-05-05T04:08:56
1696977011
{ "authors": [ "TechDufus", "chrisbenincasa", "eremem", "luispabon", "noahccoker" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2855", "repo": "Telmate/terraform-provider-proxmox", "url": "https://github.com/Telmate/terraform-provider-proxmox/issues/763" }
gharchive/issue
All disk properties config show -> null as needed change. Terraform v1.4.6 on linux_amd64 + provider registry.terraform.io/telmate/proxmox v2.9.11 Small context, I'm unable to upgrade to newer provider versions. (See https://github.com/Telmate/terraform-provider-proxmox/issues/762) When running a plan, even after applying, all VMs show every property needs to change -> null even if I set every value in the disk {} block, which isn't necessary. - disk { - backup = 0 -> null - cache = "none" -> null - file = "vm-112-cloudinit" -> null - format = "raw" -> null - iops = 0 -> null - iops_max = 0 -> null - iops_max_length = 0 -> null - iops_rd = 0 -> null - iops_rd_max = 0 -> null - iops_rd_max_length = 0 -> null - iops_wr = 0 -> null - iops_wr_max = 0 -> null - iops_wr_max_length = 0 -> null - iothread = 0 -> null - mbps = 0 -> null - mbps_rd = 0 -> null - mbps_rd_max = 0 -> null - mbps_wr = 0 -> null - mbps_wr_max = 0 -> null - media = "cdrom" -> null - replicate = 0 -> null - size = "4M" -> null - slot = 1 -> null - ssd = 0 -> null - storage = "lvm-thin-06" -> null - storage_type = "lvmthin" -> null - type = "scsi" -> null - volume = "lvm-thin-06:vm-112-cloudinit" -> null } Naturally, being flooded with all VMs needing changed for each plan gets in the way of seeing real pending state changes. Are you creating a VM via cloning? I'm hitting the same thing (and it seems like many others are too #655). I'm trying to figure out if this has to do with cloning a VM and not overriding the set parameters of the template? Correct.. I set up a VM that I use as a template (I don't actually make it a template) and I clone from it and use cloud-init. Yea that is a exactly what I'm doing as well. My guess is that the provider isn't ignoring all of the settings in the clone case Exactly the same problem when cloning from a template. Terraform v1.5.3 on windows_amd64 + provider registry.terraform.io/telmate/proxmox v2.9.14 Having the same issue too, hope this can get prioritized soon Ping. any updates on this? The workaround for me is to specify a disk stanza that matches the disk set up on the templated VM, like so: VM template: Matching disk config:
2025-04-01T06:37:37.109771
2024-12-01T14:13:03
2709231771
{ "authors": [ "Darth-Carrotpie", "grandmazboy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2856", "repo": "TencentARC/InstantMesh", "url": "https://github.com/TencentARC/InstantMesh/issues/178" }
gharchive/issue
Does anyone check if installation works regularly? I mean does it hurt to update the requirements.txt with exact versions.. since you wont know when a package is updated and if its compatible with other versions. Does anyone ever compile code to work with newer versions of Cuda, Python or VS? Seems like stable diffusion and other tools similar to this finish something then just give up on maintaining. C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.1\include\crt/host_config.h(153): fatal error C1189: #error: -- unsupported Microsoft Visual Studio version! Only the versions between 2017 and 2022 (inclusive) are supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk. texture.cu ninja: build stopped: subcommand failed. VS Build Tools 2022 17.12.2 / 14.42.34433 VS Professional 2022 14.29.30133 Desktop development with C++ / tried both MSVC v142/143 Cuda 12.1 Python 3.10 Windows 11 / Not using Conda Nvidia RTX 4080 (venv) PS E:\InstantMesh-main> python run.py configs/instant-mesh-large.yaml examples/bird.jpg --save_video --no_rembg Seed set to 42 Loading diffusion model ... Loading pipeline components...: 12%|██████▌ | 1/8 [00:00<00:00, 7.82it/s]The config attributes {'dropout': 0.0, 'reverse_transformer_layers_per_block': None} were passed to UNet2DConditionModel, but are not expected and will be ignored. Please verify your config.json configuration file. Loading pipeline components...: 100%|████████████████████████████████████████████████████| 8/8 [00:01<00:00, 6.42it/s] Loading custom white-background unet ... Loading reconstruction model ... Some weights of ViTModel were not initialized from the model checkpoint at facebook/dino-vitb16 and are newly initialized: ['encoder.layer.7.adaLN_modulation.1.bias', 'encoder.layer.6.adaLN_modulation.1.weight', 'encoder.layer.4.adaLN_modulation.1.bias', 'encoder.layer.0.adaLN_modulation.1.weight', 'encoder.layer.6.adaLN_modulation.1.bias', 'encoder.layer.1.adaLN_modulation.1.bias', 'encoder.layer.10.adaLN_modulation.1.bias', 'encoder.layer.2.adaLN_modulation.1.weight', 'encoder.layer.7.adaLN_modulation.1.weight', 'encoder.layer.9.adaLN_modulation.1.weight', 'encoder.layer.3.adaLN_modulation.1.weight', 'encoder.layer.2.adaLN_modulation.1.bias', 'encoder.layer.8.adaLN_modulation.1.weight', 'encoder.layer.11.adaLN_modulation.1.bias', 'encoder.layer.8.adaLN_modulation.1.bias', 'encoder.layer.5.adaLN_modulation.1.bias', 'encoder.layer.3.adaLN_modulation.1.bias', 'encoder.layer.5.adaLN_modulation.1.weight', 'encoder.layer.9.adaLN_modulation.1.bias', 'encoder.layer.10.adaLN_modulation.1.weight', 'encoder.layer.4.adaLN_modulation.1.weight', 'encoder.layer.1.adaLN_modulation.1.weight', 'encoder.layer.11.adaLN_modulation.1.weight', 'encoder.layer.0.adaLN_modulation.1.bias'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. Traceback (most recent call last): File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 2100, in _run_ninja_build subprocess.run( File "C:\Program Files\Python310\lib\subprocess.py", line 524, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "E:\InstantMesh-main\run.py", line 132, in model.init_flexicubes_geometry(device, fovy=30.0) File "E:\InstantMesh-main\src\models\lrm_mesh.py", line 79, in init_flexicubes_geometry renderer = NeuralRender(device, camera_model=camera) File "E:\InstantMesh-main\src\models\geometry\render\neural_render.py", line 73, in init self.ctx = dr.RasterizeCudaContext(device=device) File "E:\InstantMesh-main\venv\lib\site-packages\nvdiffrast\torch\ops.py", line 184, in init self.cpp_wrapper = _get_plugin().RasterizeCRStateWrapper(cuda_device_idx) File "E:\InstantMesh-main\venv\lib\site-packages\nvdiffrast\torch\ops.py", line 125, in _get_plugin torch.utils.cpp_extension.load(name=plugin_name, sources=source_paths, extra_cflags=common_opts+cc_opts, extra_cuda_cflags=common_opts+['-lineinfo'], extra_ldflags=ldflags, with_cuda=True, verbose=False) File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 1308, in load return _jit_compile( File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 1710, in _jit_compile _write_ninja_file_and_build_library( File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 1823, in _write_ninja_file_and_build_library _run_ninja_build( File "E:\InstantMesh-main\venv\lib\site-packages\torch\utils\cpp_extension.py", line 2116, in _run_ninja_build raise RuntimeError(message) from e I have uninstalled Python and moved to 3.10.6 and separately installed StableDiffusion and its working . I am getting a shyt ton of package dependancy conflicts and one that always pops up is huggingface_hub Check https://github.com/TencentARC/InstantMesh/issues/175#issuecomment-2495631712 If it's a problem in a Dockerfile, which should be stable in theory, it's gonna be even a bigger problem in regular installs... In the end I did run it via docker, but when generating a mesh it errored out without any message. Not sure where to go on after that :/ I tried on a different computer today, with just the bare minimum of 3.10, visual studio community 2022, build tools, cuda 12.1 etc and that is getting very similar errors to what I'm getting on this PC. I suspect mass package dependency version issues just don't know the right combo..
2025-04-01T06:37:37.135435
2017-05-26T08:55:55
231565552
{ "authors": [ "anapanadero", "kyleledbetter" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2857", "repo": "Teradata/covalent", "url": "https://github.com/Teradata/covalent/issues/629" }
gharchive/issue
Design for autocomplete with custom inputs or custom chips The placeholders and labels for autocomplete with custom inputs or custom chips doesn't follow the official Material Design spec guidelines. That spec if for the standalone autocomplete this is chips autocomplete so slightly different usage and we're using a non float label that's a feature of angular material, and we're also following the pattern of chips autocomplete from angular-material v1
2025-04-01T06:37:37.145967
2024-09-19T23:12:30
2537506747
{ "authors": [ "ACGaming", "Mysticpasta1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2859", "repo": "TerraFirmaCraft-The-Final-Frontier/FileDirector", "url": "https://github.com/TerraFirmaCraft-The-Final-Frontier/FileDirector/pull/30" }
gharchive/pull-request
For The Director on 1.20.1 all credit to juan While this is highly appreciated, we're going to revert the rebranding, StopModReposts integration since its unmaintained and other aspects we deem unfitting. Because of my other projects, this is going to take a while.
2025-04-01T06:37:37.149038
2016-09-15T15:25:57
177208224
{ "authors": [ "ljacomet", "mathieucarbou", "myronkscott" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2860", "repo": "Terracotta-OSS/terracotta-apis", "url": "https://github.com/Terracotta-OSS/terracotta-apis/issues/152" }
gharchive/issue
Being able to configure a @Builtin Service Currently, a @Builtin service cannot be configured through XML. This request is to be able to optionally configure a @Builtin service if an XML config is present. The main use case I see is if we need, one day, to make our service that was builtin optionally configurable because some options are added. So to keep backward compat', the service will need to support both @Builtin AND if provided, be configured by the XML config. notify: @chrisdennis @ChrisGreenaway @myronkscott @jd0-sag @ljacomet Might be something we may need indeed. But that's also something that can be added while being completely compatible IMO. So I do not see the point of adding it now unless really needed. Makes sense? Yes, for M&M, but I know @ChrisGreenaway might need it also. This has been implemented. If a configuration is in tc-config, that will be used instead of the builtin.
2025-04-01T06:37:37.170386
2018-12-05T18:19:11
387883795
{ "authors": [ "Bootjack", "kalimar", "mirandawang", "rudokemper" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2861", "repo": "Terrastories/terrastories", "url": "https://github.com/Terrastories/terrastories/pull/186" }
gharchive/pull-request
[176] Allow for more granular permissions based on user access control groups This adds a new concept in the Demographic model which links users and stories through an access control grouping. I'm open to suggestions for the name, but it seems like these will mostly be based on real-world social groupings, such as elders from a certain village or descendants of one individual. The term "demographics" felt like a decent fit. This is a different concept from both user roles and story permissions. The three complementary concepts we'd have after this PR are: Story permissions - an individual story is marked as either: :anonymous meaning public, or :user_only meaning any signed-in user may see it User roles - an individual user has a single role that's one of: user basic signed-in user, may view some restricted content editor same as user, and also may edit story and location content admin same as editor, and also may edit other users (just added as an option, not yet implemented/enforced) Demographics - a group of users and stories that defines restricted access levels using the following data (basically describing two new HABTM relations) name a human-readable plain-language name for the group users a list of users included in this group stories a list of stories this group has access to Punch List [x] Create the Demographic model [x] Add relationships between Demographic, User, and Story models [x] Customize the demographic dashboard in administrate [x] Filter stories based on Demographics [x] Remove the direct User-Story access control association [x] Replace the Story.permission_level field with Story.is_public [x] Add pundit policies so that editors: [x] may not modify User models [x] may not view Demographics they don't belong to [x] may not add stories to Demographics they don't belong to [x] can only see appropriate Users and Stories when modifying Demographics [x] may not add Demographics outside of their permissions to Stories they have access to [ ] Automatically add the user who creates a Demographic to that Demographic (otherwise only admins can access it) [x] add the current user to the demographic [ ] prevent duplicate records [x] Do not list Demographics under Stories that when an editor doesn't belong to them [ ] Prevent limited Demographics scope of editor from deleting existing Demographics when updating Stories For Discussion I think we can streamline these concepts a bit, but we probably do want some form of all three. Check out this branch, go to the new Demographics dashboard in the admin and see if creating demographic groups and associating users and stories seems logical. Here are some specific questions I have for the team. Story permissions is currently a text field at Story.permission_level but I think at this point it only needs to be a boolean indicating whether a story should be shown to the public, something like Story.is_public. Similarly, should we expect that the combination of user role plus demographics will always result in the correct editing permission for stories. In other words, if a user is an editor, should they have editing access to all stories their demographics allow them to see? Or should there be a distinction between view-only demographics and editor-specific demographics? Does the admin user role make sense? My expectation would be that members of the ACT who are setting things up (plus maybe a few tech-savvy locals) would be the only admins. Everyone else would go to that group to get elevated editing permissions or to modify demographics for access control. I also modified the models to allow for direct user-to-story access levels, but I'm curious if this is really needed or if it just adds extra complexity and confusion. The same could be accomplished by creating a demographic group with only one user. If that's a rarity it should be fine; if it's common the number of single-user demographics could become a problem. Note, there's no dashboard UI for setting up these direct user-story relationships yet. AWESOME!!! :100: :tada: Thanks for getting this started! Discussion points: I think updating that filed to a boolean is_public? makes total sense here, since we only really have that distinction now for Stories. Editing and viewing to me are distinct. Viewing stories is on a user level, but some users also have the ability to add/update/delete stories. That also means they have access to the /admin backend. There could be an "editor-only" demographic for stories which aren't ready to be published yet, or something. admin makes sense to me! In the future these would be the users that could add and update custom fields for the models too. Maybe @rudokemper can help us answer that question, is it more common that stories would be visible to certain groups or usually just individuals? Thanks for the feedback, @mirandawang! To clarify my second question, I really just mean should an editor be able to edit every story they can view? I think we're agreeing that they should, but I wanted to make sure. Brilliant stuff Jason!! 👍👍👍 I would agree here. Yeah, I mean, theoretically there could be a scenario where a community member gets to view stories X, Y, and Z, but only have permission to edit X. Let's say that they are a hunter and story A is a hunting story, whereas Y and Z are stories told by elders which they should be able to hear, but not modify. But I really can't imagine this playing out in reality, certainly not anywhere where we work for instance. Let's just keep it simple. Fully agreed here and also with @mirandawang's suggestion about adding and custom fields for the models in the future. My feeling is that the demographic groups (with one-user groups if necessary) are perfectly sufficient, and that having the additional direct-to-user setup could likely result in more confusion. Cool so, to sum up: Editors can edit or delete any story that they can view / they have created. Story permission levels will be based on Demographic groups. For the rare case of individual level permissions, we can model this as a one person group. @rudokemper @Bootjack @kalimar does that all sound right? @kalimar In my defense, jQuery was already a part of administrate 😜 Also, some general explanation: Those last few commits (including the jQuery stuff) are in support of honoring pundit policies in the multiple-select widget for Stories/Demographics/Users in the administrate forms. To do that I created a custom field HasManyScopedField that subclasses the normal Field::HasMany from administrate. That felt a little hacky because I had to copypasta two methods from the base class and tweak them. Even more hacky is that in order to pick up the styles and js from the base field I had to run their rails generator that copied all the assets for all fields and dumped them into our app. I'm sure a more clever dev could have come up with a cleaner way to do that. Eventually, though, I think the in-map editing will supplant this entire administrate interface so I'm not super worried. lol, nvm. @kalimar 's nerd snipe was successful You know me too well @bootjack. You know me too well :eyes: closing for now; thank you for this work!
2025-04-01T06:37:37.185546
2015-02-17T16:38:41
57947558
{ "authors": [ "TerryMooreII", "ThomasPe" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2862", "repo": "TerryMooreII/angular-azure-mobile-service", "url": "https://github.com/TerryMooreII/angular-azure-mobile-service/pull/19" }
gharchive/pull-request
Exposing .read() method In order to send parameters to the javascript backend that can be read through request.parameters the .read() method of the azure sdk is now exposed with angular promise wrapped around. See also http://stackoverflow.com/questions/17415257/passing-query-parameters-from-html-js-app-to-azure-server-script updated the parameter name, fixed some issues in the readme and added the new method documentation Nice, thanks. I will push this to bower and the cdn soon.
2025-04-01T06:37:37.189453
2021-03-27T21:41:33
842607673
{ "authors": [ "ritvij14" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2863", "repo": "TesseractCoding/NeoAlgo", "url": "https://github.com/TesseractCoding/NeoAlgo/issues/4296" }
gharchive/issue
Armstrong number using recursive approach using Dart 🚀 Feature Add code for Armstrong number using the recursive approach in dart. Have you read the Contributing Guidelines on Pull Requests? Yes /assign
2025-04-01T06:37:37.191455
2021-04-22T16:38:14
865164858
{ "authors": [ "hemant2705" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2864", "repo": "TesseractCoding/NeoAlgo", "url": "https://github.com/TesseractCoding/NeoAlgo/issues/6433" }
gharchive/issue
Floyd's Triangle 💥 Proposal (A clear and concise description of what the proposal is.) floyd's triangle program in kotlin 1 2 3 4 5 6 7 8 9 10 Have you read the Contributing Guidelines on Pull Requests? yes (Write your answer here.) /assign
2025-04-01T06:37:37.194420
2021-04-15T15:02:22
858973482
{ "authors": [ "Amit366" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2865", "repo": "TesseractCoding/NeoAlgo", "url": "https://github.com/TesseractCoding/NeoAlgo/pull/5913" }
gharchive/pull-request
Trapping rain water in python Have you read the Contributing Guidelines on Pull Requests? yes Description Added code for trapping rain water Checklist [ ] I've read the contribution guidelines. [ ] I've checked the issue list before deciding what to submit. [ ] I've edited the README.md and link to my code. Related Issues or Pull Requests Fixes: #5851 @HarshCasper please review @ankitaggarwal23 please review
2025-04-01T06:37:37.204167
2024-05-17T04:00:14
2301764741
{ "authors": [ "CharliePoole" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2866", "repo": "TestCentric/TestCentric.Cake.Recipe", "url": "https://github.com/TestCentric/TestCentric.Cake.Recipe/issues/75" }
gharchive/issue
Restore capability to create draft release This was temporarily blocked in issue #67. By upgrading GitReleaseManager to version 0.17, it can be restored. :tada: This issue has been resolved in version 1.2.1 :tada: The release is available on: GitHub Release NuGet Package
2025-04-01T06:37:37.249952
2015-05-08T01:18:22
74184428
{ "authors": [ "arteme", "davidlgj", "nicklasb", "subhendupsingh" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2872", "repo": "Textalk/angular-schema-form", "url": "https://github.com/Textalk/angular-schema-form/issues/383" }
gharchive/issue
Divide array in 3 columns Can we divide "type":"array" in 2 or 3 columns and have a tabarray with horizontal tabs? Please suggest a workaround. I don't think I understand, do you want the same array flowing over three columns? Columns can be achieved by using standard bootstrap classes, see http://schemaform.io/examples/bootstrap-example.html#/4fa8967ae5596fe8b0c0 tabarray type has support for horizontal tabs, just set tabType to "top", here are the docs https://github.com/Textalk/angular-schema-form/blob/development/docs/index.md#tabarray In my project I achieve columns by splitting the items array in half using a custom fieldset decorator like: <fieldset ng-disabled="form.readonly" class="schema-form-fieldset schema-form-fieldset-columns {{form.htmlClass}}"> <legend ng-show="showTitle()">{{ form.title }}</legend> <div ng-show="form.description" ng-bind-html="form.description"></div> <div class="column"> <div class="row"> <sf-decorator ng-repeat="item in form.items|arrayHalf" form="item"></sf-decorator> </div> </div> <div class="column"> <div class="row"> <sf-decorator ng-repeat="item in form.items|arrayHalf:1" form="item"></sf-decorator> </div> </div> </fieldset> The arrayHalf filter is really easy to implement... Please re-open if you do not feel the question has been answered sufficiently.
2025-04-01T06:37:37.251593
2023-01-24T04:29:00
1554299021
{ "authors": [ "pionxzh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2873", "repo": "TexteaInc/json-viewer", "url": "https://github.com/TexteaInc/json-viewer/pull/216" }
gharchive/pull-request
chore: bump version and delay dependabot related: #151 #184 This PR bumps the deps and tries to slow down dependabot to reduce the noise. We will try to adopt renovate bot next week. And I will manually update the dependencies temporarily. cc. @rtritto
2025-04-01T06:37:37.255604
2024-06-11T19:13:24
2347111128
{ "authors": [ "ketan96-m", "willmcgugan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2874", "repo": "Textualize/rich", "url": "https://github.com/Textualize/rich/issues/3379" }
gharchive/issue
[REQUEST] Rich progress bar changes color based on the number steps/files elapsed. Consider posting in https://github.com/textualize/rich/discussions for feedback before raising a feature request. Have you checked the issues for a similar suggestions? Yes How would you improve Rich? My application submits a bunch of transformation to a server and the rich bar gives the status of the transformation and eventual downloads. However, these transformation are prone to failures and I would like to change the color of my progress bar to indicate that some failure has occurred while other transformations continue. I would like to change the bar color to red because it stopped just before finishing. The update function does provide ways to make changes but I couldn't find a way to change the color of the bar using BarColumn() when passed as a keyword argument to the update method. e.g progress.update( progress_task, progress_bar_title, completed=self.current_status.files_completed, bar = BarColumn(complete_style="rgb(0,0,255)", style="rgb(255,0,0)") ) What problem does it solve for you? This will add more style to the progress bar when there are failures or if we can't reach 100%. You can override Progress.get_renderables() to display the progress however you like.
2025-04-01T06:37:37.264672
2021-06-04T15:17:33
911604909
{ "authors": [ "DrAkashic", "MrTubzy456", "Tfarcenim", "aaronhowser1", "epicyeeto", "j4rw15", "tankcr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2875", "repo": "Tfarcenim/AE2WirelessTerminalLibrary", "url": "https://github.com/Tfarcenim/AE2WirelessTerminalLibrary/issues/16" }
gharchive/issue
AE2 Wireless Terminals (ae2wtlib) encountered an error during the load_registries event phase AE2 Wireless Terminals (ae2wtlib) encountered an error during the load_registries event phase java.lang.RuntimeException: Cannot find class appeng/container/implementations/MEMonitorableContainer crash-2021-06-04_18.12.47-fml.txt very undescriptive report, but this looks like your ae2 version is too new Wouldn't that be more like, this mod is outdated? If it requires an older version of AE2 Came here with same issue btw, crash log here: https://pastebin.com/H3zFJNmK Bump. Running into the same issue now. Here is a link to my crash log as well: https://pastebin.com/HUKwVjKR Same issue, looks like this addon hasn't been updated since May same issue here Repeating an already known issue won't fix it any faster. I just updated my server and ran into this issue. In the crash report it lead me here so I've the crash report to pastebin and if that helps let me know. https://pastebin.com/KrVtgELd I've got to figure out how to get my server to run. If you need anything, ask. I have access to all of the files for the server and pack.
2025-04-01T06:37:37.270956
2021-11-06T04:19:35
1046391194
{ "authors": [ "Th3M4ttman", "scott91e1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2876", "repo": "Th3M4ttman/ezcolors", "url": "https://github.com/Th3M4ttman/ezcolors/issues/1" }
gharchive/issue
Advantages of ezcolors over the de facto color library? Hello, I am in the market for a python color manipulation library for a new project. Any chance of a paragraph or two of why ezcolors was created and how it compares to the other libraries out there? Many thanks, Scott. Sure thing. EzColors was created when I was trying to find a multi platform terminal colour solution that was easy to use but capable of a wide array of different colour functionalities like gradients, rainbows and various generateable colour palettes and I couldn't find anything that quite fit the bill of what I was looking for. So I started work on a few functions and classes to make ANSI escape coloured strings in the form of an ordinary str or a ColorStr class with various methods to fiddle with the colours etc I got it to where it was usable for making coloured text relatively easily with nothing but ColorStr("text", "foreground colour in RGB, integer or hex", "background colour") and sort of left it for a while but then later returned and added some extra functionality like simple coloured splash screens and separators and a coloured yes/no/whatever option you want prompt. And colored exceptions for easier reading The whole thing is somewhat of a mess and in need of cleanup but I use it in many of my projects and for quick prototyping. Overall though I'm not sure I'd recommend it other some alternatives I haven't really looked for too long for exactly what I needed. If you wanna take a look and suggest any improvements I'd be happy to hear. There's a whole suite of eztools I've got in the works EzFiles (contains a file class for much easier file manipulation) EzConfig (fully featured configs from 1 line) EzPack (dynamically generate, package and build python scripts, modules and packages) EzColors Ezcmd (decorator based system to very easily turn any script into a command line utility and a class to easily create a command prompt from a list of functions) EzTest (easy decorator based unit tests with coloured output and reports with a Test class and a Case class that can be extended easily to automatically generate and cache testcases) EzValid (a bunch of regedit powered varlidators) With all being available either separately or as 1 package called eztools with additional utilities such as 1 line simple input function output tkinter GUI, and Timers Sorry to just dump a huge wall of text. In short I'm an intermediate-advanced solo python dev I wouldn't use my EzColors for anything super important but please do mess about with it and see how you like it. I will update the documentation soon as some stuff has probably slightly changed since I did it Also to note I was high both when I wrote EzColors and when I wrote this comment. And both my comments and the entirety of EzColors development were written on a Samsung galaxy a6. So when I say it's multi platform it works everywhere. Mac, windows, Linux, Android Also only just realised how long ago you commented this haha. Sorry for such a slow reply
2025-04-01T06:37:37.274777
2023-11-06T00:23:56
1978049382
{ "authors": [ "ThaUnknown", "zoriya" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2877", "repo": "ThaUnknown/jassub", "url": "https://github.com/ThaUnknown/jassub/issues/33" }
gharchive/issue
Chrome android does not work anymore Subtitles simply don't appear on chrome android anymore (could not tell from which version, it was already the case in 1.7.9 but maybe before too). This can be reproduced from this page, for example: https://thaunknown.github.io/jassub/jassub/simple/index.html There aren't any errors, and debug messages in the console appears when the subtitle should be shown. It feels like everything is working as it should, but subtitles are not shown. can't reproduce Interesting, even with brave, it does not work for me. this is likely due to offscreen render, it's a chromium bug The issue for the sandbox was #32 (which was fixed) but my codesandbox does indeed show another issue (I have not faced in production yet) as I said, disabling offscreen render fixes this on android, I don't know why this happens, I assume it occurs when the bitmap given to skia is too big? closing as no reproduction is available
2025-04-01T06:37:37.297869
2023-06-22T11:43:32
1769507602
{ "authors": [ "frappeerpnext", "nikkothari22" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2879", "repo": "The-Commit-Company/frappe-js-sdk", "url": "https://github.com/The-Commit-Company/frappe-js-sdk/issues/31" }
gharchive/issue
Add server message to error object When we make request via db or call object, if error occur from backend frappe send friendly message to client via _server_messages key. I cannot extract it in .catch((error) { ... }). It seems that you did not include it in error object. Here is server message when I check in network response. Could you please help include this message in .catch error object? Thanks, @frappeerpnext - yes this is something that needs to be added. I'll try adding it in. Yes, I hope this will be added soon. These error information is very useful and we save a lot time to customize friendly message to end user. Thanks, Hi Guys, I have update and test it. Working perfect. Big Thanks
2025-04-01T06:37:37.304790
2023-06-21T06:04:12
1766815269
{ "authors": [ "karthikbhandary2", "khusheekapoor" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2880", "repo": "The-Data-Alchemists-Manipal/MindWave", "url": "https://github.com/The-Data-Alchemists-Manipal/MindWave/issues/454" }
gharchive/issue
Implementation of ML model on Happiness index data. 💥 Proposal I would like to apply a machine learning model to Happiness index data. By using the data I would like to predict the overall rank. Please assign me this. @karthikbhandary2 - please provide the dataset. This is present in this repo only. Link: https://github.com/The-Data-Alchemists-Manipal/MindWave/tree/main/Data Analytics/Happines_index_Data_analysis_visualization @karthikbhandary2 - you can go ahead! We are assigning you 21 days for this project, after which it will be assigned to someone else if not completed. All the best! Name the file as: algorithm_dataset.ipynb and link it in the readme of the labeled directory as algorithm - dataset. ok @khusheekapoor
2025-04-01T06:37:37.322059
2022-08-31T16:49:24
1357605921
{ "authors": [ "WebKingdom", "donn", "vijayank88" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2881", "repo": "The-OpenROAD-Project/OpenLane", "url": "https://github.com/The-OpenROAD-Project/OpenLane/issues/1299" }
gharchive/issue
Flow consistently fails regardless of configurations Description When hardening a SHA1 design inside the caravel user project, flow fails with FP_SIZING=relative and FP_CORE_UTIL=30%. Other configurations were also tried but hardening failed for all. Configuration trials can be seen in openlane/sha1_top/config.tcl (https://github.com/WebKingdom/bitcoin_asic/blob/055aed80953a4996cf8b931a32cf2a102ebe4ca6/openlane/sha1_top/config.tcl). Expected behavior The design should be able to harden and go through layout and placement. Environment Kernel: Darwin v21.6.0 Distribution: macOS 10.16 Python: v3.9.0 (OK) Container Engine: docker v20.10.17 (OK) OpenLane Git Version: f9b5781f5ef0bbdf39ab1c2bbd78be8db11b27f2 pip: INSTALLED pip:venv: INSTALLED --- PDK Version Verification Status: OK --- Git Log (Last 3 Commits) f9b5781 2022-07-01T16:04:31+02:00 Fix a bug with `-overwrite` (#1171) - Anton Blanchard - (grafted, HEAD, tag: 2022.07.02_01.38.08) Reproduction Material issue_reproducible attached as zip. issue_reproducible.zip Logs Console logs: [STEP 18] [INFO]: Running Detailed Placement... [ERROR]: during executing openroad script /openlane/scripts/openroad/opendp.tcl [ERROR]: Exit code: 1 ... [ERROR]: Last 10 lines: [INFO DPL-0035] ANTENNA__14155__A1 [INFO DPL-0035] ANTENNA__14155__A1 [INFO DPL-0035] ANTENNA__20474__A2 [INFO DPL-0035] ANTENNA__20930__A [INFO DPL-0035] ANTENNA__20930__A [INFO DPL-0035] ANTENNA__16103__A0 [INFO DPL-0035] message limit reached, this message will no longer print [ERROR DPL-0036] Detailed placement failed. Error: opendp.tcl, 32 DPL-0036 @WebKingdom Update your config.tcl with following configuration resolve the issue: set ::env(CELL_PAD) 2 Also you can use set ::env(PL_TARGET_DENSITY) 40 to utilize more core area. I set those 2 variables but set PL_TARGET_DENSITY = 0.4 (not 40). I was able to get further to [STEP 39] with the following openlane.log: [INFO]: Writing Verilog... [INFO]: Running LEF LVS... [INFO]: Running Magic DRC... [INFO]: Converting Magic DRC Violations to Magic Readable Format... [INFO]: Converting Magic DRC Violations to Klayout XML Database... [ERROR]: There are violations in the design after Magic DRC. [ERROR]: Total Number of violations is 5 [INFO]: Saving current set of views in '../Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/results/final'... [INFO]: Generating final set of reports... [INFO]: Created manufacturability report at '../Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/reports/manufacturability.rpt'. [INFO]: Created metrics report at '../Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/reports/metrics.csv'. [INFO]: Saving runtime environment... [ERROR]: Flow failed. The manufacturability.rpt contained: Design Name: sha1_top Run Directory: /Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54 Magic DRC Summary: Source: /Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/reports/signoff/drc.rpt Violation Message "Min area of metal1 holes > 0.14um^2 (met1.7) "found 5 Times. Total Magic DRC violations is 5 LVS Summary: Source: /Users/somasz/Documents/GitHub/mpw_6c/caravel_design/caravel_bitcoin_asic/openlane/sha1_top/runs/22_09_01_01_54/logs/signoff/38-sha1_top.lvs.lef.log LVS reports no net, device, pin, or property mismatches. Total errors = 0 Antenna Summary: No antenna report found. If you update open_pdks to latest it may resolve this error. I do have the latest version of OpenLane and still get the error. Any suggestions? Thanks! OpenLane Git Version: f9b5781f5ef0bbdf39ab1c2bbd78be8db11b27f2 While creating this issue, you shared above OpenLane version. Can you post latest OpenLane version do you tried? sha1_top design tested at my end and am not seening magic drc error. [INFO]: Running LEF LVS... [STEP 43] [INFO]: Running Magic DRC (log: logs/signoff/43-drc.log)... [INFO]: Converting Magic DRC Violations to Magic Readable Format... [INFO]: Converting Magic DRC Violations to Klayout XML Database... [INFO]: No DRC violations after GDS streaming out. [STEP 44] [INFO]: Running OpenROAD Antenna Rule Checker (log: logs/signoff/44-antenna.log)... [STEP 45] [INFO]: Running CVC (log: logs/signoff/45-erc_screen.log)... [INFO]: Saving current set of views in 'results/final'... [INFO]: Saving runtime environment... [INFO]: Generating final set of reports... [INFO]: Created manufacturability report at 'reports/manufacturability.rpt'. [INFO]: Created metrics report at 'reports/metrics.csv'. [WARNING]: There are max fanout violations in the design at the typical corner. Please refer to 'reports/signoff/33-rcx_sta.slew.rpt'. [INFO]: There are no hold violations in the design at the typical corner. [INFO]: There are no setup violations in the design at the typical corner. [SUCCESS]: Flow complete. [INFO]: Note that the following warnings have been generated: [WARNING]: There are max fanout violations in the design at the typical corner. Please refer to 'reports/signoff/33-rcx_sta.slew.rpt'. I think this is fixed now?
2025-04-01T06:37:37.326790
2023-07-18T22:25:40
1810809629
{ "authors": [ "annapetrosyan26", "maliberty", "oharboe" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2882", "repo": "The-OpenROAD-Project/OpenROAD", "url": "https://github.com/The-OpenROAD-Project/OpenROAD/issues/3658" }
gharchive/issue
make write_lef and others write out files atomically Description A half-written .lef file can be lying around after a crash which complicates e.g. deltaDebug.py. Suggestion: modify all write_* commands to write to a temp file when the write completes, rename temp file to final name This way there are no partial files lying around Suggested Solution No response Additional Context No response Hello @maliberty @oharboe , I'd like to work on this issue. Could you please assign it to me ? @annapetrosyan26 You can simply create a pull request without being assigned this task, I don't have the role to assign github issues in this project. I haven't seen that @maliberty assigns github issues to new contributors, but perhaps that's something that would be approperiate? Thank you @oharboe for the clarification. Then I'll work on the issue and once the code is ready I will create a pull request. @maliberty Fixed, no? Fixed in https://github.com/The-OpenROAD-Project/OpenROAD/pull/5109
2025-04-01T06:37:37.332806
2024-09-26T18:21:12
2551259012
{ "authors": [ "jeffng-or", "oharboe" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2883", "repo": "The-OpenROAD-Project/megaboom", "url": "https://github.com/The-OpenROAD-Project/megaboom/pull/118" }
gharchive/pull-request
updated SDCs to define both read and write clocks PR to update SRAM SDC file to define both clocks Isnt this also used by BoomTile? In which case we need an additional .sdc file. Also, the BUILD bazel need to be updated to use the file. If I read the BUILD.bazel correctly, the current constraints.sdc is used by the SRAM, regfile, and L1MetadataArray abstract generation flows. It is not used by BoomTile (it uses constraints-boomtile.sdc). But, yes, the proposed change wouldn't work for dataArrayB, tag_array*, or L1MetadataArray, since their clocks are named either RW0_clk, R*_clk, W*_clk or just clock. So, it seems like we might be able to: Modify the existing constraints.sdc to use *_clk to find the clock ports and use it for all SRAMs and regfiles Have L1MetadataArray continue to use the existing constraints-boomtile.sdc since they both have just "clock" Please correct me if I'm wrong, but I think that will hook up the appropriate clocks for the abstract generation flow. Yes, this should work. lgtm. should this have gone through jenkins CI? Interesting, the Jenkins CI jobs got kicked off and passed: https://jenkins.openroad.tools/job/megaboom-Public/view/change-requests/job/PR-118-head/ https://jenkins.openroad.tools/job/megaboom-Public/view/change-requests/job/PR-118-merge/ Maybe there needs to be some hook up in GitHub so that it shows up in the Rules section? @vvbandeira , how do we register the CI as a check?
2025-04-01T06:37:37.338989
2021-06-29T18:04:43
932972008
{ "authors": [ "Swastyy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2884", "repo": "TheAlgorithms/C-Plus-Plus", "url": "https://github.com/TheAlgorithms/C-Plus-Plus/pull/1518" }
gharchive/pull-request
Create n_bonacci.cpp Description of Change Checklist [x] Added description of change [x] Added file name matches File name guidelines [x] Added tests and example, test must pass [x] Added documentation so that the program is self-explanatory and educational - Doxygen guidelines [x] Relevant documentation/comments is changed or added [x] PR title follows semantic commit guidelines [x] Search previous suggestions before making a new one, as yours may be a duplicate. [x] I acknowledge that all my contributions will be made under the project's license. Notes: @Panquesito7 See now this PR, I think you will be able to see it clearly now as I have recreated it from the correct fork. Thanks in advance 😄
2025-04-01T06:37:37.369491
2022-09-01T15:19:13
1359021652
{ "authors": [ "TheArkive", "gcailly" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2885", "repo": "TheArkive/GuiCtlExt_ahk2", "url": "https://github.com/TheArkive/GuiCtlExt_ahk2/issues/5" }
gharchive/issue
No property named GuiControl Hi, (Thank you for all your AHK libraries !) Here is an error message with AutoHotKey 2.0 beta 7. Oh crud, I may not have tested that as thoroughly as I should have. Thanks for the report. fixed
2025-04-01T06:37:37.412466
2021-03-22T13:57:02
837743819
{ "authors": [ "TheDiscordian" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2886", "repo": "TheDiscordian/ipfs-sync", "url": "https://github.com/TheDiscordian/ipfs-sync/issues/29" }
gharchive/issue
Do less files/mkdir calls Related to #28, currently we call files/mkdir for every single file updated/added. It'd speed up these operations a lot if we only did this once per directory. For context I just attempted to update 18k files totalling ~1.1GiB, it's taking over 12 hours on a powerful system with an NVME SSD. This is obviously not good enough. Each call can take 1-3s!! Testing a possible patch for this right now. I think my set took 13 or 14 hours to process. I'll see how long this one takes. Fixed in 02a37e96cf6e118647b0da899ca7573332f19fb8
2025-04-01T06:37:37.459292
2022-05-13T04:10:04
1234715041
{ "authors": [ "Spappz", "TheGiddyLimit", "jbowensii", "revilowaldow" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2887", "repo": "TheGiddyLimit/plutonium-addon-automation", "url": "https://github.com/TheGiddyLimit/plutonium-addon-automation/issues/26" }
gharchive/issue
Complex items and optionality Many things in 5e contain options, whereby the AE to activate depends on a choice made. This leads to two options for data: Split each option out into a different Foundry item (and therefore a different button on the sheet) Use a macro to create a pop-up querying which option the player would like to trigger The latter is neater and more intuitive imo, but it might be slow if a player never uses a certain option. We need to agree how to handle this—and whether we are happy to accept 'half-complete' items until a general solution is found—because there's probably a fucctonne of items that it applies to Only one item please imo. You're only importing one thing from plutonium, you only expect one thing. So needs to be a macro. But also how would you implement that macro, is there a template you'd need to use for every item like that, or can you structure the data such that the module/plutonium generates a standard macro from an array of options? is there a template you'd need to use for every item like that Midi QoL has a (non-AE) option to trigger a macro when an item is activated (i.e. clicked). Using Item Macro, we can bundle this macro with the item itself rather than import it into the normal macros directory. I agree it would be nice to have all the popups be 'consistent' with each other in styling, but the issue is that the macro will invariably have more programming than just the popup. That is, once the popup appears, you need to write code to handle what each button does, and that has to be included in the same macro! I suppose in principle the popup could be triggered by a core Plutonium function with reliable behaviour, so the PAD-only, item-specific macro only needs to call the initial 'create modal' function and provide instructions for each result. await plutonium.optionsPopup.create( "Body text <b>with HTML</b> and whatnot like.", "Button 1", "function1", "Button 2", "function 2" ); function function1 { executeCommands(); } function function2 { executeCommands(); } function closePopup { execute Commands(); } I agree, one item/feature in the game equals one item on the character sheet, so please use macros. I personally do not see a reason to dictate code structure, format, or styling. The reason to at least have a style guide is so other people can review the code easily, if it ever needs updates! This could be fixing a bug, an API change, Midi overhauling itself, etc. The only requirements I'd like to 'enforce', to be clear, are: Write comments (obvious) No minified code unless the human-readable one is available somewhere else (see my post above) Use the same pop-up formatting as everything else (which we can draft whether or not Giddy makes some Plutonium magic to make it easier; this is to provide a 'consistent' experience so users don't have to constantly reread every single pop-up because they're all phrased differently) Whatever solution I end up cobbling together for jamming macros into proceedings, I'll make it eslint-able, so a minimum bar for code styling will be in place Writing good code is a whole different kettle of fish, and (un)fortunately not machine-enforceable (yet 😓) kinda made irrelevant by CPR/etc. managing it better than anything any of us clearly have the effort to maintain 😏
2025-04-01T06:37:37.469558
2024-12-06T21:32:17
2723964282
{ "authors": [ "fercer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2888", "repo": "TheJacksonLaboratory/activelearning", "url": "https://github.com/TheJacksonLaboratory/activelearning/issues/28" }
gharchive/issue
Generated layers do not follow OME standard The generated layers, when saved as zarr files, do not follow the OME standard and cannot be opened back again correctly in napari. Solved in PR #29
2025-04-01T06:37:37.475736
2023-09-14T11:59:25
1896417738
{ "authors": [ "TheLartians", "ldeng-ustc" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2889", "repo": "TheLartians/ModernCppStarter", "url": "https://github.com/TheLartians/ModernCppStarter/issues/178" }
gharchive/issue
Suggestion to define main project name in a separate file The main project name (Greeter) is currently used across multiple files. So it requires modifications in all these files to use this template, which can be time-consuming and error-prone. I suggest defining the main project name in a separate file. This way, any changes to the name would only need to be made in one place, reducing potential errors and increasing efficiency. for example, we can create a new file info.cmake: # Note: update this to your new project's name and version set(MAIN_PROJECT_NAME Greeter) And then, include this file in other files: # ./CMakeLists.txt cmake_minimum_required(VERSION 3.14...3.22) # ---- Project ---- include(info.cmake) project( ${MAIN_PROJECT_NAME} VERSION 1.0 LANGUAGES CXX ) Then, one can start its own projects with changing only one Greeter. I write a demo in this branch, but I haven’t updated the documentation. If this suggestion is accepted, I can submit a PR. Hey thanks for the input! I've also been bugged by this, but decided to not bother as it would introduce extra complexity without not adding long term benefits. Also in an actual project, you would expect the project name to be hardcoded in multiple places instead of being encoded in a generic variable name.
2025-04-01T06:37:37.480357
2023-03-08T18:08:10
1615696830
{ "authors": [ "The-Great-Nothing", "TheLastBen", "blazing", "cxyzdroid90" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2890", "repo": "TheLastBen/fast-stable-diffusion", "url": "https://github.com/TheLastBen/fast-stable-diffusion/issues/1705" }
gharchive/issue
module not found error after entering Path_to_HuggingFace:XpucT/Deliberate I'm encountering this error ModuleNotFoundError Traceback (most recent call last) in 3 from IPython.utils import capture 4 from IPython.display import clear_output ----> 5 import wget 6 7 #@markdown - Skip this cell if you are loading a previous session that contains a trained model. ModuleNotFoundError: No module named 'wget' NOTE: If your import is failing due to a missing package, you can manually install dependencies using either !pip or !apt. To view examples of installing some common dependencies, click the "Open Examples" button below. Same here. Me too but yesterday it worked. Solution: put !pip install wget before import wget. Thank You! Still, later in the training cell an avalanche of errors happen so I guess we wait for the updated colab. Oh yes I see, hope they will fix it ASAP. fixed
2025-04-01T06:37:37.483560
2022-12-04T04:55:33
1474454438
{ "authors": [ "AlternativelyMaybe", "TheLastBen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2891", "repo": "TheLastBen/fast-stable-diffusion", "url": "https://github.com/TheLastBen/fast-stable-diffusion/issues/859" }
gharchive/issue
"LayerNormKernelImpl" not implemented for 'Half' No flames please, be gentle, it's my first time posting. :) I got this error message after successfully (so I thought) installing and then running a request: "LayerNormKernelImpl" not implemented for 'Half'. Any words of wisdom? What else can I tell you to help someone help me out? Thanks so much! Where exactly do you get this error ? After putting in search terms and hitting the Generate button. I just went back to the main screen and saw the code output that had been generated: https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb#scrollTo=PjzwxTkPSPHf Did this work? I copied and pasted but it wasn't getting the entire output as for some reason the paste results were truncated. I can also try copying to a .txt file if you need. That phrase shows up a few times in the text at the end of each block. Thanks! For clear instructions how to use it, look for AUTOMATIC1111 colab tutorial on YouTube. Thanks much, I'll look into that!
2025-04-01T06:37:37.494879
2019-06-19T08:59:56
457890301
{ "authors": [ "TheM4hd1", "canaksoy", "sbertix" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2892", "repo": "TheM4hd1/SwiftyInsta", "url": "https://github.com/TheM4hd1/SwiftyInsta/issues/25" }
gharchive/issue
/feed/reels_media/ Great repo. Is there any way we could have story highlights added as well? make sure update to version 1.0.5 Example: https://github.com/TheM4hd1/SwiftyInsta/blob/10db0b97b7508a5c492c9f178c35b63768b2811b/SwiftyInstaTests/SwiftyInstaTests.swift#L1330 Great! Unfortunately it does not seem to populate items for every Higlight, only the first one, and not reliably either. Quickly googling it seems to suggest that to retrieve the actual items (i.e. stories) we need to pass the id to the "reels_media" endpoint. Any plans to add that as well? Thanks @TheM4hd1 did you try passing ID like this? highlight:123882132324123 I see, I'll check it for you. Thanks man There is an endpoint: /feed/reels_media/ it accepts an array of user_ids like highlight:123882132324123 I've tried to implement it but I'm receiving an error from server {"message": "Invalid reel id list", "status": "fail"} try it yourself, see if you can fix it, maybe I'm doing wrong somewhere. {"message": "Invalid reel id list", "status": "fail"} I have to admit that I've been trying to implement this for a while, without success. I was hoping you would fare better, but we seem to have stumbled on the same error unfortunately. I'l let you know if I find anything, but as I see similar libraries used this endpoint but I didn't test them to see if they works or not. maybe we're missing something... idk I try with "supported_capabilities_new" but same response. anyone fixed? sample request POST: https://i.instagram.com/api/v1/feed/reels_media/ DATA: signed_body=bca1cf35fe840d851d3d37f488665e955fa517e06a521843b619fd44183e28e3.{"supported_capabilities_new":"[{"name":"SUPPORTED_SDK_VERSIONS","value":"9.0,10.0,11.0,12.0,13.0,14.0,15.0,16.0,17.0,18.0,19.0,20.0,21.0,22.0,23.0,24.0,25.0,26.0,27.0,28.0,29.0,30.0,31.0,32.0,33.0,34.0,35.0,36.0,37.0,38.0,39.0,40.0,41.0,42.0,43.0"},{"name":"FACE_TRACKER_VERSION","value":"10"},{"name":"segmentation","value":"segmentation_enabled"},{"name":"WORLD_TRACKER","value":"WORLD_TRACKER_ENABLED"}]","source":"feed_timeline","_csrftoken":"lQElejwdXHJToUUryVZWghOEN2X8GFn0","user_ids":["archiveDay:17960651299222776"],"_uid":"5889897609","_uuid":"1df4e0f8-fc98-4250-a6a4-56b455e75699"}&ig_sig_key_version=4 @TheM4hd1 how can we create signed_body with array inside. how to know which part will be sign? @canaksoy Here is an example of how to sign body. https://github.com/TheM4hd1/SwiftyInsta/blob/c2beaad164a49b84fe8254af25ac75e36e192f72/SwiftyInsta/API/Services/UserHandler.swift#L164 The sample request you attached above, is it a working version? I tried with signed_body samples but its hard to handle [String] with encoder. Yes its a working php sample. @TheM4hd1 @canaksoy Okay, I'll check it. Thanks. I tested function with 2 id, if you see there are some missing data in returned model, you can decode the returned data to access full info about request.
2025-04-01T06:37:37.545882
2022-12-18T02:35:22
1501710358
{ "authors": [ "P4ranoidAndroid" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2893", "repo": "TheRealArthurDent/renovate", "url": "https://github.com/TheRealArthurDent/renovate/pull/32" }
gharchive/pull-request
Update renovatebot/github-action action to v34.62.1 This PR contains the following updates: Package Type Update Change renovatebot/github-action action minor v34.60.0 -> v34.62.1 Release Notes renovatebot/github-action v34.62.1 Compare Source See the the changelog for changes in all releases. 34.62.1 (2022-12-17) Bug Fixes deps: update renovate/renovate docker tag to v34.62.1 (afc5d3c) v34.61.0 Compare Source See the the changelog for changes in all releases. 34.61.0 (2022-12-17) Bug Fixes deps: update renovate/renovate docker tag to v34.61.0 (57b8e85) Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Enabled. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 Ignore: Close this PR and you won't be reminded about this update again. [ ] If you want to rebase/retry this PR, check this box This PR has been generated by Renovate Bot. Branch automerge failure This PR was configured for branch automerge. However, this is not possible, so it has been raised as a PR instead.
2025-04-01T06:37:37.851045
2024-07-11T15:36:57
2403502904
{ "authors": [ "TheRedHatter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2894", "repo": "TheRedHatter/CE", "url": "https://github.com/TheRedHatter/CE/pull/501" }
gharchive/pull-request
[Snyk] Security upgrade com.amazonaws:aws-java-sdk-core from 1.12.239 to 1.12.760 This PR was automatically created by Snyk using the credentials of a real user.![snyk-top-banner](https://github.com/andygongea/OWASP-Benchmark/assets/818805/c518c423-16fe-447e-b67f-ad5a49b5d123) Snyk has created this PR to fix 61 vulnerabilities in the maven dependencies of this project. Snyk changed the following file(s): api/pacman-api-admin/pom.xml Vulnerabilities that will be fixed with an upgrade: Issue Score Upgrade Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-608664   780   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 Reachable Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-450917   705   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Mature Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-467015   675   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Mature Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONCORE-3038426   670   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 Reachable Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1054588   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056416   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056418   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056420   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056421   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056426   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056427   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-174736   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-548451   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-559094   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-559106   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-560762   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561585   630   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1009829   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1047324   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056414   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056417   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056419   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056424   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1056425   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-467016   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-560766   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561362   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561373   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561586   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-561587   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-564887   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-564888   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-570625   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-572300   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-572314   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-572316   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72448   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72449   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72450   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72451   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72882   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72883   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-72884   563   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit XML External Entity (XXE) Injection SNYK-JAVA-COMFASTERXMLJACKSONCORE-1048302   560   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1052449   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1052450   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-1061931   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-455617   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-467014   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-469674   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-469676   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-471943   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-472980   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-540500   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-6056407   555   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONCORE-2421244   525   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONDATAFORMAT-1047329   525   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit Denial of Service (DoS) SNYK-JAVA-COMFASTERXMLJACKSONCORE-3038424   520   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Deserialization of Untrusted Data SNYK-JAVA-COMFASTERXMLJACKSONCORE-450207   520   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found Proof of Concept Information Exposure SNYK-JAVA-COMMONSCODEC-561518   485   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 Reachable No Known Exploit Improper Input Validation SNYK-JAVA-ORGAPACHEHTTPCOMPONENTS-1048058   415   com.amazonaws:aws-java-sdk-core: 1.12.239 -> 1.12.760 No Path Found No Known Exploit [!IMPORTANT] Check the changes in this PR to ensure they won't cause issues with your project. Max score is 1000. Note that the real score may have changed since the PR was raised. This PR was automatically created by Snyk using the credentials of a real user. Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs. For more information: 🧐 View latest project report 📜 Customise PR templates 🛠 Adjust project settings 📚 Read about Snyk's upgrade logic Learn how to fix vulnerabilities with free interactive lessons: 🦉 Deserialization of Untrusted Data 🦉 XML External Entity (XXE) Injection 🦉 Denial of Service (DoS) 🦉 More lessons are available in Snyk Learn Checkmarx One – Scan Summary & Details – f1ea51e7-8150-4369-b139-d654d1e8728c New Issues Severity Issue Source File / Package Checkmarx Insight Cleartext_Submission_of_Sensitive_Information /lambda-functions/notification-es-logging-service/src/main/java/com/paladincloud/notification_log/config/AuthManager.java: 44 Attack Vector Cleartext_Submission_of_Sensitive_Information /lambda-functions/notification-es-logging-service/src/main/java/com/paladincloud/notification_log/config/AuthManager.java: 44 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/commons/autofix/manager/AuthManager.java: 41 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-aqua-enricher/src/main/java/com/tmobile/cso/pacman/aqua/jobs/AquaDataImporter.java: 37 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/commons/autofix/manager/AuthManager.java: 41 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/commons/autofix/manager/AuthManager.java: 41 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-asset/src/main/java/com/tmobile/pacman/api/asset/service/AssetServiceImpl.java: 856 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/executor/PolicyExecutor.java: 141 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-rule-engine-2.0/src/main/java/com/tmobile/pacman/executor/PolicyExecutor.java: 141 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 281 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 72 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 67 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 281 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 72 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 67 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/service/AmazonCognitoConnector.java: 121 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-vulnerability/src/main/java/com/tmobile/pacman/api/vulnerability/service/VulnerabilityService.java: 973 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 281 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 72 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/azure-discovery/src/main/java/com/tmobile/pacbot/azure/inventory/auth/AzureCredentialProvider.java: 67 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 143 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/gcp-discovery/src/main/java/com/tmobile/pacbot/gcp/inventory/auth/GCPCredentialsProvider.java: 144 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 391 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 373 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 391 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 373 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 391 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AssetGroupUtil.java: 373 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Cleartext_Submission_of_Sensitive_Information /lambda-functions/notification-template-formatter-service/src/main/java/com/paladincloud/HttpUtil.java: 154 Attack Vector Cleartext_Submission_of_Sensitive_Information /lambda-functions/notification-send-email-service/src/main/java/com/paladincloud/utils/HttpUtil.java: 152 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 132 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 125 Attack Vector Cleartext_Submission_of_Sensitive_Information /jobs/pacman-data-shipper/src/main/java/com/tmobile/cso/pacman/datashipper/util/AuthManager.java: 22 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 125 Attack Vector Cleartext_Submission_of_Sensitive_Information /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/TenableAccountServiceImpl.java: 119 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20540 Attack Vector Client_Potential_XSS /webapp/src/app/shared/searchable-dropdown/searchable-dropdown.component.ts: 142 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21841 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21838 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21831 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21828 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21825 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 21822 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger.js: 63 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20550 Attack Vector Client_Potential_XSS /commons/pac-api-commons/src/main/resources/docs/v1/js/swagger-ui.js: 20547 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 106 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/controller/AccountsController.java: 59 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 89 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AwsAccountServiceImpl.java: 213 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AwsAccountServiceImpl.java: 216 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 155 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/QualysAccountServiceImpl.java: 158 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 151 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AquaAccountServiceImpl.java: 154 Attack Vector Privacy_Violation /api/pacman-api-admin/src/main/java/com/tmobile/pacman/api/admin/repository/service/AbstractAccountServiceImpl.java: 111 Attack Vector More results are available on AST platform
2025-04-01T06:37:37.867481
2024-05-18T08:44:00
2303935190
{ "authors": [ "TheRedHatter" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2895", "repo": "TheRedHatter/vulhub", "url": "https://github.com/TheRedHatter/vulhub/pull/61" }
gharchive/pull-request
[Snyk] Security upgrade php from 7.1-apache to 7.4.33-apache This PR was automatically created by Snyk using the credentials of a real user. Keeping your Docker base image up-to-date means you’ll benefit from security fixes in the latest version of your chosen image. Changes included in this PR base/gitlist/0.6.0/Dockerfile We recommend upgrading to php:7.4.33-apache, as this image has only 225 known vulnerabilities. To do this, merge this pull request, then verify your application still works as expected. Some of the most important vulnerabilities in your base image include: Severity Priority Score / 1000 Issue Exploit Maturity 929 Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740 Mature 929 Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740 Mature 929 Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740 Mature 929 Server-Side Request Forgery (SSRF) SNYK-DEBIAN10-APACHE2-1585740 Mature 886 Out-of-bounds Write SNYK-DEBIAN10-APACHE2-2322058 Mature Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs. For more information: 🧐 View latest project report 🛠 Adjust project settings Note: This is a default PR template raised by Snyk. Find out more about how you can customise Snyk PRs in our documentation. Learn how to fix vulnerabilities with free interactive lessons: 🦉 Server-Side Request Forgery (SSRF) Checkmarx One – Scan Summary & Details – 7b6409b7-8ce8-417a-8cec-254bc00144e0 No New Or Fixed Issues Found
2025-04-01T06:37:37.905061
2016-06-08T19:08:29
159247251
{ "authors": [ "Cat-McCatface", "Rumpelstiltskinny", "TheTemportalist" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2896", "repo": "TheTemportalist/EsoTeriCraft", "url": "https://github.com/TheTemportalist/EsoTeriCraft/issues/8" }
gharchive/issue
<EMAIL_ADDRESS> Server crash on killing blow to EnderZoo Wither Witch while using Tconstruct broadsword. fml-server-latest.log.txt crash-2016-06-08_11.57.10-server.txt Can you provide more ways and steps to reproduce? I got a similar crash. Happened when I entered menu to pause the game for an unrelated issue of hostile mobs ignoring players. Last entity killed was the only actual hostile entity, ironically a bat (part of rough mobs). forge: 1954 mc: 1.9.4 EsoTeriCraft-1.9.4-<IP_ADDRESS> Origin-1.9.4-9.1.6 log.txt
2025-04-01T06:37:37.941625
2017-09-26T09:42:49
260558830
{ "authors": [ "bassosimone", "coveralls" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2897", "repo": "TheTorProject/ooni-backend", "url": "https://github.com/TheTorProject/ooni-backend/pull/111" }
gharchive/pull-request
fix(report/handlers): accept more semver versions The current problem we have, specifically, is that since v1.2.0 the mobile apps emit full semver including also the build. This requires several parts of our infrastructure to improve the regexp being used so to validate also this kind of input. See also: measurement-kit/measurement-kit#1388 An earlier version of this diff was blessed by @hellais on Slack and since then I just changed comments. RET=$? +RET=0 if [[ "$(ps -ef | awk -v pid=$prog '$2==pid{print}{}')" != "" ]]; then kill -ALRM $prog wait $prog fi +kill -ALRM 6046 ps -ef | awk -v pid=$prog '$2==pid{print}{}' killitwithfire ++killitwithfire ++trap - ALRM ++kill -ALRM 6047 ++kill -9 6049 ++ps -ef ++awk -v pid=6047 '$2==pid{print}{}' /home/travis/.travis/job_stages: line 57: 6046 Segmentation fault (core dumped) ./.travis.test.sh 30 ./bin/oonib The command "chmod +x .travis.test.sh && ./.travis.test.sh 30 ./bin/oonib" exited with 139. I don't understand very well what it could be, and it seems like this is something related to the way in which travis is handling the test. I am going to restart it to see if it's deterministic (I actually hope so). I am experiencing more errors with the build, like that the keyserver doesn't know the key. 😡 Alright, it seems it's time to play golf. Coverage increased (+0.04%) to 78.469% when pulling 9e0eca426cacf7ec4e301fd612ed25d9c824e22a on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master. Coverage increased (+0.09%) to 78.527% when pulling ec91e00d53ea57413c7cb5301fc2d747433ee7d9 on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master. Coverage increased (+0.04%) to 78.469% when pulling 9bb6db5e2a453911873e6fa069f12a85466473fe on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master. I really hate these coveralls annoying comments that don't serve any purpose. Coverage increased (+0.09%) to 78.527% when pulling d09c1c59e09743a3986ef27ca4d31f6878125b80 on fix/version_regexp into 03d738b8daa86ed103e435a23adc16e9bac64127 on master. In the end, I decided to rewrite the test that caused a segfault. I believe the problem appeared now that travis has upgraded its infrastructure to 14.04: the previous travis build was for 03d738b and occurred on Apr, 6 when travis was still using 12.04. The original script was brilliant: it did two nested waits to make sure the running process was either killed gracefully or with fire. But probably was too brilliant and triggered some edge case. I did not want to wrestle too much with travis. Also, reading the script, it seems to me it's fine to rewrite it such that, if we cannot kill the background process, the build will hang and then fail (on travis). I guess having the local build hang and the travis build fail is good enough for our purposes. Alright, I have read the diff once more. I am going to self bless this as good, given that @hellais already blessed the diff improving the regexp on Slack and that the test changes "look good to me".
2025-04-01T06:37:37.955958
2022-10-04T04:52:48
1395664400
{ "authors": [ "NordicGamerFE", "Speiger" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2899", "repo": "TheUsefulLists/UsefulMods", "url": "https://github.com/TheUsefulLists/UsefulMods/issues/137" }
gharchive/issue
[Addition] World Pregenerators Checklist [X] I've checked that my mod suggestion hasn't already been added to one of the lists. [X] I've checked the latest issues to make sure that this mod hasn't already been suggested. What is your mod's name? Chunk Pregenerator What is your mod's main version? Other (Forge) What is your mod's other versions? 1.8.X or earlier, 1.12.X, 1.15.X (Forge), 1.16.X (Forge), 1.18.X (Forge), 1.19.X (Forge) What is your mod's type? Enhancement What side dose your mod need to run on? Worldgen lag is a real issue. Even in Singleplayer, and worldpregeneration is actually a really important thing you should do. This can turn it from "unplayable" to "playable" especially if your core count (including threads) is stuck at 4 or less or the CPU isn't the newest anymore... And even then its suggested to do so. That's why I am suggestion ChunkPregenerator. Which supports 1.4.7-1.7.x-1.8.9-1.10.x-1.11.x-1.12.x-1.14.x-1.15x-1.16x-1.18x-1.19.x Also another mod that I would suggest is Chunky Which is another pregenerator that provides a lot of performance gains and covers forge-fabric-customservers while chunkpregenerator is dedicated for forge. (OPTIONAL) Give some extra information about the mod. Chunk Pregenerator is a tool for generating your world before you actually play it. It also includes maintenance tools such as: Chunk Deletion/Trimming Performance Tracking (up to 1.12) World Maintenance tools (up to 1.12) Retro-generation that isn't relying on the Player Itself. Harddrive protection (1.14 or newer where it becomes nessesary) Memory Leak fixes that become apparent when pregeneration. And a few other things. I've added your mod to performance mods as i think that is a better fit then Enhancement. Love your mods, Keep up the good work :D @NordicGamerFE Small extra note since i saw this. Know issues + Fixes are actually tracked by myself. Here you find both lists. 1.12 or older: https://github.com/TinyModularThings/Chunk-Pregenerator-Issue-Tracker/issues/1 1.14 or newer: https://github.com/TinyModularThings/Chunk-Pregenerator-Issue-Tracker/issues/2
2025-04-01T06:37:37.962842
2024-11-29T12:42:40
2705012586
{ "authors": [ "AakashJaiswal-beta", "freeboub" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2900", "repo": "TheWidlarzGroup/react-native-video", "url": "https://github.com/TheWidlarzGroup/react-native-video/issues/4313" }
gharchive/issue
[BUG]: Mute/Volume 0.0 not working while screen recording Version 6.8.2 What platforms are you having the problem on? Android System Version All On what device are you experiencing the issue? Real device Architecture Old architecture What happened? I encountered an issue, when I screen record my application using, react native video, Screen Recorded video have the sound of the muted videos. In some phone the file is corrupted or is a lot laggy The saved video in the gallery have alot of sounds all over the components where the video is muted and volume is 0; DO NOTE: the sound is not being played in the app itself, muting and unmuting is working in the app, BUT on screen recording it, the sound of all those unpaused and muted video is there in the recorded video. The code snippet is: video config= Platform.select({ ios: { automaticallyWaitsToMinimizeStalling: false, bufferConfig: { minBufferMs: 2000, maxBufferMs: 5000, bufferForPlaybackMs: 1000, bufferForPlaybackAfterRebufferMs: 2000, }, }, android: { bufferConfig: { minBufferMs: 2000, maxBufferMs: 5000, bufferForPlaybackMs: 1000, bufferForPlaybackAfterRebufferMs: 2000, }, }, }); <Video source={{ uri: videoUri, }} ref={ref => { this.player = ref; }} onBuffer={this.onBuffer} onError={this.videoError} style={styles.thumbnailImage} paused={isPaused} repeat={true} resizeMode="stretch" muted={true} {...videoConfig} /> I even tried adding volume: 0, but it disnot work as well Reproduction Link Reproduction Step to reproduce this bug are: Initially when my video is paused nothing is happening, But when video is changed from paused to played, the video sound is coming in the screen recording of my app, DO NOTE: the sound is not being played in the app itself, muting and unmuting is working in the app, BUT on screen recording it, the sound of all those unpaused and muted video is there in the recorded video. look strange, but I don't know how we can fix it... I usually use another package to control device volume instead of player volume :/ maybe switching to texture view can fix the issue. you can have a try I think. Tried doing it on adding a viewType prop and changing it to textured view, still I am facing the same issue and sometimes the video is not openable by the phone, it says corrupted file or screen recording cannot be played, even on showing it on meet or any other screen sharing calls, It is creating an issue Ok, another solution can be to unselect audio tracks I think
2025-04-01T06:37:37.999878
2023-07-08T11:00:43
1794891640
{ "authors": [ "DerpDerpling", "Therkelsen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2902", "repo": "Therkelsen/echoes_of_the_wilderness", "url": "https://github.com/Therkelsen/echoes_of_the_wilderness/issues/8" }
gharchive/issue
Update Fadeless to 1.20.1 Currently, the mod is made for 1.20-Snapshot, and not 1.20.1. Update to 1.20.1, when the update comes out. Curse Forge Page It works fine on 1.20.1, no update is required. Another mod it works well together with is remove reloading screen , which removes the resource pack reloading screen. Great, I just noticed that it wasn't specifically for 1.20.1, so I figured that might cause issues. Also yeah, that mod is also part of the pack :D
2025-04-01T06:37:38.039593
2022-03-07T17:59:16
1161745137
{ "authors": [ "Thiritin", "mowny" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2904", "repo": "Thiritin/identity", "url": "https://github.com/Thiritin/identity/issues/10" }
gharchive/issue
It should log you out of all devices instead. Session handling is a bit weird anyway. After the session times out, a reload redirects to https://identity.eurofurence.org/auth/choose. Clicking "Login with existing account" gets me directly to the dashboard without having to actually log in. Hey @mowny Thanks for your comment. Yes that is wanted as the openid connect Server saves a cookie depending on if you habe set the remember me. The clients all got a limited session time. Although that session time does not end the cookie session at the idp. E.x. the idp may have a cookie lifetime of 180 days meanwhile the apps only got an hour. The right solution should be to redirect when the session times out. Fixed by implementing #11, sessions are not concern of the idp. So session management can only be done by the idp for the IDP Apps not for the Reg as example. A fix for this could be backchannel logout.
2025-04-01T06:37:38.041979
2015-03-27T01:51:32
64672965
{ "authors": [ "ThirteenAG", "jm10087" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2905", "repo": "ThirteenAG/Widescreen_Fixes_Pack", "url": "https://github.com/ThirteenAG/Widescreen_Fixes_Pack/issues/28" }
gharchive/issue
[GTA:SA] IVRadarScaling Would it be possible to add the IVRadarScaling option to the GTA: San Andreas fix like in the Vice City and GTA 3 fixes? I guess, but maybe later.
2025-04-01T06:37:38.111505
2023-07-29T00:09:54
1827229336
{ "authors": [ "scala-steward" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2907", "repo": "ThoughtWorksInc/enableIf.scala", "url": "https://github.com/ThoughtWorksInc/enableIf.scala/pull/127" }
gharchive/pull-request
Update scalafmt-core to 3.7.11 About this PR 📦 Updates org.scalameta:scalafmt-core from 3.7.5 to 3.7.11 📜 GitHub Release Notes - Version Diff Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scalameta", artifactId = "scalafmt-core" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "org.scalameta", artifactId = "scalafmt-core" } }] labels: library-update, early-semver-patch, semver-spec-patch, commit-count:1 Superseded by #128.
2025-04-01T06:37:38.126839
2021-07-28T02:16:25
954407753
{ "authors": [ "Tienisto", "ffshy1214" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2908", "repo": "Tienisto/flutter-fast-i18n", "url": "https://github.com/Tienisto/flutter-fast-i18n/issues/32" }
gharchive/issue
'in' can't be used as an identifier because it's a keyword. iso 3166 country code of India have problem, as below: Thanks for this issue. I will probably add a rename from in to indian By the way, it is more common to use language codes. Or language code + country code. Using country code only is pretty rare Fixed in 5.0.3. By the way, it is more common to use language codes. Or language code + country code. Using country code only is pretty rare. Also be aware that country codes must be in uppercase, e.g. CN not cn. Otherwise it is interpreted as language code like zh of zh-Hant-TW Thanks a lot for the quick response, and We'll consider your suggestion. Thanks
2025-04-01T06:37:38.127981
2018-09-25T03:52:26
363400142
{ "authors": [ "eteters" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2909", "repo": "TigerHacks/app-ios", "url": "https://github.com/TigerHacks/app-ios/pull/49" }
gharchive/pull-request
Api/sponsors implementation I did it I think... I may have missed some good error handling and may have left a todo in there somewhere but I'll let you guys tell me if things still need to change Unless I'm reading this wrong (which I definitely could be) the only travis errors I'm getting are for naming enum cases with a lowercase letter, which I do to match up with the variable name in the structs and the name of the api fields for decodable to work. Can we add a thing to ignore the rule for enums? I'm hoping that would be the easiest fix.
2025-04-01T06:37:38.141527
2022-03-22T14:33:10
1176879889
{ "authors": [ "Shelnutt2", "thetorpedodog" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2910", "repo": "TileDB-Inc/TileDB-Cloud-Py", "url": "https://github.com/TileDB-Inc/TileDB-Cloud-Py/pull/242" }
gharchive/pull-request
Allow raw installation outside of the context of PyPI. If the package is installed outside of PyPI or an active checkout (e.g. via pip install [some-path]), a version.py file might not be created. Use a local version number in that case. @thetorpedodog I've used python setup.py develop --user as my go-to for a while. Are you trying to get it working from the top level repo folder without any install? Figured out what spurred me to do this. It would cause an error if you pip-installed directly from git: pip install --upgrade git+https://github.com/TileDB-Inc/TileDB-Cloud-Py.git@some-version-hash
2025-04-01T06:37:38.145059
2024-11-07T02:06:46
2639644963
{ "authors": [ "jdblischak" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2911", "repo": "TileDB-Inc/cellxgene-census-feedstock", "url": "https://github.com/TileDB-Inc/cellxgene-census-feedstock/pull/17" }
gharchive/pull-request
Bump to 1.16.2 and update requirements This PR does a few related things: Bumps version to 1.16.2 (and resets build number to 0) Builds for Python 3.12 (replaces #15 because 1.15.0 couldn't be built against 3.12) Use the new conda-forge syntax to control and test the minimum supported Python version. Upstream supports >=3.10 Also note that this version 1.16.2 will not be able to be installed in the TileDB Cloud py39 environment since it requires py>=310. The goal is to install it in the updated py312 environment that is in progress Confirmed that the version number is still working as expected (#12): import: 'cellxgene_census' + python -c 'import cellxgene_census; print(cellxgene_census.__version__)' 1.16.2 + pip check No broken requirements found.
2025-04-01T06:37:38.147026
2018-12-17T11:28:16
391669454
{ "authors": [ "TimBarham", "ruslan-bikkinin" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2912", "repo": "TimBarham/send", "url": "https://github.com/TimBarham/send/pull/1" }
gharchive/pull-request
Fix security vulnerabilities This PR fixes security vulnerabilities for current version of send-transform. All tests are passed. If you're going to go the route of updating my fork, I'd rather keep it in sync with the original as much as possible. That would mean applying my changes on top of the latest release of send. However, I notice you've updated some of the dependencies here to newer versions than those used by the latest release of send - does that mean the versions currently used by send have security vulnerabilities? @TimBarham just a friendly ping on this.
2025-04-01T06:37:38.149785
2024-07-04T20:51:27
2391452582
{ "authors": [ "Christopher-Chianelli", "triceo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2913", "repo": "TimefoldAI/timefold-solver", "url": "https://github.com/TimefoldAI/timefold-solver/pull/934" }
gharchive/pull-request
fix: Do not throw an exception when the solution or entity classes are interfaces The superclass of an interface is null. This caused this line of code to throw an exception for solution classes that are interface: var superclass = bottomClass.getSuperclass(); lineageClassList.addAll(getAllAnnotatedLineageClasses(superclass, annotation)); Since getAllAnnotatedLineageClasses expected superclass to not be null. getAllAnnotatedLineageClasses now returns an empty list for null arguments. I'm wondering if solutions as interfaces should be supported. What would be the use case? What would be the downsides? Considering that in all the years we have not seen anyone ask us for this, maybe we don't need this.
2025-04-01T06:37:38.191346
2023-01-25T10:08:54
1556378592
{ "authors": [ "AurelVU", "Tkko", "bugrevealingbme" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2914", "repo": "Tkko/Flutter_Pinput", "url": "https://github.com/Tkko/Flutter_Pinput/pull/118" }
gharchive/pull-request
replaced FocusTrap to TapRegion I replaced FocusTrap to TapRegion because FocusTrap is not supporten now in Flutter 3.7.0 (https://github.com/flutter/flutter/pull/107262) .. Hey @AurelVU, I'm sorry for closing your PR without merging it but I was already working on it. The update is live on pub Version 2.2.22
2025-04-01T06:37:38.192257
2023-02-23T10:38:02
1596608685
{ "authors": [ "StunninglyWrong" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2915", "repo": "Tlesis/SquakePlusPlus", "url": "https://github.com/Tlesis/SquakePlusPlus/pull/4" }
gharchive/pull-request
Update to 1.19.3 Update required libraries. By the way, do you ever plan on uploading this to either or both Curseforge and Modrinth? Also, what about cutting off sharking from the mod if it keeps being unusable?
2025-04-01T06:37:38.200529
2024-07-09T10:34:07
2397818486
{ "authors": [ "Chaotixu", "TobyAdd", "exploitle" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2916", "repo": "TobyAdd/GDH", "url": "https://github.com/TobyAdd/GDH/issues/243" }
gharchive/issue
frame stepper doesnt work in geode annoying frame stepper isn't even a feature in the mod yet. https://github.com/TobyAdd/GDH/issues/198 the feature is back in 4.6.4
2025-04-01T06:37:38.247602
2019-10-10T20:23:34
505491054
{ "authors": [ "FutureFutureTo", "TomLingham" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2917", "repo": "TomLingham/Laravel-Searchy", "url": "https://github.com/TomLingham/Laravel-Searchy/issues/103" }
gharchive/issue
SQLite Testing Environment Does this package work w/ a sqlite testing environment? I'm getting the following error when using 'TEST' as the query string: +errorInfo: array:3 [ 0 => "HY000" 1 => 1 2 => "near "'T%E%S%T%'": syntax error" ] The query in my controller is as follows: $query = Searchy::search('products') ->fields('title', 'producer') ->query($q) ->getQuery() ->having('relevance', '>', 20) ->limit(20) ->pluck('id') ->toArray(); Unfortunately, this package only works with a MySQL database. It uses MySQL specific features to calculate match relevance.
2025-04-01T06:37:38.281488
2018-08-09T14:42:40
349159039
{ "authors": [ "LabShareLouie", "TonyGermaneri" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2919", "repo": "TonyGermaneri/canvas-datagrid", "url": "https://github.com/TonyGermaneri/canvas-datagrid/issues/154" }
gharchive/issue
Questions on columnOrder API Hi @TonyGermaneri, Thanks for the quick feature release for columns sorting. I've been playing around with it, and am able to retrieve the order using grid.columnOrder but cannot set the order. Looking over the docs, I assumed grid.columnOrder([1, 0 , 2, 3]) would work but got the error: grid.columnOrder is not a function May I get an example of how to set column order? Thanks, Louis PS: Theres a duplicate entry for columnOrder in the API documentation https://tonygermaneri.github.io/canvas-datagrid/docs/#canvasDatagrid.columnOrder All fixed! https://tonygermaneri.github.io/canvas-datagrid/docs/#canvasDatagrid.columnOrder Thanks again for pointing this out.
2025-04-01T06:37:38.292460
2014-07-10T22:18:55
37610798
{ "authors": [ "TooTallNate", "anprogrammer", "bdunlay" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2920", "repo": "TooTallNate/node-gyp", "url": "https://github.com/TooTallNate/node-gyp/pull/473" }
gharchive/pull-request
Adds support for custom GYP include When developers have libraries in non-standard locations (or don't wish to dump them into System32 on Windows), they need to run "node-gyp configure", edit the generated files to add include paths, and then run "node-gyp install" This is doubly painful when you wanted to "npm install" something, but it depends on a library being in a default path on your system. I've added the ability for users to include an additional ".gyp" file in all node-gyp builds by setting an environment variable (NODE_GYP_ADDITIONAL_CONFIG), which they can use to configure include/library paths. No longer must I manually install packages or dump files into global include directories. Has this been resolved in subsequent releases? I'm currently stuck trying to install packages that require libraries I've installed in custom locations. The preferred way to do this is to add a common.gypi file to the root of your module. For example: https://github.com/TooTallNate/node-vorbis/blob/master/common.gypi
2025-04-01T06:37:38.333590
2023-01-22T11:12:17
1552080517
{ "authors": [ "TousstNicolas", "Xyntexx" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2921", "repo": "TousstNicolas/JLC2KiCad_lib", "url": "https://github.com/TousstNicolas/JLC2KiCad_lib/issues/33" }
gharchive/issue
Proper Arc support for schematic and footprint The arcs for the schematic and footprint are rarely properly placed. It needs to be correctly handled. I pushed some work on the Fix_ARC branch, but I cannot figure out how to draw them, there are always components for which it does not work. For now, I tried to use the GetCenterParam function, which is reversed from https://easyeda.com/editor/6.5.5/js/editorPCB.min.js This function seems to return the coordinates of the center (or possibly the midpoint in some occasions ? ), and two angles, but when using theses to calculate the start point, end point and center, the result is inconsistent, sometimes it works, sometimes it doesn't. The schematic equivalent is more consistent and seems to have less issues. The following components have an arc in their footprint and could be used to test: C55684 C185659 C86002 C312983 C1341701 C307522 C689358 C403695 C602208 C152951 C688068 C163798 C661330 This concern the h_ARC function in the footprint : https://github.com/TousstNicolas/JLC2KiCad_lib/blob/b5c38c2beff6f710eb8ac427622384717539ce84/JLC2KiCadLib/footprint/footprint_handlers.py#L174-L278 I made my version of the code and it seems to be working pretty well so far: https://github.com/Xyntexx/JLC2KiCad_lib/tree/my_ARC_Fixes Thanks a lot for your work, merged in 12c6860aecd43dd8b4262347f74a3bb06de85959
2025-04-01T06:37:38.338065
2022-08-05T02:34:20
1329369236
{ "authors": [ "BCoyler", "Disctanger", "Narann" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2922", "repo": "Tow-Boot/Tow-Boot", "url": "https://github.com/Tow-Boot/Tow-Boot/issues/169" }
gharchive/issue
USB Mass Storage mode - eMMC unrecognized with pinephoneA64 Successful Tow-Boot Install - Tow-Boot 2021.10-004 was successfully installed on a pinephoneA64. - I first verified booting pOS from eMMC. - I then verified the Mass Storage mode successfully using a Windows10 host. - USB device was recognized as a "PinePhone (A64)", and the eMMC partitions are viewable. Mass Storage mode issue - After install; rebooted phone with USB plug still plugged in, but battery out. - Entered into Mass Storage mode, (blue LED). - Windows host now reporting, "Unknown USB Device (Device Descriptor Request Failed)". Mass Storage mode issue reoccurring across phone reboots - Tried again by disconnecting USB cable, inserting battery, and rebooting phone. - Entered into Mass Storage mode, (blue LED). - Plugged in USB cable. - Windows host still reporting, "Unknown USB Device (Device Descriptor Request Failed)". @BCoyler I was having very similar issue. Turns out i connected my pinephone to USB 2.0 port of the host device. Tried with different port(USB 3.0) port and device was recognized. If you haven't already found the solution. Give above method a try maybe it will resolve your issue as well. Had the same problem. For me, the problem cames from the provided (red) cable. Using another USB-C cable works at first try.
2025-04-01T06:37:38.341605
2023-11-27T20:00:02
2012999629
{ "authors": [ "louisfd" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2923", "repo": "Tracel-AI/burn", "url": "https://github.com/Tracel-AI/burn/pull/1006" }
gharchive/pull-request
fix/docs/chunk @dcvz I was a bit late to review your chunk operation. It's very nice but I found something was wrong with the doc. I made a test that explains why I'm changing it. Is this the behaviour you want? @dcvz now it should have the right behaviour
2025-04-01T06:37:38.383426
2023-08-29T13:19:43
1871662043
{ "authors": [ "CLAassistant", "Fluder-Paradyne", "andreacasarin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2924", "repo": "TransformerOptimus/SuperAGI", "url": "https://github.com/TransformerOptimus/SuperAGI/pull/1149" }
gharchive/pull-request
Using a standard .env file instead of config.yaml for docker deployment. Refactoring docker image example docker deployments to take a standard .env file with variables defined in the docker-compose itself to simplify docker custom deployments (eg. on Portainer). Description You will customize the installation via .env file instead of YAML. Related Issues No direct issue, just a Discord conversation. Solution and Design This setup helps deployments via docker for users which have a production-like environment. It's focused on the docker installation, it will not impact non-docker usage. Test Plan Type of change [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [x] Breaking change (fix or feature that would cause existing functionality to change) [ ] Docs update Checklist [x] My pull request is atomic and focuses on a single change. [x] I have read the contributing guide and my code conforms to the guidelines. [x] I have documented my changes clearly and comprehensively. [ ] I have added the required tests. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. With the lastest release I'm having another issue here: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 145, in __init__ self._dbapi_connection = engine.raw_connection() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 3288, in raw_connection return self.pool.connect() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 452, in connect return _ConnectionFairy._checkout(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 1267, in _checkout fairy = _ConnectionRecord.checkout(pool) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 716, in checkout rec = pool._do_get() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/impl.py", line 284, in _do_get return self._create_connection() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 393, in _create_connection return _ConnectionRecord(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 678, in __init__ self.__connect() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 902, in __connect with util.safe_reraise(): File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/util/langhelpers.py", line 147, in __exit__ raise exc_value.with_traceback(exc_tb) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 898, in __connect self.dbapi_connection = connection = pool._invoke_creator(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/create.py", line 637, in connect return dialect.connect(*cargs, **cparams) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/default.py", line 615, in connect return self.loaded_dbapi.connect(*cargs, **cparams) File "/opt/venv/lib/python3.10/site-packages/psycopg2/__init__.py", line 122, in connect conn = _connect(dsn, connection_factory=connection_factory, **kwasync) psycopg2.OperationalError: could not translate host name "super__postgres" to address: Name or service not known The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/opt/venv/bin/alembic", line 8, in <module> sys.exit(main()) File "/opt/venv/lib/python3.10/site-packages/alembic/config.py", line 632, in main CommandLine(prog=prog).main(argv=argv) File "/opt/venv/lib/python3.10/site-packages/alembic/config.py", line 626, in main self.run_cmd(cfg, options) File "/opt/venv/lib/python3.10/site-packages/alembic/config.py", line 603, in run_cmd fn( File "/opt/venv/lib/python3.10/site-packages/alembic/command.py", line 385, in upgrade script.run_env() File "/opt/venv/lib/python3.10/site-packages/alembic/script/base.py", line 582, in run_env util.load_python_file(self.dir, "env.py") File "/opt/venv/lib/python3.10/site-packages/alembic/util/pyfiles.py", line 94, in load_python_file module = load_module_py(module_id, path) File "/opt/venv/lib/python3.10/site-packages/alembic/util/pyfiles.py", line 110, in load_module_py spec.loader.exec_module(module) # type: ignore File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/app/migrations/env.py", line 94, in <module> run_migrations_online() File "/app/migrations/env.py", line 82, in run_migrations_online with connectable.connect() as connection: File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 3264, in connect return self._connection_cls(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 147, in __init__ Connection._handle_dbapi_exception_noconnection( File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 2426, in _handle_dbapi_exception_noconnection raise sqlalchemy_exception.with_traceback(exc_info[2]) from e File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 145, in __init__ self._dbapi_connection = engine.raw_connection() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", line 3288, in raw_connection return self.pool.connect() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 452, in connect return _ConnectionFairy._checkout(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 1267, in _checkout fairy = _ConnectionRecord.checkout(pool) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 716, in checkout rec = pool._do_get() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/impl.py", line 284, in _do_get return self._create_connection() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 393, in _create_connection return _ConnectionRecord(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 678, in __init__ self.__connect() File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 902, in __connect with util.safe_reraise(): File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/util/langhelpers.py", line 147, in __exit__ raise exc_value.with_traceback(exc_tb) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/pool/base.py", line 898, in __connect self.dbapi_connection = connection = pool._invoke_creator(self) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/create.py", line 637, in connect return dialect.connect(*cargs, **cparams) File "/opt/venv/lib/python3.10/site-packages/sqlalchemy/engine/default.py", line 615, in connect return self.loaded_dbapi.connect(*cargs, **cparams) File "/opt/venv/lib/python3.10/site-packages/psycopg2/__init__.py", line 122, in connect conn = _connect(dsn, connection_factory=connection_factory, **kwasync) sqlalchemy.exc.OperationalError: (psycopg2.OperationalError) could not translate host name "super__postgres" to address: Name or service not known Looks like alembic.ini has hardcoded db credentials. We have a pr for this, https://github.com/TransformerOptimus/SuperAGI/pull/1136 waiting on its review @andreacasarin What if I make a new docker compose definition, something like: docker-compose.prod.yaml which has the preferred vector store and explicit env vars for production like deployments? I can let the example there and add a new one. I think it would also be great to have a nginx image built so that you can actually deploy without pulling the whole repo. Perfect then, I'll restore the example and go with docker-compose.prod.yaml, just a couple more questions: which would be the preferred vector store? i'd keep the env file minimal, just the required env vars to start the project, then if I got this right, the rest can be configured via gui, right? Also, we still need https://github.com/TransformerOptimus/SuperAGI/pull/1149#issuecomment-1698503784 done to make it viable. which would be the preferred vector store? By default we are using redis vector store keep .env.dist the same, because we can set variables from the gui, but the vector stores in docker hub we have setup autobuilds so every push to main and dev gets build which would be the preferred vector store? By default we are using redis vector store Ok, but does it work adding knowledge to that? Because in my installation it asks to configure a new vector store to add knowledge. keep .env.dist the same, because we can set most variables from the gui, but the vector stores I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications. in docker hub we have setup autobuilds so every push to main and dev gets build Ok! Pushed, it requires nginx/DockerfileNginx to be added to the autobuild as superagi/superagi-proxy. Pushed, it requires nginx/DockerfileNginx to be added to the autobuild as superagi/superagi-proxy. i'll add it keep .env.dist the same, because we can set most variables from the gui, but the vector stores I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications. To explain a bit more my point-of-view: I'd go with 1 base docker compose with the bare minimum and then maybe a couple more which will be used to override the base one for different applications (like local dev, local llm, prod) they could only override the base one in specific keys; we'll discuss more towards this, there are some issues with gui hmr I'd add a .env.dist with the minimum variables needed to start the project and point out that you can actually configure all the rest from the gui. The reason why I wanted to keep env.dist with all the keys is to let users know all the keys in the project in one place we'll comment everything out other than the bare minimum Just my 2 cents, I'll stop arguing :) No issues which would be the preferred vector store? By default we are using redis vector store Ok, but does it work adding knowledge to that? Because in my installation it asks to configure a new vector store to add knowledge. sorry forgot about knowledge, asked @Tarraann ( main contributor to knowledges ) to use knowledge for now only pinecone, weviate and qdrant are supported so keeping the weviate commented out should be good for users to quickly set it up keep .env.dist the same, because we can set most variables from the gui, but the vector stores I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications. in docker hub we have setup autobuilds so every push to main and dev gets build Ok! keep .env.dist the same, because we can set most variables from the gui, but the vector stores I'm not sure I got it right, but I'll leave it like that if that is what you prefer. I'd only like to point out that it gets a bit overwhelming to have so many vars in there, especially since there are also 6 docker compose specifications. To explain a bit more my point-of-view: I'd go with 1 base docker compose with the bare minimum and then maybe a couple more which will be used to override the base one for different applications (like local dev, local llm, prod) they could only override the base one in specific keys; we'll discuss more towards this, there are some issues with gui hmr 👍 I'd add a .env.dist with the minimum variables needed to start the project and point out that you can actually configure all the rest from the gui. The reason why I wanted to keep env.dist with all the keys is to let users know all the keys in the project in one place we'll comment everything out other than the bare minimum I see, looks fine to me, I took the time to rearrange the variables there and put them under some comment headings, that might help this process.
2025-04-01T06:37:38.418044
2024-03-21T09:02:21
2199584151
{ "authors": [ "oscorops" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2926", "repo": "Tresjs/tres", "url": "https://github.com/Tresjs/tres/issues/594" }
gharchive/issue
如何设置背景颜色为透明? Description 如何设置背景颜色为透明? Suggested solution 如何设置背景颜色为透明? Alternative No response Additional context No response Validations [X] I agree to follow this project's Code of Conduct [X] Read the Contributing Guidelines. [X] Read the docs. [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. Ok its <TresCanvas clearColor: ' '>
2025-04-01T06:37:38.551890
2022-08-20T12:59:02
1345159555
{ "authors": [ "akinaritsugo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2927", "repo": "TripInsurance/devopsoh78828", "url": "https://github.com/TripInsurance/devopsoh78828/pull/3" }
gharchive/pull-request
Features/challenge4/userprofile PR Template Purpose ... Does this introduce a breaking change? [ ] Yes [ ] No Pull Request Type What kind of change does this Pull Request introduce? [ ] Bugfix [ ] Feature [ ] Code style update (formatting, local variables) [ ] Refactoring (no functional changes, no api changes) [ ] Documentation content changes [ ] Other... Please describe: How to Test Get the code git clone [repo-address] cd [repo-name] git checkout [branch-name] npm install Test the code What to Check Verify that the following are valid ... Other Information NG
2025-04-01T06:37:38.556185
2023-07-22T03:42:16
1816594901
{ "authors": [ "TripitakaBC", "spilegi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2928", "repo": "TripitakaBC/ampinvt_esphome", "url": "https://github.com/TripitakaBC/ampinvt_esphome/issues/4" }
gharchive/issue
Multiple Charge Controllers Sorry for raising an issue for this, but I have multiple Charge controllers. You can assign them numbers. I don't see anywhere in the code where it's referencing the unit number (It's 1 by default) Is it possible to specify that in the protocol? Not a problem! The charge controller is referenced in the hex code string sent over UART by the ESP32. Look for this code block in your YAML file: time: - platform: homeassistant id: esptime on_time: - seconds: 0,30 then: - uart.write: id: uart_bus data: [ 0x01, 0xB3, 0x01, 0x00, 0x00, 0x00, 0x00, 0xB5 ] # Reads only real-time data I think the controller address is the first hex block so you would need to change that and recalculate the checksum. That part is pretty easy. The bigger part of the work is that you would need to duplicate the ampinvt.h file to create separate sensors for each charge controller. If you don't need or want all the sensors then you could cut a lot of it out and just keep the sensors you did want but essentially, you have two parts to this; 1) sending the command to each controller (easy and quick) and 2) processing the response from each controller (harder and more time). Thanks, I'll go ahead an give it a shot. I currently have 6 of these, and was hoping to not have to purchase 6 ESP32's. Lot's of cabling and mess. Two questions I'm a Network Engineer who only dabbles in Computer Science. How would I calculate the checksum? If I were to duplicate only specific sensors, how would I differentiate between which charge controller it was coming from? I do not see a reference to the controller address in the ampinvt.h file. Thanks, PS. I'm ok with being told it's too far over my head, in that case I'll just purchase more ESP32's and do it that way. ESP32 I have setup connect to one of my charge controllers has been rock solid this entire time during testing. One ESP32 will do the job and even 1 ampinvt.h file but what you will need is to setup the ampinvt.h to read 6 x sensor sets and replicate all those custom sensors in your YAML file. If you go the 6 x ESP32 setup, you will still need to recode the sensors because they all come together in HA and HA will not know how to process 6 instances of the same sensors. When you figure out the renaming for those, just work backwards to the ampinvt sensor naming.
2025-04-01T06:37:38.567278
2015-11-04T17:20:43
115102875
{ "authors": [ "ItzDan", "TheSpaceArmy" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2929", "repo": "Tromino/PolyExtend", "url": "https://github.com/Tromino/PolyExtend/issues/12" }
gharchive/issue
Team groups going off edge of page. Most likely known of but here: http://i.imgur.com/c4aJxd0.png its a Known Bug <B Would love This Fixed https://surl.im/i/8qvwj
2025-04-01T06:37:38.601633
2017-05-11T16:17:47
228047273
{ "authors": [ "coveralls", "kevinansfield" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2930", "repo": "TryGhost/Ghost-Admin", "url": "https://github.com/TryGhost/Ghost-Admin/pull/690" }
gharchive/pull-request
[WIP] ✨ use markdown-it for markdown previews refs https://github.com/TryGhost/Ghost/pull/8451 replaces SimpleMDE's default marked rendering with markdown-it adds markdown-it plugins to more closely match legacy Showdown behaviour footnotes highlight/mark named headers don't require a space after the # for headers adds ember-browserify so that markdown plugins that only provide CommonJS modules can be imported Coverage decreased (-0.2%) to 71.556% when pulling f190198f49cb38b32133abd721f528564f89d1d1 on kevinansfield:markdown-it into fbb46dc72c4380a632e73db37c4ae8f2370b2087 on TryGhost:master. Coverage decreased (-0.2%) to 71.556% when pulling 0f7dd8a26025fd7a05a389190746bebcc077cee1 on kevinansfield:markdown-it into 627a71e1a4ea95a9d0b61dd92b90c0d822c2f907 on TryGhost:master.
2025-04-01T06:37:38.630936
2020-04-09T16:40:23
597410275
{ "authors": [ "gargol", "matthanley", "naz", "vikaspotluri123" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2931", "repo": "TryGhost/Ghost", "url": "https://github.com/TryGhost/Ghost/issues/11729" }
gharchive/issue
[API] Ordering by meta_title is not possible Issue Summary It's not possible to make an API request (internally or externally) that sorts by meta_title. Moreover, sorting by a related table (e.g. posts_meta) is allowed, but the request fails. I think this is a regression from 2.x, after the post metadata got split out into a new table. Technical details: Ghost Version: 3.13.1 Forum Ref: https://forum.ghost.org/t/order-by-meta-title-doesnt-seem-to-work/13349 Possible related issue: The model base parseOrderOptions pulls in a list of permitted attributes from the child modal, including allowed relations. This allows you to set the order to one of those relations, which kind-of breaks the API request. For example, ordering by authors asc will give you a 400, but title asc works perfectly fine, and potato asc succeeds because you can’t sort by potatoes, and therefore the order is ignored. Might be related to this issue - https://github.com/TryGhost/Ghost/issues/11572, should be checked when investigated deeper. After deeper digging into the issue, confirming that problem is how we create the order query here and the way Post model's permittedAttributes function returns only posts table specific fields ignoring posts_meta fields. A closest candidate to explore a solution is tinkering bookshelf-relations plugin, which should be able to abstract the table split in a way that permittedAttributes gives back correct fields from posts_meta . This issue needs appetite do dig into possible solution through bookshelf-relations. Guesstimating 1 day. Note, while loading up on the context have observed another place which will need an update when the ordering is implemented - pagination plugin. Specifically, options.order should most likely come with resolved tableName because pagination plugin should not know about relations and not do the tableName + '.' + property concatenation. Braindump after playing extensively with ordering There are 2 separate problems that this issue raises: Main one - cannot order post resource by a legitimate field, which comes from a 1:1 relation (posts:posts_meta tables). Ordering behaves inconsistently for fields that are not meant to be ordered by: ignores unknown properties and returns 400 when trying to order by property like authors, which comes from a relation name. In case of any field that can't be ordered by, it should consistently ignore that field (throwing a validation error would break API compatibility, so this is a no-go) The key to both problems is how parseOrderOption function relies on permittedAttributes to calculate all the fields that could be ordered upon. (pt.1) Ordering by fields which come from 1:1 relation The missing piece here is a mechanism in model layer which would recognize a 1:1 relation and extract "orderable" fields out of related table. Additionally, the fields would have to come in a format that includes related table name so that pagination plugin could correctly form the query, e.g: ['posts_meta.meta_title ASC'] instead of ['meta_title ASC']. The solution I'm thinking of here is adding a orderAttributes method which would be used instead of permittedAttributes in parseOrderOption. orderAttributes would have the same values as permittedAttributes for tables with no additional relations logic, and would have special overrides on each model with "expanded" field names coming from related table. To make above solution more maintainable would need to create a declarative way of describing which fields from related model the orderAttributes could be taken from. One of the possible directions to explore is expanding hasOne relation in bookshelf-relations - plan to timebox this direction to half a day tomorrow (cc @matthanley). (pt. 2) Treating all non-orderable fields the same The solution here somewhat relates to the solution from pt. 1. When parsing order options the function should not rely on permittedAttributes because those are not the same fields that could be "ordered" upon. Adding a orderAttributes method to the base model would possibly solve the problem. Thoughts on holistically solving posts<>posts_meta ordering/filtering/change detection problems I've dug through bookshelf-relations aiming to figure out an abstraction which would allow solving current ordering, related filtering, and change-detection(1) problems(2). I have not found a way that would solve all of them in a holistic way. These are 3 distinct areas that have some or no relation to bookshelf-relations: Ordering - has to do with findPage method from pagination plugin and depends on buggy/outdated handling in parseOrderOption function. Nothing to do with bookshelf-relations because the ordering fields should be calculated independently IMO not just for Post model but all other models that have relations. Filtering - problems here stem from lack of explicit NQL configuration for posts_meta table in combination with mapping posts_meta filter fields to correct ones (using posts_meta. prefix). This kind of mapping should be done in serialization layer same way other field mappings are done there. Change detection - to fix this problem on a deeper level (the fix that was done through override to wasChanged() was a "patch") we could look into fixing bookshelf-relations change tracking. This might come through some special parameter passed with hasOne relation that posts_meta is declared with or invent a whole new relation for this specific situation. New relation could also help with elimination of patchwork that has to be done in Post model. This needs to be researched. Additional note, bookshelf-relations inherently deals with create/update/delete operations only and doesn't have anything to do with read operations that are needed in case of ordering and filtering. Conclusion with regards to current issue Two problems that were described in the comment above should be solved outside of bookshelf-relations. This proposed solution: adding a orderAttributes method which would be used instead of permittedAttributes in parseOrderOption. orderAttributes would have the same values as permittedAttributes for tables with no additional relations logic, and would have special overrides on each model with "expanded" field names coming from related table. seems like the most viable solution for now and would solve both problems. @matthanley would love to know if you have any feedback on this? I estimate implementing the proposed solution would take about a day. adding a orderAttributes method which would be used instead of permittedAttributes in parseOrderOption. orderAttributes would have the same values as permittedAttributes for tables with no additional relations logic, and would have special overrides on each model with "expanded" field names coming from related table @naz this looks like a reasonable approach to me 👍 After implementing the ordering idea I had in mind, got faced with a Bookshelf limitation I completely forgot existed. Have my experimentation available on this spike branch - https://github.com/TryGhost/Ghost/compare/master...naz:ordering-for-posts-meta-fields. To sum up the problem. In pagination plugin, when self.fetchAll is done it doesn't have post_meta table loaded into the query and it fails with "no such column" error (when trying to order by posts_meta.meta_title for example). The problem of not having loaded relations when fetching records in Bookshelf is summed up with references here - https://github.com/bookshelf/bookshelf/issues/1707#issuecomment-351026830. There are two possible ways to get around this problem which I've been thinking of: Extend pagination with detection of relations and build in a join to related table when the ordering query is built. We extend the query in similar way in filtering plugin through NQL (it adds joins and filters records at the same stage). The downside for this method, is possibly keeping yet another configuration that might be similar to one in filter. This could end up hard to maintain longterm. Explore fixing/extending Bookshelf itself and load up relations like posts_meta into the queryBuilder automatically. This potentially, would be a more maintainable approach. The downside is, I have no clue how hard this would be to achieve, would need to dive into bookshelf codebase to understand more. The problem seems to have been around in Bookshelf since 2014. @matthanley the point Hannah made earlier about approaching the problem through bookshelf-relations is now clearer to me. Probably the limitation I've rediscovered was the reason. This issue could become a scope creep, so I think approaching it by timeboxing and researching possible directions would be best approach forward. I'm thinking of timeboxing first approach above to half a day and see what comes out and then if we really end up with need for configs would research the second approach. Let me know what you think or have any questions about the issue itself! Summary of the discussion around future plan for this issue Short term plan (to be done now) would be implementing a solution which expands query builder object inside of pagination (or ordering) plugin based on additional configurations similar to ones done with NQL/filtering. Long term, aiming to develop a maintainable solution which is not based on configuration but rather on special relation type coming from Bookshelf or some other fix which might also address https://github.com/bookshelf/bookshelf/issues/202. Will timebox 1 day for research to figure out right questions and maybe possible solutions. A solution that is in mind right now is expansion of hasOne relation with an "alwaysFetched" parameter - this might be done through a plugin or from within bookshelf itself (needs digging around). Extend pagination with detection of relations and build in a join to related table when the ordering query is built. We extend the query in similar way in filtering plugin through NQL (it adds joins and filters records at the same stage). The downside for this method, is possibly keeping yet another configuration that might be similar to one in filter. This could end up hard to maintain longterm. This solution has landed in master through https://github.com/TryGhost/Ghost/pull/12226. Next up will be pushing pt. 2 and figuring out more generic way to handle relation inclusion in query builder object through Bookshelf/special plugin. This should help solve filtering/ordering and attribute change tracking all together.
2025-04-01T06:37:38.637726
2022-04-18T22:17:12
1207504489
{ "authors": [ "ErisDS", "guidefox" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2932", "repo": "TryGhost/Ghost", "url": "https://github.com/TryGhost/Ghost/issues/14508" }
gharchive/issue
Impersonation tokens do signup action if no account is found Issue Summary When a user is deleted it is expected that all tokens associated with that account are revoked. This is currently not the case. Creating a token before deleting a user creates the account again when pasted into the browser. Steps to Reproduce Create an access token for an account. Delete that account. Paste that token into your browser. User is created again. The user should NOT be created again and all tokens should be revoked when a user is deleted. Ghost Version 4.44.0 Node.js Version 16.14.2 How did you install Ghost? OS - Debian 11 with MariaDB 10.5.15 Database type MySQL 8 Browser & OS version No response Relevant log / error output No response Code of Conduct [X] I agree to be friendly and polite to people in this repository Hey there @guidefox. Ghost's magic links are based on JWTs, the tokens aren't stored and there's not really a concept of revocation here. What's happening is that the magic link has a fall back behaviour of creating a new account if no matching account is found. I realise that's a little jarring, but it's a brand new account that is created, not an old one being restored. I think it would make sense to pin the impersonation links to only be allowed to do signin, rather than falling back to signup, to make this a little less weird. Ah, that makes more sense. I think that the current behavior can be improved because it is a little bit jarring right now. Perhaps a dedicated button for having impersonation tokens re-create the account instead of doing it automatically would be a better solution. And maybe make the tokens one use only? or at least provide the option to have it expire after one use. This has cropped up in other forms recently, and is something we want to prioritise fixing.
2025-04-01T06:37:38.665761
2023-07-27T02:22:40
1823498846
{ "authors": [ "ErisDS", "Grasume", "TheLaurenBarger", "daniellockyer", "davedub", "hussainb", "joe-blocher", "kilmarnock", "vikaspotluri123" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2933", "repo": "TryGhost/Ghost", "url": "https://github.com/TryGhost/Ghost/issues/17514" }
gharchive/issue
Make this site private - not working in Chrome and Opera Issue Summary Change or delete the row 58: versions/5.54.4/core/frontend/apps/private-blogging/lib/middleware.js return session({ name: 'ghost-private', maxAge: constants.ONE_MONTH_MS, signed: false, sameSite: 'lax' <----- row 58: instead of 'none' or delete row })(req, res, next); Or you can delete the row 58 because sameSite: 'Lax' is the default value. You can't code 'secure' within an object - secure: true will not work. Works now in Chrome and Opera. See https://web.dev/i18n/en/samesite-cookies-explained Steps to Reproduce See https://forum.ghost.org/t/make-this-site-private-not-working/39938/1 Ghost Version 5.54.4 Node.js Version v18.15.0 How did you install Ghost? local, macos Database type SQLite3 Browser & OS version No response Relevant log / error output No response Code of Conduct [X] I agree to be friendly and polite to people in this repository In core/frontend/apps/private-blogging/lib/middleware.js: ´´´ const privateBlogging = { … return session({ name: 'ghost-private', maxAge: constants.ONE_MONTH_MS, signed: false, // sameSite: 'none' <——— replace this with 2 lines below sameSite: urlUtils.isSSL(config.get('url')) ? 'none' : 'lax', secure: urlUtils.isSSL(config.get('url')) })(req, res, next); }, ´´´ and all is fine! Implement in the same way as you did in core/server/services/auth/session/express-session.js Hey there, thank you so much for the detailed bug report. That does look like something that shouldn't happen! A PR to fix this issue would be very welcome 🙂 I have made the PR Hi, I am facing the same issue in the latest version of Ghost, unable to login to the private site using Chrome based browsers. It's nearly three month later ... an nothing happened. But for me it closed, because I'm working locally :-) https://forum.ghost.org/t/make-this-site-private-not-working/39938 It's nearly three month later ... an nothing happened. But for me it closed, because I'm working locally :-) https://forum.ghost.org/t/make-this-site-private-not-working/39938 yeah, they didn't care to merge it. but I appreciate you for your troubleshooting and the fix. I think just a rerun would be required to pass the build, otherwise the PR is already approved: https://github.com/TryGhost/Ghost/actions/runs/6057836235/job/16821076886?pr=17938 I have made my first PR. I think the 2. one failed: Merge branch 'main' into joe-blocher-patch-1 https://github.com/TryGhost/Ghost/pull/17938/commits/ae0f64eb578f4aa43248da1b3e807a1f0c3b9bef I don't really know what is for and how I can delete this PR... Am 16.10.2023 um 08:45 schrieb Hussain @.***>: It's nearly three month later ... an nothing happened. But for me it closed, because I'm working locally :-) https://forum.ghost.org/t/make-this-site-private-not-working/39938 https://forum.ghost.org/t/make-this-site-private-not-working/39938 yeah, they didn't care to merge it. but I appreciate you for your troubleshooting and the fix. I think just a rerun would be required to pass the build, otherwise the PR is already approved: https://github.com/TryGhost/Ghost/actions/runs/6057836235/job/16821076886?pr=17938 https://github.com/TryGhost/Ghost/actions/runs/6057836235/job/16821076886?pr=17938 https://user-images.githubusercontent.com/4962633/275414297-2abc6c33-ed0c-442d-9e81-9cdbd0e405cd.png — Reply to this email directly, view it on GitHub https://github.com/TryGhost/Ghost/issues/17514#issuecomment-1763832578, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWHNLAX5ONSU2FFFUG6MRN3X7TJZDANCNFSM6AAAAAA2ZM7QIU. You are receiving this because you modified the open/close state. Maybe @daniellockyer can help You did't fix the error: Make this site private - not working not working in Chrome and Opera SOLUTION - it told you in August 2023 and I have made the PR! versions/5.82.2/core/frontend/apps/private-blogging/lib/middleware.js `const privateBlogging = { .... return session({ name: 'ghost-private', maxAge: constants.ONE_MONTH_MS, signed: false, sameSite: urlUtils.isSSL(config.get('url')) ? 'none' : 'lax', <------------ insert this secure: urlUtils.isSSL(config.get('url')) <------------------------------ insert this //sameSite: 'none' <------------------------------------------ row 58: remove })(req, res, next); },` The pull request still not merged in version 5.82.2: Fixed private mode cookie for local development #17938 What makes you say the PR wasn't merged? The commit shows that it's been in releases starting from 5.70.0. I've downloaded the code: versions/5.82.2/core/frontend/apps/private-blogging/lib/middleware.js But the code is still the same: ` return session({ name: 'ghost-private', maxAge: constants.ONE_MONTH_MS, signed: false, sameSite: 'none' <------------------------------------------ why this? })(req, res, next); },`` The code being the same does not mean your PR was not merged. In this case it looks like this change ended up possibly breaking something else so it was reverted: https://github.com/TryGhost/Ghost/pull/19298 The code being the same does not mean your PR was not merged. In this case it looks like this change ended up possibly breaking something else so it was reverted: #19298 OK so that means it is still a problem. I am running 5.79.6 (released Feb 26) and cannot make the site private because of this bug. What's the ETA on solving this? The code being the same does not mean your PR was not merged. In this case it looks like this change ended up possibly breaking something else so it was reverted: #19298 OK so that means it is still a problem. I am running 5.79.6 (released Feb 26) and cannot make the site private because of this bug. What's the ETA on solving this? My solution: I change always the code by myself, when I install an update. You have to change only 2 lines. The first time I reported the solution in August 2023. Maybe they will fix the bug sometimes ... Hey guys, any update on this one? I couldn't access the links behind why the commit was reverted. So not sure on the details or complexity of the bug, is there any progress towards figuring it out? Thanks, and I'm a huge Ghost fan 😊 Downloaded version 5.89.1 - this bug still not fixed You have to change only 2 lines in your code. The first time I reported the solution in August 2023. Why this is impossible? @daniellockyer this is disappointing that this is still an issue exp with docker involved. I can confirm that I have no access via chromium, chrome and edge. Works with firefox. Will tell that to my customers not. Hey, I'm sorry that it wasn't made clear when the related PR was reverted. Unfortunately the fix broke the theme preview in Ghost admin for private sites, when the admin and site URLs are configured differently (the recommended configuration). We are clearly missing some test coverage there, as the PR looked good to merge. In the meantime, whilst trying to understand this issue I wasn't able to reproduce it in Arc, Chrome or Chromium. There's something really janky going on here, because whilst there are clearly a couple of people here on this issue experiencing the problem, there's not a lot of wider noise despite private sites being used widely very successfully. Meanwhile, when we merged the PR and broke the theme preview for private sites, we heard about it instantly from many people. So there has to be a caveat that's not being covered here in the reproduction steps. I have a feeling that is something to do with SSL, which shouldn't be impacting production sites. I'm going to close this bug as it stands. If anyone has the detailed reproduction case, feel free to open a new issue and we can work through what cases should and shouldn't work & making sure that fixing this issue doesn't cause a more widespread issue for private site users. I wasn't able to reproduce it in Arc, Chrome or Chromium. I've updated Chrome and the bug disappeared. It was a bug in Chrome with the error-message: OK, many thanks
2025-04-01T06:37:38.672269
2018-07-30T09:22:11
345677414
{ "authors": [ "c0derabbit", "kevinansfield" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2934", "repo": "TryGhost/Ghost", "url": "https://github.com/TryGhost/Ghost/issues/9760" }
gharchive/issue
Ajax call from external website - Response 200 but no posts Issue Summary I’d like to fetch blog posts into my website client-side. When I fetch posts, I get this response, and no response body: ok: true redirected: false status: 200 statusText: "OK" type: "cors" url: "http://my_website/ghost/api/v0.1/posts/?limit=3&client_id=ghost-frontend&client_secret=MY_SECRET To Reproduce Follow this guide: https://api.ghost.org/docs/ajax-calls-from-an-external-website Try to fetch posts from an external website. Expected behaviour is to get a list of posts. When I curl it works fine. Technical details: Ghost Version: 1.25.1 Node Version: 6.14.3 Browser/OS: Chrome 67.0.3396.99 on the client, backend running on Ubuntu 16.04.5 (aws ec2 t2.micro) Database: mysql Thanks for looking into it :) Hey @c0derabbit 👋 We ask that you please do not use GitHub for help or support, the default issue template pointed you to our forum for this type of question 😄 We use GitHub solely for bug-tracking and on-going feature development so we try to keep it noise free. Many questions can be answered by reviewing our docs for self-hosters, our theme API, or our public API. If you can't find an answer then our forum is a great place to get community support, plus it helps create a central location for searching problems/solutions. FYI: Many projects have their own support guidelines and GitHub will highlight them for you, or the project owners will use issue templates to point you in the right direction, please read them before opening issues Hi @kevinansfield, I started there but no response, also it seems to me more like a bug than a question, as I did follow the instructions, and I suspect it's either not working as it should, or there's a missing step in the docs.
2025-04-01T06:37:38.738097
2018-03-08T11:16:29
303444258
{ "authors": [ "daniellockyer", "scidran" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2935", "repo": "TryGhost/node-sqlite3", "url": "https://github.com/TryGhost/node-sqlite3/issues/949" }
gharchive/issue
Sqlite3 doen't install into dory-node android app Hy. I'm testing a proget with dory node (app into google play from tempage.io). I've installed into a Samsung Tab A10.1 with android 7.0 and kernel 3.18.14, the gekko project from git hub. Then i used: npm install --only=production But there was error to install sqlite3. I tryed to install itmanually with npm install sqlite3 But i have error I attach this log npm-debug (npm bugs sqlite3).log npm-debug (npm install --build-from-source).log npm-debug (npm install sqlite3 --loglevel=info).log Before try to root my Tablet, i would like to know if there is an other solution... Thanks I'm trying to find different solution for the problem... I tried also to root my tablet but I had issue also with this procedure because my tablet is unsupported yet. I read carefully the wiki but I'm not able to understand how to do it. I'm a very newbie with github and npm so I can't understand some things. I have a package to clone from github (gekko from askmike); Before putting into terminal npm install --only=production after the cloning, I have understand that I must: download the package of sqlite3 (I downloaded the zip file from github: node-sqlite3-master.zip); then, where should I put the sqlite3 unzipped files? Into what directory of the cloned package (there isn't a node-modules directory after the cloning process... the directory is created after the install process)? After this, I have to open the file node.gyp but info the cloned directory, there isn't any *.gyp file... only after the install process a node-pre-gyp file exist and isn't a node-gyp file like the wiki. So I'm very confused! Would you be able to try on the latest version v5.0.3? 🙂
2025-04-01T06:37:38.744587
2023-08-04T08:17:07
1836288865
{ "authors": [ "Tsingularity", "ZY123-GOOD" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2936", "repo": "Tsingularity/dift", "url": "https://github.com/Tsingularity/dift/issues/6" }
gharchive/issue
How to choose features from OpenCLIP? Hello! Thanks for your great work. I don't know how to use OpenCLIP to find correspondences. Could you please share these codes? Thanks again. Hi, thanks for your interest in our work! The original OpenCLIP codebase didn't support input image size larger than the training resolution (at least it was still the case when we wrote our paper), so we follow the common practice and manually interpolate position encoding to support larger input resolution. This pull request could be very helpful as a reference. Feel free to let us know if you have more questions. @ZY123-GOOD I attached a cleaned-up version of our implementation here. Hope this helps!
2025-04-01T06:37:38.754745
2020-11-27T05:01:49
751982563
{ "authors": [ "RheingoldRiver", "turtleworks" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2937", "repo": "TsubakiBotPad/pad-cogs", "url": "https://github.com/TsubakiBotPad/pad-cogs/issues/272" }
gharchive/issue
transforminfo command Shows an overview of a card including the awakenings of both base & transformed version, as well as the cooldown it takes to transform Overview tab Description Show types of transformed version Show both pre-transform and transform (bc we wanna count total SB, SBR, etc) Show latents but don't bother showing caption that mentions pre-xform cos we know that Stats etc Leave as-is for now Skills Active skill (2cd) (Base: 30 -> 30) Leave description as-is for now Leave leader skill as-is for now Additional tabs Overview tab of base Overview tab of transformed card (Similar to how ^ls works) Now I see why you've asked me to comment: https://github.com/isaacs/github/issues/100 Now I see why you've asked me to comment: https://github.com/isaacs/github/issues/100
2025-04-01T06:37:38.764707
2024-05-07T02:52:05
2282163518
{ "authors": [ "joobus" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2938", "repo": "Tsuzat/NeoSolarized.nvim", "url": "https://github.com/Tsuzat/NeoSolarized.nvim/issues/21" }
gharchive/issue
How do I change the background color of a highlighted variable? In the screenshot below, what is the name of the highlight group which is making "lunarvim/colorschemes" appear with a yellow background? I would like to change the color. Figured it out: CurrentWord. Add this to your config to change the color: -- Add specific highlight groups on_highlights = function(highlights, colors) highlights.CurrentWord.bg = colors.blue end,
2025-04-01T06:37:38.772845
2016-04-12T15:01:48
147779523
{ "authors": [ "BoBeR182", "Tudmotu" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2940", "repo": "Tudmotu/gnome-shell-extension-bettervolume", "url": "https://github.com/Tudmotu/gnome-shell-extension-bettervolume/issues/5" }
gharchive/issue
Update for Gnome 3.20 Please update the extension to be compatible with Gnome version 3.20 Thanks :) Uploaded a new version.
2025-04-01T06:37:38.785817
2023-11-08T10:21:44
1983262428
{ "authors": [ "clizbe" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2941", "repo": "TulipaEnergy/TulipaEnergyModel.jl", "url": "https://github.com/TulipaEnergy/TulipaEnergyModel.jl/issues/237" }
gharchive/issue
Define the pipeline & decide ecosystem Description What do we need from Toolbox/ESDL and what's our pipeline going to be? [x] What are all the pieces of the pipeline? [ ] How much of the pipeline can SpineToolbox cover? [ ] How much of the pipeline can ESDL/EDR cover? [ ] What is left uncovered? [ ] Do we need to change the input of the model to "match" SpineToolbox? Related Issues Blocking #94, #106, #105, #118, #88, #89, #115, #36 The more I look at this, the more I think we should use SpineToolbox to integrate everything and the EDR as just one of several data sources. And maybe also be able to put results back into ESDL to use the MapEditor etc for analysis (depending on its capability). What I'm a bit concerned about is the whole system still working in 5 years. Seems a bit complex. Assigned myself although this is a group effort. WHAT WE WANT Build the network once (in a while) Use draft networks to build new networks Sufficient flexibility for ad-hoc code for experimentation Definition of temporal stuff Definition of solver specifications Be able to mix data sources (ESDL + ENTSO-E for example)
2025-04-01T06:37:38.851811
2016-01-25T22:20:03
128656772
{ "authors": [ "ArthurHoaro", "ksamuel", "rugk", "sametmax", "wankbank" ], "license": "WTFPL", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2942", "repo": "Tygs/0bin", "url": "https://github.com/Tygs/0bin/issues/101" }
gharchive/issue
zerobin.py runserver: Unknown option '--host' Using the command: python zerobin.py --host <IP_ADDRESS> --port 80 --compressed-static returns the error: zerobin.py runserver: Unknown option '--host' Changing it to: python zerobin.py host <IP_ADDRESS> port 8001 compressed-static returns: Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 3099, in run server = server(host=host, port=port, **kargs) File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 2723, in __init__ self.port = int(port) ValueError: invalid literal for int() with base 10: '<IP_ADDRESS>' Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 3099, in run server = server(host=host, port=port, **kargs) File "/usr/lib/python2.7/site-packages/bottle-0.12.9-py2.7.egg/bottle.py", line 2723, in __init__ ... Non-stop... Either the docs are incorrect, there's a bug, or I'm just stupid? I have the same issue. It looks like the way 0bin handles parameters changed. I've found this digging a bit in the code: def runserver(host='', port='', debug=None, user='', group='', settings_file='', compressed_static=None, version=False, paste_id_length=None, server="cherrypy"): So I managed to run 0bin on a custom port with: python zerobin.py <IP_ADDRESS> 8006 But it doesn't really make any sense: How am I supposed to enable compressed-static without setting previous parameters? I have to put dummy parameters to set the settings files? I'm probably missing something, but yes, the docs are outdated. All of this happen because I merged some PR and trusted the content instead of reading it all. I'm guilty of lazyness. I will have to roll back everything, and fix all the bugs one by one. I'm sorry for the mess, especially since I'm so slow at fixing it. Any news on this issue? BTW I also get issues with the command line suggested by @ArthurHoaro: $ python zerobin.py --host <IP_ADDRESS> --port 80 --compressed-static Traceback (most recent call last): File "zerobin.py", line 4, in <module> from zerobin.cmd import main File "/home/zerobin-python/0bin/zerobin/cmd.py", line 12, in <module> from sigtools.modifiers import annotate, autokwoargs ImportError: No module named sigtools.modifiers $ python zerobin.py <IP_ADDRESS> 8006 Traceback (most recent call last): File "zerobin.py", line 4, in <module> from zerobin.cmd import main File "/home/zerobin-python/0bin/zerobin/cmd.py", line 12, in <module> from sigtools.modifiers import annotate, autokwoargs ImportError: No module named sigtools.modifiers Can/Do I need to install sigtools.modifiers somehow? This has been fixed in the V2 branch. It will be merged in master and pushed to pypi soon.
2025-04-01T06:37:38.868609
2019-11-12T03:26:19
521305936
{ "authors": [ "Perryvw", "Sanjo", "ark120202" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2943", "repo": "TypeScriptToLua/TypeScriptToLua", "url": "https://github.com/TypeScriptToLua/TypeScriptToLua/issues/746" }
gharchive/issue
Transpiling Math.atan2 to math.atan2 Currently Math.atan2(y, x) is transpiled to math.atan(y / x). math.atan2 is available in Lua 5. math.atan2 (y, x) Returns the arc tangent of y/x (in radians), but uses the signs of both parameters to find the quadrant of the result. (It also handles correctly the case of x being zero.) Therefore Math.atan2(y, x) can be transpiled to math.atan2(y, x) for the extra functionality of quadrant checking and handling the case of x being zero. It is a little more complicated: Lua 5.1, 5.2 and I'm assuming lua JIT (should be checked) do support math.atan2. Lua 5.3 however does not have math.atan2, but instead takes an optional second argument to math.atan, which allows it to function as atan2 https://www.lua.org/manual/5.3/manual.html#pdf-math.atan The math.atan2 function is labelled as deprecated but still available in Lua 5.3. See https://www.lua.org/manual/5.3/manual.html#8.2 and I have tested if the math.atan2 function exists in Lua 5.3.5. So until the function is removed in a future version of Lua, it could be just transpiled directly to math.atan2. I have also verified that math.atan2 is available in LuaJIT 2.1.0-beta3. I have tested if the math.atan2 function exists in Lua 5.3.5 Are you sure about that? I don't have it in a standard 5.3.5 build, maybe you had some compatibility flags enabled?
2025-04-01T06:37:38.879747
2020-01-09T10:48:21
547400000
{ "authors": [ "Gerrit0", "jeremyrea", "kobezzza", "nknapp", "socsieng" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2944", "repo": "TypeStrong/typedoc", "url": "https://github.com/TypeStrong/typedoc/issues/1159" }
gharchive/issue
TypeDoc doesn't render comments with Handlebars 4.6.0 [x] I have checked issues with bug label and found no duplicates Expected Behavior Comments should be rendered when generating documents using globally installed typedoc. (i.e. npm install typedoc --global) Note that it works as expected when installed locally. # doesn't work - documentation rendered without comments typedoc --out docs # works - documentation rendered with comments node_modules/.bin/typedoc --out docs Expected: Actual Behavior Comments aren't included when run using typedoc global install. The comment section is missing from the html output: <section class="tsd-panel tsd-comment"> <div class="tsd-comment tsd-typography"> <div class="lead"> <p>Base class for animals</p> </div> </div> </section> Steps to reproduce the bug Command: npm install typedoc --global git clone https://github.com/socsieng/typedoc-plugin-typescript-declaration.git cd typedoc-plugin-typescript-declaration/example typedoc --out docs open docs/classes/_index_.example.animal.html Environment Typedoc version: 0.15.6 Node.js version: 8.16.0 npm version: 6.4.1 nvm version: 0.34.0 OS: macOS Catalina 10.15.1 (19B88) I'm experiencing a similar issue, although it's the local installation that's not rendering the comments in the HTML output. When I add the --json flag, I can see that my tags and shortText are picked-up in the object's comment property. +1 This is.... really weird. I can confirm the global/local issue, no idea what's causing it yet. Looking into it. @jeremyrea could you provide a repo with a repro for the issue when run locally? It seems like the global install has been broken for a long time<EMAIL_ADDRESS>also has this issue. @Gerrit0 https://github.com/jeremyrea/typedoc-comment-repro Looks like Handlebars is the cause of this break (global + local I'm guessing, I bet the local install that works has a lower version of handlebars pinned in package-lock.json) - https://github.com/wycats/handlebars.js/pull/1633... I'm not exactly sure how we should go about fixing this... listing out all of the prototype methods that we expect a template (as suggested in the handlebars PR) to be able to access isn't feasible and is very likely to break in the future whenever a new method is added. For now, I'll pin handlebars to a lower version and release a patch with that change. Fixed in v0.15.7, thanks for the report @socsieng + @jeremyrea! Leaving this open to track finding a better solution. I don't want to be stuck on an old version of handlebars forever. Thanks @Gerrit0, can confirm that it works for me. Handlebars 4.7.0 has been release with options to disable prototype restrictions: https://handlebarsjs.com/api-reference/runtime-options.html#options-to-control-prototype-access Thanks @nknapp! I'll release 0.15.8 with a handlebars version bump to 4.7.0 later today :) v0.15.8 is released
2025-04-01T06:37:38.882644
2022-06-01T08:40:25
1255373926
{ "authors": [ "Gerrit0", "Kivylius" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2945", "repo": "TypeStrong/typedoc", "url": "https://github.com/TypeStrong/typedoc/issues/1947" }
gharchive/issue
Use typedoc as remote type Search terms typedoc, typescript, remote, micro frontends Question I love typedoc and its generated docs, but is it possible to use the output of typedoc or even write some plugin to be able to use typedoc output as remote type (comment or import) e.g. /** typedoc:url website.com/typedoc/MyComponentType **/ const MyComponent = import('website.com/components/MyComponent'); - or - import MyComponentType from `website.com/typedoc/MyComponentType`; const MyComponent: MyComponentType = import('website.com/components/MyComponent'); This pattern would allow remote types, when using patterns like micro frontend and there no direct importing of the different repos. This could also aid in the generation of better docs, where the type is remote, we could say something like: import MyComponentType from `website.com/typedoc/MyComponentType`; type Something { MyComponent: MyComponentType } Let me know if this is possible, if not maybe point me in the right direction. You're probably after renderer.addUnknownSymbolResolver - https://github.com/TypeStrong/typedoc/blob/master/internal-docs/third-party-symbols.md
2025-04-01T06:37:38.901348
2021-07-05T17:39:27
937273081
{ "authors": [ "EremitaDelle6Vie", "KavyaKinjalk", "M123-dev", "berkayyildi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2946", "repo": "Tyrrrz/YoutubeDownloader", "url": "https://github.com/Tyrrrz/YoutubeDownloader/issues/222" }
gharchive/issue
MP3 downloaded music missing background image Version 1.8.2 Details When i download a song in MP3 version I don't get the image when I play it with VLC Media Player, there is only a black screen instead of the background (which is the thumbnail of the video I guess) Is that bug or a problem with my media player or you intended to remove this feature in this version? Steps to reproduce Download any video in MP3 version Open the file with a media player This only extracts audio from the video from YouTube, so there won't be any "tags" for the audio file, so this is not a bug, I suppose, as you are getting the audio only @KavyaKinjalk yeah but with the previous version I could do it, so I don't know if they removed this feature @EremitaDelle6Vie, I have the suffer from issue, I tried around with some older versions but it did not work there either, maybe its caused by some changes on the serverside. Same problem ++ Please add thumbnail images to mp3 files. I can add on my previous comments. In the Audi entertainment system the thumbnails were shown, but not in Windows. But I used a older version in the process of getting it to work again.
2025-04-01T06:37:38.952143
2021-07-28T15:48:40
954999569
{ "authors": [ "astrochun", "yhan818" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2948", "repo": "UAL-RE/ReQUIAM", "url": "https://github.com/UAL-RE/ReQUIAM/pull/161" }
gharchive/pull-request
Merge develop into master - Include Grouper 2.5 migration This is the primary PR now for testing Grouper 2.5 and implementation. Note: That the original feature was so outdated that required a git merge --no-ff master. Before this is merged, here are a few things to do: [ ] Bump the version from v1.0.0 to v1.0.1 [ ] setup.py [ ] requiam/__init__.py [ ] docs/source/conf.py [ ] Update CHANGELOG.md Note that we have two config files for v2.2 and v2.5 deployment. It's likely that when Grouper 2.5 goes public, that the grouper25.iam.arizona.edu will become grouper.iam.arizona.edu Closes #116 To test this PR, we do: ./scripts/script_run --config config/figshare_grouper25.ini --persistent_path /mnt/block3_sfo2/ --ldap_password $eds_pass --grouper_password $eds_pass --portal --quota --sync A dry run tested 99% of the cases except it does not update the quota etc. So after the grouper 2.5 upgrade, partial failures exists for automatic run. A manual run was carried out. A dry run was successful, but a "--sync" run failed in timeout. Updated "timeout =100" (originally timeout =60 in figshare.ini) solved this issue. The first batch of 100 took 73s. (that is the reason to timeout =60 failed). 17:41:48 - INFO: batch size = 100, batch timeout = 100 seconds, batch delay = 0 seconds 17:41:48 - INFO: processing drops: 17:41:48 - INFO: processing adds: 17:43:02 - INFO: added batch 1, 100 entries, 73.567426 seconds 17:43:12 - INFO: added batch 2, 100 entries, 10.436354 seconds 17:43:22 - INFO: added batch 3, 100 entries, 10.194059 seconds 17:43:32 - INFO: added batch 4, 100 entries, 9.949762 seconds 17:43:42 - INFO: added batch 5, 100 entries, 10.029692 seconds 17:43:45 - INFO: added batch 6, 33 entries, 2.717927 seconds 17:43:45 - INFO: QUOTA : Total time: 0 hours 7 minutes 17.11 seconds 17:43:45 - INFO: Total time: 0 hours 15 minutes 39.70 seconds 17:43:45 - INFO: ****************************** 17:43:45 - INFO: SUMMARY DATA num_EDS num_Grouper adds drops total fine_arts 258 258 0 0 0 performing_arts 426 426 0 0 0 architecture 178 178 0 0 0 arts_design 41 41 0 0 0 business_econ 424 424 0 0 0 management 272 272 0 0 0 economics 110 110 0 0 0 education 1301 1301 0 0 0 english 296 296 0 0 0 lang_culture 392 392 0 0 0 humanities 271 271 0 0 0 law 402 402 0 0 0 nursing 369 369 0 0 0 med_health 1949 1949 0 0 0 clinical 857 857 0 0 0 ped_reprod 197 197 0 0 0 neurology 104 104 0 0 0 oncology 280 280 0 0 0 pharmacology 723 723 0 0 0 physiology 134 134 0 0 0 public_health 492 492 0 0 0 astro 646 646 0 0 0 cognitive_sci 360 360 0 0 0 life_sci 1846 1846 0 0 0 sci_math 1898 1898 0 0 0 earth_sci 1047 1047 0 0 0 physics 200 200 0 0 0 lpl 310 310 0 0 0 agriculture 143 143 0 0 0 anthropology 207 207 0 0 0 social_sci 870 870 0 0 0 cultural_studies 169 169 0 0 0 history 72 72 0 0 0 journalism 245 245 0 0 0 engineering 1268 1268 0 0 0 technology 144 144 0 0 0 libraries 277 277 0 0 0 536870912 15097 15097 0 0 0 2147483648 24393 23860 533 0 533 17:43:45 - INFO: ****************************** 17:43:45 - INFO: Exit 0
2025-04-01T06:37:38.966050
2020-11-19T07:27:03
746322230
{ "authors": [ "chuangw46", "huan-ds" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2949", "repo": "UBC-MDS/Abalone_Age_Prediction", "url": "https://github.com/UBC-MDS/Abalone_Age_Prediction/pull/1" }
gharchive/pull-request
create code-of-conduct and contribution file I also created a .gitignore file which ignores unnecessary files. Reviewed by Huan