id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
13570801
AMQP Payloads Should Be Validated. AMQP payloads/messages are currently trusted to be in the correct format, and contain the correct data. Sensu needs to be more resilient to invalid payloads/messages, and report them. Validation is currently done on input instead of consumption: configuration client socket(s) API POSTs Trusting producers.
gharchive/issue
2013-04-24T04:43:34
2025-04-01T06:45:45.898943
{ "authors": [ "portertech" ], "repo": "sensu/sensu", "url": "https://github.com/sensu/sensu/issues/528", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
55163634
Client list does not indicate that it is truncated/paginated The client page (/#/clients) does not indicate that there are more clients than are currently listed. It would be appreciated if a "more" or "next set" button were provided toward the bottom of the page, and an indication along the lines of "showing 50 of approximately 300 clients" was shown at the top :+1: Wouldn't hurt if searching covered all pages as well. Yes, this. This enhancement is available since 0.6.0 so I'll close this issue! :+1:
gharchive/issue
2015-01-22T15:05:14
2025-04-01T06:45:45.901081
{ "authors": [ "dcws", "gwaldo", "palourde", "zarry" ], "repo": "sensu/uchiwa", "url": "https://github.com/sensu/uchiwa/issues/252", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2378402463
Make seqera-kit installable without using conda 'defaults' channel Hi, I noticed this in the docs for seqera-kit Conda You can install seqerakit and its dependencies via Conda. Ensure that you have the correct channels configured: conda config --add channels defaults conda config --add channels bioconda conda config --add channels conda-forge conda config --set channel_priority strict It sounds as if the "defaults" channel is a requirement to install? Is this really the case? To be honest I am not 100% sure what the "defaults" channel on conda entails, but it sounds as if its a "paid" licensed channel for commercial use? Sources; https://stackoverflow.com/questions/77572223/forbid-usage-of-defaults-channel-in-conda https://stackoverflow.com/questions/42309333/explanation-of-different-conda-channels We know that there are licensing issues with the usage of non-free conda package channels such as anaconda, and possibly defaults (?) https://stackoverflow.com/a/75881378 On the other hand, conda channels conda-forge and bioconda appear to be free for use without license requirements. Is it possible to get the install requirements for seqera-kit updated so that the defaults channel is not needed? Assuming that defaults does indeed evaluate to paid, licensed conda channels and packages. Otherwise, it would cause us to require a paid conda license in order to be able to use this channel to be able to install seqera-kit from conda (which is the preferred method). Thanks! I don't see a particular need to keep defaults, conda-forge should suffice for the most part, it encompasses almost all packages of the defaults channel and we definitely want to avoid conflicting with Anaconda's ToS. Thanks for bringing this our attention. Will remove this requirement in the next release. Changes made in #155, thanks for reporting this.
gharchive/issue
2024-06-27T14:59:21
2025-04-01T06:45:45.929349
{ "authors": [ "ejseqera", "stevekm" ], "repo": "seqeralabs/seqera-kit", "url": "https://github.com/seqeralabs/seqera-kit/issues/150", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
133954843
CLOUD-50248 if security block is defined only change the type @schfeca75 Jenkins build finished, all tests passed. Refer to this link for build results: http://ci.sequenceiq.com/job/ambari-rest-client-pull-request/121/
gharchive/pull-request
2016-02-16T11:25:18
2025-04-01T06:45:45.957976
{ "authors": [ "jenkins-sequenceiq", "keyki" ], "repo": "sequenceiq/ambari-rest-client", "url": "https://github.com/sequenceiq/ambari-rest-client/pull/109", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
342062394
Please add "Following" section Site now by default shows the recommended pins on the main page. And then, what you signed, moved to another section: https://pinterest.com/following/ Do you plan to add to the library a function that allows you to receive only what you are subscribed to? Thanks! @seregazhuk ? :'( It is already implemented. Use this method. @seregazhuk Method behaves like https://pinterest.com/following page: includes recent pins for these pinners. Не совсем. То есть, конечно, можно получить последние Пины в "recent_pin_images" юзеров, на которых ты подписан, но это не то же самое, что показывается в https://pinterest.com/following, потому что я, например, подписан на Доски (в большинстве случаев на 1 доску у 1 пользователя), а не на все Пины из всех их Досок. В https://pinterest.com/following показываются последние Пины только из тех досок, на которые ты подписан, а в "recent_pin_images" подгружаются последние Пины из всех Досок. Вот так они замудрили :) Я в браузере больше никакх запросов, которые бы отдавали пины на странице following не вижу. Единственный это https://www.pinterest.com/resource/UserFollowingResource/get, который используется в этом методе бота: Если есть какие-то предложения/замечания, буду благодарен 💪 @misterkek закрываю? @seregazhuk Думаю, пока можно закрыть, хотя вопрос частично остается открытый. Я написал в поддержку Пинтереста, чтобы они пояснили, каким методом через их официальное API можно получить данные из этого раздела. Если ответят, то я сюда напишу.
gharchive/issue
2018-07-17T19:57:35
2025-04-01T06:45:45.969386
{ "authors": [ "misterkek", "seregazhuk" ], "repo": "seregazhuk/php-pinterest-bot", "url": "https://github.com/seregazhuk/php-pinterest-bot/issues/402", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1558205044
[Bugfix] DeepFace.represent model parameter bugfix and completed parameters docstring Description This PR bugfixes the model parameter to DeepFace.represent() and added the complete parameter docstring. Current behavior DeepFace.represent() ignores the model parameter as it is always overwritten in the code invariantly, making the model parameter useless. model = build_model(model_name) New behavior DeepFace.represent() only uses the model_name parameter and build the model only if the model parameter wasn't sent (model = None). if model is None: model = build_model(model_name) Tests [x] Run tests/unit_test.py thank you for your contribution. actually i retired the model input argument in represent function and other functions. but it seems i forgot to drop it in represent function. thank you for your contribution. actually i retired the model input argument in represent function and other functions. but it seems i forgot to drop it in represent function. That's ok. I've reverted the change and removed the model parameter, but the docstring creating for the other parameters is still necessary, correct? yes, other arguments are still valid. thank you again.
gharchive/pull-request
2023-01-26T14:11:56
2025-04-01T06:45:45.974327
{ "authors": [ "ronaldpereira", "serengil" ], "repo": "serengil/deepface", "url": "https://github.com/serengil/deepface/pull/633", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
130543362
Binary results Hi dbeaver dev, Not sure if this is an issue or I just don't know how to configure dbeaver since this is my first time using it. If I execute my query dbeaver displays binary data as binary. How can I make it display as text? When I do the same query in SQLDeveloper I get a readable string representation of the binary data. I want dbeaver to do the same. Thanks alot! Could you clarify it a bit? What is the column type (VARBINARY, BLOB, .. ?). Maybe some screenshot would help. hey @crisarceramos, you can format binary results under preferences>common>result sets>Binaries. p.s. There is also a current request (#124) open to expand the number of binary formatting options. So feel free to express interest in that if the current binary formatting options aren't sufficient. Yep. Binary formatting can be changed in preferences: If you need something more sophisticated then this is duplicate of #124 indeed.
gharchive/issue
2016-02-02T01:48:23
2025-04-01T06:45:45.994247
{ "authors": [ "crisarceramos", "r-richmond", "serge-rider" ], "repo": "serge-rider/dbeaver", "url": "https://github.com/serge-rider/dbeaver/issues/182", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
382904478
Pythran points to the wrong line on silly error When compiling this: import numpy as np # pythran export test(float) def test(x): return 2*x # pythran export test3(float,) # <- This is where the error is def test3(x): return 2*x Pythran reports CRITICAL I am in trouble. Your input file does not seem to match Pythran's constraints... E: Invalid Pythran spec near ')' (line 5) But the error is on line 7, not 5. On a much larger program this can be very confusing because the real error is nowhere near where pythran says it is. Thxs for the report! I can reproduce on my Linux box, some homework for tonight :-) Got it. The short story is: never trust a regular expression Should be fixed by #1107, thanks again for the bug report! Got it. The short story is: never trust a regular expression Very untrustworthy they are! Fixed in #1107
gharchive/issue
2018-11-21T00:28:01
2025-04-01T06:45:45.997890
{ "authors": [ "jeanlaroche", "serge-sans-paille" ], "repo": "serge-sans-paille/pythran", "url": "https://github.com/serge-sans-paille/pythran/issues/1102", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
108633311
Very high CPU usage Hi, I've used hanfire.mongo in a little project before and it works great. Thank you for the hard work. I recently tried to use it again and I get a constant 100% cpu usage even though there are no processes running. HangFire.Mongo version=0.2.2 mongocsharpdriver version=2.0.1 It seems that if I downgrade mongo and hangfire.mongo (no other changes). Then the problem goes away, mongocsharpdriver version=1.9.2 HangFire.Mongo version=0.2.1 If you have some time could you please see if there is an easy fix to the issue. Thanks Rolf Thank you for your kind of word. You are right. I noticed the same behavior by myself. The reason is mongoscharpdriver migrated to async API while HangFire still run everything synchronously. I used AsyncHelper to wrap async calls, but seems it increases CPU utilization. I will take a look when's possible to do with it when will have a spare minute and will release the fix. Hi Sergeyzwezdin, Could you please have a look at it now? I really need this to be solve, and I cannot downgrade to previous version because I'm using mongoscharpdriver 2.x in many other places of my solution. Thank you very much in advance. @Thieund I would suggest to use these options as a temprorary solution. Just pass your MongoStorageOptions object as third parameter for GlobalConfiguration.Configuration.UseMongoStorage. Thank you for your quick response, Sergeyzwezdin. I'll do that for now. For now I will keep using the older version. Let us know when you have made the update. Thanks +1 Stuck with SQL Server until fixed :) Fixed in v.0.2.3, will be released soon. Did migrating to the Mongo csharp 2.1 driver help to solve this issue or was it just changing how often the polling occurs? I am curious as I have a site which is having high cpu utilization using the 2.0.1 driver and wondering if that could be part of the problem as this mentions that simply upgrading from 1.9 to 2.0 caused high cpu. @runxc1 it isn't reated to mongo driver. In previous version, after I upgraded Hangfire version I forgot to initialize some parameters. As a result loop interated quite often, which made high CPU utilization :-) Fix for that described here - https://github.com/sergeyzwezdin/Hangfire.Mongo/commit/1586862b7059f6448f11ab3895d6507271ab7a04 But actually I checked mostly everything in code to find root cause. Thank you to profilers :-)
gharchive/issue
2015-09-28T10:51:19
2025-04-01T06:45:46.010826
{ "authors": [ "Thieund", "bluee", "rolfwessels", "runxc1", "sergeyzwezdin" ], "repo": "sergeyzwezdin/Hangfire.Mongo", "url": "https://github.com/sergeyzwezdin/Hangfire.Mongo/issues/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1065446318
Avalonia support Will there be any support for Avalonia in the near future? I'd certainly welcome a pull-request that adds support for Avalonia, if you or anyone in the community wants to give it a shot. Avalonia is not something I use in my day-to-day, so I personally don't have plans to work on this in the short-term, but would be happy to collaborate with someone that wants to take this forward. @Whatareyoulaughingat - Is your company willing to sponsor an avalonia logger? Price would be in the $1-3k range. If so, I'm interested. Just wondering if this supports Avalonia yet? @zflemingg1 No support for Avalonia yet. PR's welcome Serilog.Sinks.Observable seems to work quite well in my protoype Avalonia v11 app. IObservable.Subscribe in the WriteTo.Observers sink sends events to a DynamicData.SourceCache. SourceCache.Connect can sync the sequence of events to a ReadOnlyObservableCollection. A DataGrid can bind to the ReadOnlyObservableCollection.
gharchive/issue
2021-11-28T21:53:17
2025-04-01T06:45:46.039987
{ "authors": [ "TonyValenti", "augustoproiete", "moon6969", "waylaa", "zflemingg1" ], "repo": "serilog-contrib/serilog-sinks-richtextbox", "url": "https://github.com/serilog-contrib/serilog-sinks-richtextbox/issues/27", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1935734748
[Bug]: Google Calendar Infers Timezone for Floating Events Bug Description Event from the ICS file DTSTART:20231009T090000 DTEND:20231015T160000 SUMMARY:MHN: Rathian & Pink Rathian Special Appearance DESCRIPTION:Rathian and Pink Rathian will appear more frequently than usual , especially during the weekend. Pink Rathian will appear in low numbers in Forest Habitats for all hunters HR11 and above. While appearance rat es are boosted, Rathian will appear more frequently than usual and Pink Rathian will also be easier to find. END:VEVENT Not the lazy of a Z at the end of the datetime (this implies floating time). When utilizing floating times, Google Calendar infers a timezone for the event (which seems to be UTC). This doesn't work for MHN events as they're timezone agnostic. If the event is 1-4pm, it's 1-4pm in every single timezone which means a fixed point in time isn't possible. At this point the only solution I can think of is to have timezone specific subscriptions, but that only works if you never leave your timezone. Retaining the floating time would be much better. Expected Behavior Google Calendar should show events at the proper time, similar to how Apple Calendar does. Steps to Reproduce (Optional) Subscribe to the calendar feed via Google Calendar Verify the MHN event times are incorrect The more I look in to this, the less I think a timezone will fix things. In order to have an event that starts at the same time regardless of timezone, I'd either need to use floating times (which I do already) or create a separate event per timezone. Taking Eastern, Central and Pacific time, I'd need 3 different calendar feeds with events just for that timezone, or 3 different events: BEGIN:VEVENT UID:1@example.com DTSTART;TZID=America/New_York:20230101T130000 DTEND;TZID=America/New_York:20230101T160000 SUMMARY:Event for Eastern Time END:VEVENT BEGIN:VEVENT UID:2@example.com DTSTART;TZID=America/Chicago:20230101T130000 DTEND;TZID=America/Chicago:20230101T160000 SUMMARY:Event for Central Time END:VEVENT BEGIN:VEVENT UID:3@example.com DTSTART;TZID=America/Los_Angeles:20230101T130000 DTEND;TZID=America/Los_Angeles:20230101T160000 SUMMARY:Event for Pacific Time END:VEVENT Which isn't really a solve at all. Short of creating a web server or some sort to infer user location and automatically assign the timezone, then send back a user specific file, I don't see this working out. To humor myself, I asked ChatGPT what it thought about the issue: Prompt: Apple Calendar on iOS works with floating timezones, but microsoft outlook and google calendar appear not to. The issue is that the events I'm tracking occur for a fixed duration at the same time regardless of timezone. So if the event is 1-4pm, it's 1-4pm anywhere you are in the world. This doesn't seem to work with fixed times. If I was to use GMT and set the event to 1-4pm, each consumer of the ics file would alter the time and date based on their own particular offsets. Do you have any suggestions on how to handle that issues? Answer The situation you're describing is a unique one. Typically, events have a fixed time in a specific timezone, and users in different timezones see the event's local equivalent. But you want the event to appear at the exact same local time regardless of the user's timezone. Floating times in iCalendar format are indeed intended for this purpose, but as you've noticed, not every calendar software handles them correctly. Here are a few potential solutions to tackle the problem: Create Multiple ICS Files: You could generate separate ICS files for each major timezone you're targeting. This is tedious and doesn't cover every individual case, but for a broad audience, targeting the major timezones (PST, CST, EST, etc.) might be enough. Users would need to pick or be directed to the appropriate ICS file based on their location. Use an Intermediary Service: Instead of directly offering an ICS file, you could have users sign up for calendar updates on a website. Once they choose their timezone, your backend could generate personalized calendar events for them. This is a more interactive solution and requires a bit of infrastructure but offers a personalized touch. Educate & Instruct: Since there is no perfect technical solution for this because of varying client behavior, you could use the DESCRIPTION or LOCATION field of the event to remind users to adjust the time as necessary for their timezone. E.g., "This event is set for 1-4pm YOUR LOCAL TIME. Please adjust if it appears differently." Duplicate Events: This is a bit of a hack, but you could add the same event multiple times to the ICS file, once for each major timezone. This will cause the event to appear correctly for most users, but users will also see duplicates of the event at other times of the day. This is not an ideal solution, but it's an option. Feedback to Calendar App Developers: Given the nature of this problem and the inconsistencies across calendar apps, it could be valuable to submit feedback to major calendar software developers like Google and Microsoft. It's a long shot, but if enough people have this use case and raise the issue, it could lead to improved handling of floating times in future software versions. The core of the issue is the lack of consistent behavior across calendar clients, which makes it challenging to find a one-size-fits-all solution. You'll need to consider the trade-offs of each approach and decide which one aligns best with your user base and project goals. Outlook actually shows the events correctly, also.
gharchive/issue
2023-10-10T16:18:06
2025-04-01T06:45:46.057854
{ "authors": [ "seriouslysean" ], "repo": "seriouslysean/monster-hunter-now-events", "url": "https://github.com/seriouslysean/monster-hunter-now-events/issues/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2618173965
Could not find org.jetbrains.kotlin:kotlin-reflect:. Ref: https://github.com/serpro69/kotlin-faker/issues/250#issuecomment-2440920179 Changes since last rc: https://github.com/serpro69/kotlin-faker/compare/v2.0.0-rc.6...8e77afdc7ece6b265099e61f1f530699fec5777e @dees91 , I couldn't reproduce your issue on my end. I've just made a simple project in intellij w/o any dependencies but kotlin-faker, and that seems to work: https://github.com/serpro69/kotlin-faker-issue-253 Can you see if the same project works for you? And can you try to reproduce your issue there? I wonder if the missing dependencies are due to some android specifics, where you need to explicitly add kotlin-reflect or smth like that? Also, when I try to list dependencies from your repro project, I can see reflect dependency is reported as failed in a lot of places. So something is definitely going awry in the build process: ➜ ./gradlew :app:dependencies > deps.txt mp.oKtCHvux8B on  main [?] is 📦 unspecified via 🅺 v2.0.20 via ☕ v17.0.12 … ➜ cat deps.txt | grep reflect +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED | +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED | +--- org.jetbrains.kotlin:kotlin-reflect:1.6.10 +--- org.jetbrains.kotlin:kotlin-reflect:1.6.10 +--- org.jetbrains.kotlin:kotlin-reflect:1.6.10 +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED | +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED | +--- org.jetbrains.kotlin:kotlin-reflect FAILED | +--- org.jetbrains.kotlin:kotlin-reflect FAILED I get a similar error in the project https://github.com/serpro69/kotlin-faker-issue-253: kotlin-faker-issue-253:main: Could not find org.jetbrains.kotlin:kotlin-reflect:. Required by: project : > io.github.serpro69:kotlin-faker:2.1.0-SNAPSHOT:20241028.081202-24 Possible solution: - Declare repository providing the artifact, see the documentation at https://docs.gradle.org/current/userguide/declaring_repositories.html more info: $ ./gradlew dependencies > deps.txt $ cat deps.txt | grep reflect | +--- org.jetbrains.kotlin:kotlin-reflect:1.6.10 +--- org.jetbrains.kotlin:kotlin-reflect:1.6.10 +--- org.jetbrains.kotlin:kotlin-reflect:1.6.10 +--- org.jetbrains.kotlin:kotlin-reflect FAILED +--- org.jetbrains.kotlin:kotlin-reflect FAILED Thanks for the info @dees91 ! I've cleaned local caches for both maven and gradle, and I can still run the code locally. However, the failed dependency I'm also getting now in the graph. So I'll try to find the root cause of that, and hopefully that will fix this issue.
gharchive/issue
2024-10-28T11:41:23
2025-04-01T06:45:46.081510
{ "authors": [ "dees91", "serpro69" ], "repo": "serpro69/kotlin-faker", "url": "https://github.com/serpro69/kotlin-faker/issues/253", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1094413255
DEP: Drop Python 3.7 support Python 3.7 is gradually abandonned by the community. References: NEP-29 says numpy dropped support for Python 3.7 in December 2021. pandas (Drop support for Python 3.7 pandas-dev/pandas#41678) pyproj (DEP: Python 3.7 support pyproj4/pyproj#930) xarray (DEP: Drop Python 3.7 Support pydata/xarray#6138) rioxarray (DEP: Drop Python 3.7 support corteva/rioxarray#451) However, ArcGis Pro still uses Python 3.7 😓 So, drop Python 3.7 support as soon as ArcGis Pro supports Python 3.8 ? Arcgis Pro seems to be using Python 3.9 in the 3.0 release, scheduled sometimes in Q2 2022 Should happen anytime soon (in fr.). However, wait for the end of the year to settle on which ArcGis version we will be using. geopandas has dropped support for Python 3.7 for version 0.11.0 rasterio has dropped support for Python 3.7 for version 1.3.0 (and this version is needed to support Pleiades-Neo RPC...) Arcgis Pro 3.0 is available at SERTIT :)
gharchive/issue
2022-01-05T14:31:41
2025-04-01T06:45:46.087249
{ "authors": [ "remi-braun" ], "repo": "sertit/eoreader", "url": "https://github.com/sertit/eoreader/issues/18", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2086009310
🛑 [KaKi87] Analytics is down In ee65c20, [KaKi87] Analytics (https://analytics.kaki87.net) was down: HTTP code: 500 Response time: 19679 ms Resolved: [KaKi87] Analytics is back up in 8330bab after 7 minutes.
gharchive/issue
2024-01-17T11:40:35
2025-04-01T06:45:46.090323
{ "authors": [ "server-KaTys" ], "repo": "server-KaTys/status", "url": "https://github.com/server-KaTys/status/issues/243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
898992173
How to mimic the commandHook properties from the NodejsFunction? 1st, Thanks for this awesome lib, very promising idea and works flawlessly, I have a small problem when deploying Basically, I want to reproduce the following logic: // The code that defines your stack goes here new NodejsFunction(this, 'prisma', { runtime: lambda.Runtime.NODEJS_14_X, handler: 'handler', entry: lambdaFn, timeout: cdk.Duration.seconds(10), memorySize: 1024, environment: { DB_URL: process.env.DB_URL || '', }, bundling: { nodeModules: ['@prisma/client', 'prisma'], commandHooks: { beforeBundling(_inputDir: string, _outputDir: string) { return [] }, beforeInstall(_inputDir: string, outputDir: string) { return [`cp -R ../prisma ${outputDir}/`] }, afterBundling(_inputDir: string, outputDir: string) { return [ `cd ${outputDir}`, `yarn prisma generate`, `rm -rf node_modules/@prisma/engines`, `rm -rf node_modules/@prisma/client/node_modules node_modules/.bin node_modules/prisma`, ] }, }, }, }) in this procedure, we need to generate something in the node_modules and remove some of the generated sub-folders We do not have a commandHooks but a copyFiles in the bundle object, but not enough for this case, anyway we can mimic this? Or if we can skip the build step in sst deploy, then we can control what would be uploaded, currently, it will always rebuild the bundle. Saw some discussion here, https://github.com/serverless-stack/serverless-stack/pull/205 @thdxr Could you plz share how you solve the prisma-bundling problem in the end? How did define that new sst.Api() I tried many ways just not work, Thanks :) Hi @Albert-Gao, yeah I can see these command hooks being useful on sst deploy. Do you need to run them for sst start? To add a bit of context, when running sst start, functions are built and run locally. The functions are not zipped up and sent to AWS Lambda. And because the functions are run locally, some of the bundle options like copyFiles are ignored, b/c we assume the files you are going to copy exist locally, and the files only need to be copied on sst deploy. So do these command hooks need to run in sst start? No, no need to run for sst start, only for deployment, because thanks to this godlike framework, the sst start can just connect to my local db, so beautiful and elegant.... But when doing deployment, I need to assemble the proper bundle, in this case, I need that commandHook 🤝 Ah that makes sense. Let me put something together and run it by you. Appreciate the kind words! That made my day :) Thanks, take your time, not a blocker for me at all. Keep the awesome work! 😺 I am back to serverless solely because of your work. similar issue #119 @Albert-Gao sorry it took me so long to see this - gonna copy paste some code that I have and hopefully you can pick through it. Some of this might be outdated because of updates in SST but I think it's mostly relevant This is me creating the Prisma layer const layer_root = ".build/layer" const layer_node = path.join(layer_root, "nodejs") const to_copy = [ "node_modules/@prisma/client", "node_modules/@prisma/engines/migration-engine-rhel-openssl-1.0.x", "node_modules/@prisma/engines/query-engine-rhel-openssl-1.0.x", "node_modules/@prisma/engines/package.json", "node_modules/@prisma/engines/dist", "node_modules/prisma", "node_modules/.prisma", ] execSync(`rm -rf ${layer_root}`) for (let src of to_copy) { const parent = path.join(src, "../") execSync(`mkdir -p ${layer_node}/${parent}`) console.log(`Copying ${src} to ${layer_node}/${parent}`) execSync(`cp -a ${src} ${layer_node}/${parent}`) } const to_delete = [ "node_modules/prisma/query-engine-rhel-openssl-1.0.x", "node_modules/prisma/query-engine-debian-openssl-1.1.x", "node_modules/.prisma/query-engine-rhel-openssl-1.0.x", "node_modules/.prisma/query-engine-debian-openssl-1.1.x", ] for (let src of to_delete) { execSync(`rm -rf ${src}`) } const layer = new lambda.LayerVersion(stack, "Prisma", { code: lambda.Code.fromAsset(layer_root), compatibleRuntimes: [lambda.Runtime.NODEJS_14_X], }) execSync(`rm -rf ${layer_root}`) return { layers: [layer], externalModules: [ "@prisma/client", "prisma", ".prisma/client", "@prisma/engines", ], } For when I'm using the function I make sure those modules aren't bundled bundle: { externalModules: props.prisma.externalModules, } Here's an example of a custom resource to run migrations export const trigger: CloudFormationCustomResourceHandler = async ( event, _ctx ) => { if (event.RequestType === "Delete") return const result = execSync( `${process.execPath} /opt/nodejs/node_modules/prisma/build/index.js migrate deploy --preview-feature`, { env: { DATABASE_URL: Config.postgres.url(), }, } ) console.log(result.toString()) } ah, layers, then that makes sense now.... by making the prisma layers, we can gain the control of building back and we do not need to ship it with every lambda, just need to reuse the layer THANKS! @thdxr Savior! Hey @Albert-Gao, sorry it took me so long to get back to this. Did @thdxr's solution work for you? Let me know if you still need the commandHooks! @fwang Hi, mate, it works! works quite well, however, my current remote database maybe far from Amazon server, so it takes nearly 2 sec to get the response, make it impossible to do any real device testing, will just use sst local dev flow for now, until I have time to setup the whole AWS Aurora Serverless with VPC with Lambda, and a proper migration setup with Prisma..... God....just hurt to even think about it... So, put in short, might not need in the long run, since Lambda Layer will give me smaller bundle size. but might need it for the short-term, just because I do not want to touch the giant CDK setup until the prototype is nearly finished. xD FWIW I ultimately decided to let go of Prisma. I wanted the full serverless experience and ultimately Postgres was never made for that as hard as AWS has tried with serverless Aurora. All in on Dynamo now @thdxr making sense, for me, I need to save stock market information which makes DynamoDB too expensive to use, (BTW, found a DynamoDB price calculator https://twitter.com/albertgao/status/1401037182412808192) I was thinking about separate user data to DynamoDB for a faster lookup as a large cheaper in-mem DB, but just do not want to add that complexity. If DynamoDB suits your use case, that gonna be fantastic DX, for me, I am stuck. In release v0.30.0 Here's my setup with layer and bash script! Lemme know if I'm missing something :) bash script to prepare prisma layer #!/bin/bash PRISMA_LAYER_DIR=".build/layers/prisma-client" function prepare_prisma_client_lambda_layer() { echo "Cleaning up workspace ..." rm -rf $PRISMA_LAYER_DIR echo "Creating Prisma client layer ..." mkdir -p "${PRISMA_LAYER_DIR}"/nodejs/node_modules/.prisma mkdir -p "${PRISMA_LAYER_DIR}"/nodejs/node_modules/@prisma echo "Prepare Prisma Client lambda layer ..." cp -r ../../node_modules/.prisma/client "${PRISMA_LAYER_DIR}"/nodejs/node_modules/.prisma cp -r ../../node_modules/@prisma "${PRISMA_LAYER_DIR}"/nodejs/node_modules } prepare_prisma_client_lambda_layer usage within the sst.Api new sst.Api(this, `${SERVICE_NAME}Api`, { defaultFunctionProps: { srcPath: SRC_PATH, bundle: { externalModules: ['@prisma/client'], }, layers: [ new lambda.LayerVersion(this, 'PrismaClientLayer', { code: lambda.Code.fromAsset(`${SRC_PATH}/.build/layers/prisma-client`), }), ], }, httpApi: HttpApi.fromHttpApiAttributes(this, 'Api', { httpApiId, }), routes: { [`ANY /{proxy+}`]: 'src/lambda.handler', }, }); @mladibejn I use the same approach, but if I share the API between other stacks, then the Prisma layer can not be updated during the deployment process cause the layer is used by "nested" stacks. More details in issue #549
gharchive/issue
2021-05-23T10:57:00
2025-04-01T06:45:46.109724
{ "authors": [ "Albert-Gao", "fwang", "mladibejn", "pabacham", "thdxr" ], "repo": "serverless-stack/serverless-stack", "url": "https://github.com/serverless-stack/serverless-stack/issues/383", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
923601430
Review doc for StaticSite Added 2 examples for overriding defaultBehavior and using Lambda@Edge with explanations on how Lambda@Edge works. Reivew the doc changes: https://github.com/serverless-stack/serverless-stack/pull/463 Updated https://github.com/serverless-stack/serverless-stack/commit/42f19a689db973c2991409853229a1c39df2da18
gharchive/issue
2021-06-17T07:58:19
2025-04-01T06:45:46.112420
{ "authors": [ "fwang", "jayair" ], "repo": "serverless-stack/serverless-stack", "url": "https://github.com/serverless-stack/serverless-stack/issues/464", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
988061657
WebSocketApi CORS error I'm getting a CORS error when I try to use the WebSocketApi with my frontend running on the my local host. On an ApiGatewayApi, I can pass in cors: true to enable CORS, but I can't find anything similar for websockets. Am I missing something? Hey @tysoncadenhead, sorry for the late reply. Please correct me if I'm wrong, I think CORS doesn't apply to WebSocket. You're right! Sorry I should have closed this out. The error was from me trying to use socket.io with API Gateway and they apparently don't play nice. I switched to the native WebSocket API and the CORS error went away.
gharchive/issue
2021-09-03T19:51:06
2025-04-01T06:45:46.114362
{ "authors": [ "fwang", "tysoncadenhead" ], "repo": "serverless-stack/serverless-stack", "url": "https://github.com/serverless-stack/serverless-stack/issues/782", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1535178290
Deployed Frontend not local Frontend folder Hi guys, I've gone through the full tutorial and everything seems to work great. However, the frontend doesn't seem to deploy to AWS. Instead, when I go to the SiteUrl, I see the SST logo on a turqoise-ish background with some text. I'm assuming I must have missed some step, somewhere, but can't find what that is. Many thanks That's what it looks like when you are working locally. We deploy a placeholder so you can run your frontend through vite on your machine. You'll need to do a sst deploy at the end, like we do here: https://docs.sst.dev/learn/deploy-to-prod#deploy-with-the-cli Thanks. It seemed strange that a placeholder would be uploaded when running locally, and I guess that's what threw me off.
gharchive/issue
2023-01-16T16:14:06
2025-04-01T06:45:46.116931
{ "authors": [ "Lazarus404", "jayair" ], "repo": "serverless-stack/sst", "url": "https://github.com/serverless-stack/sst/issues/2339", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1580839162
Python: Cannot deploy v2 (EINVAL) Description I upgraded our stack to V2 but now I'm running into an error with sst build. Curious, I tried using a blank python project and ran into the same problem. I think it's a bug. Error: ❯ npm run build > sst-py@0.0.0 build > sst build Error: Invalid src or dest: cp returned EINVAL (cannot copy /Users/mitch/<dir>/sst-py to a subdirectory of self /Users/mitch/<dir>/sst-py/.sst/artifacts/c8e724473dea4ba7e3c24ed8758fa55c990c9fe7a6) /Users/mitch/<dir>/sst-py/.sst/artifacts/c8e724473dea4ba7e3c24ed8758fa55c990c9fe7a6 Trace: SystemError [ERR_FS_CP_EINVAL]: Invalid src or dest: cp returned EINVAL (cannot copy /Users/mitch/<dir>/sst-py to a subdirectory of self /Users/mitch/<dir>/sst-py/.sst/artifacts/c8e724473dea4ba7e3c24ed8758fa55c990c9fe7a6) /Users/mitch/<dir>/sst-py/.sst/artifacts/c8e724473dea4ba7e3c24ed8758fa55c990c9fe7a6 at __node_internal_captureLargerStackTrace (node:internal/errors:478:5) at new SystemError (node:internal/errors:238:5) at new NodeError (node:internal/errors:349:7) at checkPaths (node:internal/fs/cp/cp:109:11) at async cpFn (node:internal/fs/cp/cp:65:17) at async Object.build (file:///Users/mitch/<dir>/sst-py/node_modules/sst/runtime/handlers/python.js:79:13) at async task (file:///Users/mitch/<dir>/sst-py/node_modules/sst/runtime/handlers.js:34:31) at async Object.build (file:///Users/mitch/<dir>/sst-py/node_modules/sst/runtime/handlers.js:78:23) at async file:///Users/mitch/<dir>/sst-py/node_modules/sst/constructs/Function.js:188:32 at process.<anonymous> (file:///Users/mitch/<dir>/sst-py/node_modules/sst/cli/sst.js:58:17) at process.emit (node:events:525:35) at process.emit (node:domain:489:12) at process._fatalException (node:internal/process/execution:149:25) at processPromiseRejections (node:internal/process/promises:279:13) at processTicksAndRejections (node:internal/process/task_queues:97:32) Reproduction steps Have any function with a python3.* rutime configured Run sst build or sst deploy Get this error I think the issue is here: https://github.com/serverless-stack/sst/blob/2e897bc8beb343a70b9e3911e83e67d758e93fbd/packages/sst/src/runtime/handlers/python.ts#L88 This seems to fix it (?) await fs.cp(input.out, src + "/", { recursive: true, filter: (src) => { if (src.includes(".sst")) return false; return true; }, });
gharchive/issue
2023-02-11T13:20:00
2025-04-01T06:45:46.121312
{ "authors": [ "ecumene" ], "repo": "serverless-stack/sst", "url": "https://github.com/serverless-stack/sst/issues/2469", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1675953020
🛑 geray.serverip.my.id is down In cfd807c, geray.serverip.my.id (geray.serverip.my.id) was down: HTTP code: 0 Response time: 0 ms Resolved: geray.serverip.my.id is back up in 7b4abf8.
gharchive/issue
2023-04-20T03:51:03
2025-04-01T06:45:46.144329
{ "authors": [ "serversshnet" ], "repo": "serversshnet/server-status", "url": "https://github.com/serversshnet/server-status/issues/1667", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1726754134
🛑 sg2socks.serverssh.net is down In b5ba60c, sg2socks.serverssh.net (sg2socks.serverssh.net) was down: HTTP code: 0 Response time: 0 ms Resolved: sg2socks.serverssh.net is back up in 7cfb937.
gharchive/issue
2023-05-26T01:59:02
2025-04-01T06:45:46.147743
{ "authors": [ "serversshnet" ], "repo": "serversshnet/server-status", "url": "https://github.com/serversshnet/server-status/issues/1745", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1870715676
🛑 ukray.serverip.my.id is down In 5f7b3d3, ukray.serverip.my.id (ukray.serverip.my.id) was down: HTTP code: 0 Response time: 0 ms Resolved: ukray.serverip.my.id is back up in 7dd9447 after 9 hours, 39 minutes.
gharchive/issue
2023-08-29T01:10:48
2025-04-01T06:45:46.151117
{ "authors": [ "serversshnet" ], "repo": "serversshnet/server-status", "url": "https://github.com/serversshnet/server-status/issues/2218", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2138829914
🛑 id2-ray.ipservers.xyz is down In 40f0d27, id2-ray.ipservers.xyz (id2-ray.ipservers.xyz) was down: HTTP code: 0 Response time: 0 ms Resolved: id2-ray.ipservers.xyz is back up in 5da3bc6 after 28 minutes.
gharchive/issue
2024-02-16T15:18:42
2025-04-01T06:45:46.155859
{ "authors": [ "serversshnet" ], "repo": "serversshnet/uptime", "url": "https://github.com/serversshnet/uptime/issues/5148", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
231027382
Create new Github repository for Blueprint MS Task involves: create new repository push "hello world" project add license Created new repository: https://github.com/servicecatalog/blueprint-service Added simple Spring Boot project - starting tomcat on port 8090 (configurable in application.properties), displaying page with "Hello world" header (localhost:8090). Applications is currently built as executable "uber jar" with all dependencies. There is one unit test added using Junit 4 and Mockito 2. Added license file and headers.
gharchive/issue
2017-05-24T12:45:03
2025-04-01T06:45:46.159097
{ "authors": [ "grubskim" ], "repo": "servicecatalog/development", "url": "https://github.com/servicecatalog/development/issues/727", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
116094100
Não estão sendo exibidos os solicitantes No menu lateral com links para as seções: DADOS BÁSICOS SOLICITANTES ETAPAS DO SERVIÇO Agendamento Atendimento Presencial OUTRAS INFORMAÇÕES Não estão sendo exibidos os solicitantes http://104.236.231.239.xip.io/editar/servico/carlos-teste @cevmelo é uma boa idea mas está fora do escopo.
gharchive/issue
2015-11-10T12:04:41
2025-04-01T06:45:46.164136
{ "authors": [ "cevmelo", "gustavo-freitas" ], "repo": "servicosgovbr/editor-de-servicos", "url": "https://github.com/servicosgovbr/editor-de-servicos/issues/412", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
578262609
Reduce the number of crates library users need to care about Right now we have a lot of crates, and this is causing friction for potential library users. One option would be to have a super-crate that reexports everything important. But I worry this crate will grow without limit. We already have Pathfinder native API, canvas, SVG, PDF, SWF, and Lottie frontends and it's unlikely a user will care about all of them. Likewise with GPU API backends: we already have OpenGL, WebGL, and Metal backends and will possibly get Vulkan, DX11, DX12, and gfx-rs backends in the future; it's unlikely a user wants to run on all of these platforms. I think perhaps we could have the "front-end" crates—pathfinder_canvas and pathfinder_svg—reexport the functionality necessary to use them. Then you'd only need to import two crates in most cases: the front-end (pathfinder_canvas/pathfinder_svg) and a GPU backend (pathfinder_gl, etc.) Another option would be to have a front-end pathfinder_vg crate that reexports all the subcrates behind feature flags. (Unfortunately it can't be named just pathfinder because that name is already in use on crates.io.) Other suggestions and ideas are very welcome. I really like the front-end / back-end crates, because then it’s easy to know as a user explicitly which platforms you are supporting. I can also imagine the associated docs being very obvious and easy to follow: “I want canvas, and I want to support OpenGL and Metal, so I install pathfinder_X” Does reexporting add any overhead? It shouldn't add any overhead, no. I like the reexport everything behind feature flags. I would rather people depend on whatever crate they want to use than use an umbrella crate with one feature per backend. Having a big umbrella crate means that your Cargo.lock is full of things you don't actually use, which is confusing when reading the lock file (yes, sometimes you do need to do that), or when using tools such as cargo fetch (I've already seen that command download crates that I didn't need, but maybe I misread). Also, although I like the appeal for the cargo oriented user audience, I dislike the umbrella-crate proxy pattern as a Linux Vendor. Mostly, because it just creates more busywork for us, it just adds more points of failure, and, (this is the worst part): You can't treat the umbrella crate dependencies optionally, you have to handle all of them anyway (I even have to handle test dependencies, and optional test dependencies). So directly depending on what you want IME, is better. [nb: I speak on behalf of Gentoo, so my opinion may seem selective, but I also converse with debian folk and people who work on this sort of thing in a few vendors, so its not entirely unique an opinion I'm channeling here, just some of it may be more mine than theirs] I think we're going to hold off on this for now, per feedback. Closing for now, but may reopen in the future if this decision changes.
gharchive/issue
2020-03-10T00:22:49
2025-04-01T06:45:46.183712
{ "authors": [ "BurntCaramel", "kentfredric", "nox", "pcwalton", "s3bk" ], "repo": "servo/pathfinder", "url": "https://github.com/servo/pathfinder/issues/283", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
206244961
ir: Remove unused call to clang_Cursor_isAnonymous. Fixes #496 Self-reviewing, trivial enough. @bors-servo r+ :pushpin: Commit 0bbbecd has been approved by emilio @bors-servo r- Not trivial enough; Travis is failing. r? @KiChjang Forgot to remove one call in the logging code. @bors-servo r+ :pushpin: Commit e31836f has been approved by KiChjang :hourglass: Testing commit e31836f with merge 0927273... :sunny: Test successful - status-travis Approved by: KiChjang Pushing 0927273e4a5162f0723b7c15e677ec5240b00801 to master...
gharchive/pull-request
2017-02-08T16:12:06
2025-04-01T06:45:46.188102
{ "authors": [ "KiChjang", "bors-servo", "emilio" ], "repo": "servo/rust-bindgen", "url": "https://github.com/servo/rust-bindgen/pull/497", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1230282055
Re-add caching + some workflow optimizations This PR re-adds Docker cache that was temporarily removed in #42 and also introduces some optimizations in regards to QEMU setup in the GH Actions workflows. Had to revert #42, so this will have to be rebased after we figure out what went wrong and re-do cross-platform support. I'll close this and re-open after fixing cross-compile.
gharchive/pull-request
2022-05-09T21:42:09
2025-04-01T06:45:46.322034
{ "authors": [ "sethforprivacy", "theanonymousexyz" ], "repo": "sethforprivacy/simple-monerod-docker", "url": "https://github.com/sethforprivacy/simple-monerod-docker/pull/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1213274748
🛑 Blog is down In 5c476cf, Blog (https://blog.severnbronies.co.uk) was down: HTTP code: 503 Response time: 220 ms Resolved: Blog is back up in f1a4bd2.
gharchive/issue
2022-04-23T11:17:11
2025-04-01T06:45:46.349641
{ "authors": [ "querkmachine" ], "repo": "severnbronies/status", "url": "https://github.com/severnbronies/status/issues/67", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
534525013
Missing documentation for .npsrc So .npsrc was provided by this PR, https://github.com/sezna/nps/pull/198. However, there is no documentation. Good call.
gharchive/issue
2019-12-08T11:01:43
2025-04-01T06:45:46.357958
{ "authors": [ "cursedcoder", "sezna" ], "repo": "sezna/nps", "url": "https://github.com/sezna/nps/issues/215", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
53222741
No route matches [POST] "/sessions/admin_user" After I sign in (in the URL: /admin_users/sign_in ) I get this error: Routing Error No route matches [POST] "/sessions/admin_user" I have only one Deivse User: class AdminUser < ActiveRecord::Base devise :database_authenticatable, :recoverable, :rememberable, :trackable, :validatable end My config/rails_admin.rb: RailsAdmin.config do |config| ## == Devise == config.authenticate_with do warden.authenticate! scope: :admin_user end config.current_user_method(&:current_admin_user) ... end My routes.rb: Rails.application.routes.draw do devise_for :admin_users mount RailsAdmin::Engine => '/admin', as: 'rails_admin' ... end Also, if I keep active admin and I change my routes.rb to Rails.application.routes.draw do devise_for :admin_users, ActiveAdmin::Devise.config ActiveAdmin.routes(self) mount RailsAdmin::Engine => '/newadmin', as: 'rails_admin' end Everything, even the redirect back to /newadmin, works fine. I just have to login through a ActiveAdmin looking login screen. It means only the loggin in route isn't working. if the user is already loggen in, things work. I guess something might be wrong with your devise setup. If you could upload the app with minimal reproducible code, I can help you.
gharchive/issue
2015-01-01T16:09:45
2025-04-01T06:45:46.366804
{ "authors": [ "mshibuya", "vedant1811" ], "repo": "sferik/rails_admin", "url": "https://github.com/sferik/rails_admin/issues/2147", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2561705
Add spatial type We've geo-enabled a Rails 3.1 app using the RGeo gem with a PostgreSQL database and PostGIS extensions. rails_admin can't handle tables with the spatial type. I'm contributing two very simple files: A minimal lib/rails_admin/config/fields/types/spatial.rb based on the other minimal types such as float.rb lib/tasks/spatial_type.rake rake rails_admin:spatial_disable # Remove spatial field type rake rails_admin:spatial_enable # Add spatial field type for use with PostGIS The disable task removes any line in lib/rails_admin/config/fields/types/all.rb that requires spatial.rb. The enable task is more cautious, verifying that the database connector is 'PostGIS', then adding a line to lib/rails_admin/config/fields/types/all.rb requiring spatial.rb. Once enabled, rails_admin will display the Well-Known Text (WKT) form of a spatial object, e.g., POINT (-121.260685 37.917075) See http://postgis.refractions.net/documentation/manual-1.5/ch04.html#OpenGISWKBWKT The goal is to have no impact on the 99.99% of rails_admin users who have no use for PostGIS, but to make it easier for the tiny (but growing?) percent that do. It's likely I'll make additions to spatial.rb as our app develops. Thanks for your consideration, Eric Are you sure it wouldn't be a candidate for a separate gem? Why pulling it in the base code and hacking the source with a rake task? Well, Benoit, I'm not sure of the best way to accomplish this and I'd be happy to listen to your suggestions. In the week since I made the request, I myself have wondered if it wouldn't be better done as a generator, i.e., rails g rails_admin:install --with-spatial but that still puts it in the rails_admin code base. It's such a simple thing that creating a gem for it seems like overkill. Hello, I just added a template for custom fields (in the wiki), you might be interested in testing it! Hi Benoit, I will do exactly that (and close this request, because it's not the best way to approach the problem). Thanks for letting me know.
gharchive/issue
2011-12-15T00:43:46
2025-04-01T06:45:46.373096
{ "authors": [ "bbenezech", "erictheise" ], "repo": "sferik/rails_admin", "url": "https://github.com/sferik/rails_admin/issues/881", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
583591461
Fix polymorphic association field with namespaced models The current code is generating invalid jQuery selectors for models with a namespace. For instance a model Users::Account will generate a #users::account-js-options DOM ID into the polymorphic association field template, causing an invalid jQuery selectors and breaking the related javascript code. Closing in favor of #3377.
gharchive/pull-request
2020-03-18T09:31:28
2025-04-01T06:45:46.374909
{ "authors": [ "dalpo", "mshibuya" ], "repo": "sferik/rails_admin", "url": "https://github.com/sferik/rails_admin/pull/3262", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2221458428
🛑 Civitas Juzgado is down In f6124b9, Civitas Juzgado (http://civitas.catamarcaciudad.gob.ar:8080/JUZ_CATA/forms/login.jsp) was down: HTTP code: 0 Response time: 0 ms Resolved: Civitas Juzgado is back up in f60f3cf after 13 hours, 21 minutes.
gharchive/issue
2024-04-02T21:14:12
2025-04-01T06:45:46.383320
{ "authors": [ "juanml14" ], "repo": "sfvc/upptime", "url": "https://github.com/sfvc/upptime/issues/1509", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2424988679
🛑 Turnero is down In 53e6806, Turnero (https://turnos.catamarcacapital.gob.ar/) was down: HTTP code: 0 Response time: 0 ms Resolved: Turnero is back up in 002bfd8 after 4 minutes.
gharchive/issue
2024-07-23T11:40:51
2025-04-01T06:45:46.385956
{ "authors": [ "juanml14" ], "repo": "sfvc/upptime", "url": "https://github.com/sfvc/upptime/issues/1927", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126464176
Add helpers for dealing with timestamps created_at doesn't need anything from us, we can just have DEFAULT current_timestamp. updated_at requires a trigger, but we shouldn't require our users to write that. We can use the same procedure for pretty much every table: CREATE FUNCTION set_updated_at() RETURNS trigger AS $$ BEGIN IF ( NEW IS DISTINCT FROM OLD AND NEW.updated_at IS NOT DISTINCT FROM OLD.updated_at ) THEN NEW.updated_at := current_timestamp; END IF; RETURN NEW; END $$ LANGUAGE plpgsql; We should provide that so users don't have to write it repeatedly Unsure if this is too much, but I think I'd also like to have something like diesel_manage_updated_at(table_name), because I can never remember the trigger syntax. Did you want to set updated_at upon INSERT as well? We can handle that via DEFAULT NOW() Alright, and you wanted a procedure like this? CREATE OR REPLACE FUNCTION diesel_manage_updated_at(_tbl regclass) RETURNS VOID AS $$ BEGIN EXECUTE format('CREATE TRIGGER set_updated_at BEFORE INSERT OR UPDATE ON %s FOR EACH ROW EXECUTE PROCEDURE set_updated_at()', _tbl); END; $$ LANGUAGE plpgsql; I guess it should just be BEFORE UPDATE if we assume updated_at has a default. Oh yeah, had that one written prior to your response haha. Apologies if this is obvious, but where do you imagine these helpers existing? Should the updated_at/created_at -> Timestamp pairs be reserved and handled with this procedure automatically within the migration code, or exposed to the user so they can decide whether or not diesel should manage it for them? Or is there a plan to implement a migration DSL in Rust, akin to Rails (e.g. add_timestamps)? There's currently no Rust DSL, though we may add one in the future as use cases arise. I'm fine with creating this as part of the migrations code, as long as it has a sufficiently specific name. The actual code for creating these should probably live on the Connection, so we can easily adapt it for SQLite once #152 is merged. I've been looking into what exactly would it take to implement this, constrained by the SQL format of migrations. If they wanted to, a user could write valid SQL with alternating capitals and one word per line, and it'd be accepted by Pg. So implementing a check for the phrase updated_at TIMESTAMP and then parsing the table_name would most likely mean collapsing all whitespace, and then using the regex crate's capture groups. I also thought about a pseudo-DSL using SQL comments, but a good design of that basically means implementing the foundation of a DSL entirely. We shouldn't do any sort of magic around this, the user should set the trigger themselves. We just need to provide the function for them.
gharchive/issue
2016-01-13T16:55:18
2025-04-01T06:45:46.431041
{ "authors": [ "RobertMaloney", "sgrif" ], "repo": "sgrif/diesel", "url": "https://github.com/sgrif/diesel/issues/91", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
393697020
Add commitlint configuration to enable commitlint Description As I mentioned in this comment we forgot to add a commtlint configuration, which resulted in commitlint not doing anything. Why is this change required? To enable commitlint and enforce proper commit messages 😇 [x] I've read the contributing guidelines and the Code of Conduct @ericcornelissen Thanks, I forgot about it.
gharchive/pull-request
2018-12-22T19:20:37
2025-04-01T06:45:46.439537
{ "authors": [ "ericcornelissen", "sh7dm" ], "repo": "sh7dm/vue-simple-icons", "url": "https://github.com/sh7dm/vue-simple-icons/pull/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1943361540
Error: Type 'Element' is not assignable to type 'ReactNode'. in SelectContent Component Inside Select Component I recently installed the npx shadcn-ui@latest package and tried to add a SelectContent component inside the Select component. However, I encountered the following error: Type 'Element' is not assignable to type 'ReactNode' and The expected type comes from property 'children' which is declared here on type 'IntrinsicAttributes & SelectPortalProps' The error seems to be related to the children prop of the SelectContent component, and it mentions that the expected type is coming from the property 'children' on the type 'IntrinsicAttributes & SelectPortalProps'. I'm not sure how to resolve this issue. Could you provide guidance on how to fix this error and successfully use the SelectContent component within the Select component in Shadcn-UI? Additional Information: Package: npx shadcn-ui@latest Component causing the error: SelectContent inside the Select component I think this might be related to https://github.com/shadcn-ui/ui/issues/1672 Upgrading @types/react to 18.2.28 fixes it for me Upgrading @types/react to 18.2.28 fixes it for me Got this issue now and re-openning my vs-code solved it haha
gharchive/issue
2023-10-14T16:02:16
2025-04-01T06:45:46.444454
{ "authors": [ "MonaAghili", "faris-work", "farisamirmudin", "johnrackles", "leonardogrig" ], "repo": "shadcn-ui/ui", "url": "https://github.com/shadcn-ui/ui/issues/1746", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2311808089
[bug]: the Checkbox docs page manual installation section has unspecified dependency Describe the bug On this page: https://ui.shadcn.com/docs/components/checkbox , in the manual installation section, there is a dependency that should be installed - lucide-react, which is not specified on the 'packages to install' step. Affected component/components Checkbox How to reproduce Go to https://ui.shadcn.com/docs/components/checkbox Go to manual installation section See on step 1 that lucide-react is not specified as a dependency that should be installed Codesandbox/StackBlitz link No response Logs No response System Info Irrelevant Before submitting [X] I've made research efforts and searched the documentation [X] I've searched for existing issues I made a PR: https://github.com/shadcn-ui/ui/pull/3799
gharchive/issue
2024-05-23T02:29:07
2025-04-01T06:45:46.449471
{ "authors": [ "JorensM" ], "repo": "shadcn-ui/ui", "url": "https://github.com/shadcn-ui/ui/issues/3798", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1417229637
Add cache test using a graphics shader Changes: Added a new shader cache test that utilizes a graphics instead of a compute shader - This is to check that the shader is correctly split into multiple entry points in the cache, each with their own distinct entry. Closing as I have several new changes that also need to be merged before I can open an MR for integration into Falcor. The branch containing those changes was built on top of this one.
gharchive/pull-request
2022-10-20T20:28:17
2025-04-01T06:45:46.450845
{ "authors": [ "lucy96chen" ], "repo": "shader-slang/slang", "url": "https://github.com/shader-slang/slang/pull/2459", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1227895481
feat(article): 添加新增文章接口POST /api/article re # 8 re #8
gharchive/pull-request
2022-05-06T13:49:25
2025-04-01T06:45:46.451638
{ "authors": [ "shadowfish07" ], "repo": "shadowfish07/VuePressAdmin-backend", "url": "https://github.com/shadowfish07/VuePressAdmin-backend/pull/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
700591954
Add cursor: pointer to .sidebar-toggler Issue: Great thank you! @all-contributors please add @scottparkerdesign for code
gharchive/pull-request
2020-09-13T15:33:11
2025-04-01T06:45:46.490305
{ "authors": [ "scottparkerdesign", "shahednasser" ], "repo": "shahednasser/sbuttons", "url": "https://github.com/shahednasser/sbuttons/pull/110", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
272918693
RNVideoProcessingManager.swift is causing build to fail, after upgrading to Xcode 9 I've recently upgraded to Xcode 9, and after that, when building, I get failures in RNVideoProcessingManager.swift (see screenshot below). That's weird as I've been using the library for long time, it was working fine, never faced this issue before. Any idea? @shahen94 I think this maybe duplicates https://github.com/shahen94/react-native-video-processing/issues/123 @iMagdy probably you're right. What version of RN you're using ? @shahen94 0.47.2 I'm trying to upgrade RN and see if this will fix the issue. Will also try upgrading RNVideoProcessing from 1.12.5 to 1.13.0. @shahen94 upgrade seem to break other packages. Should I downgrade your lib to something lower than 1.12.5? Yes please try to downgrade this library. Upgrading react-native probably will be little bit hard rather than downgrading this library @shahen94 Do you recommend a particular version? Build succeeded with versions: react-native-video-processing: 1.12.6 react-native: 0.47.2
gharchive/issue
2017-11-10T12:34:35
2025-04-01T06:45:46.495679
{ "authors": [ "iMagdy", "shahen94" ], "repo": "shahen94/react-native-video-processing", "url": "https://github.com/shahen94/react-native-video-processing/issues/133", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1966303515
Tweets not ordered by date Hello! Is it possible to get tweets in order of date(recent to older)? I'd be interested in that too. Have you found a solution? Also facing this issue. Any progress?
gharchive/issue
2023-10-27T23:36:57
2025-04-01T06:45:46.496974
{ "authors": [ "J0shh", "Jamie1605", "JorgeRuRe" ], "repo": "shaikhsajid1111/twitter-scraper-selenium", "url": "https://github.com/shaikhsajid1111/twitter-scraper-selenium/issues/82", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1305928433
Can we find out if a device supports a particular codec? Have you read the Tutorials? Yes Have you read the FAQ and checked for duplicate open issues? Yes What version of Shaka Player are you using? 3.3.5 Please ask your question Can we find out what codecs are supported by the platform? Something like shakaPlayer.isCodecSupportedOnDevice("mp4a") You have a few options: Use web APIs directly, such as MediaSource.isTypeSupported('audio/mp4; codecs="mp4a.40.2"') Use a method in Shaka meant for debugging: await shaka.Player.probeSupport() (pass optional true argument to get it to probe DRM, but this may result in prompts on some browsers) If you want to do this to debug, but not in your live application, visit https://shaka-player-demo.appspot.com/support.html which calls probeSupport(true) and formats the output for you Please note that you shouldn't need to probe codec support in a live application. Shaka Player will negotiate with the platform at runtime to determine which tracks are playable. There is generally no need to check codec support and load different content based on the results. Instead, just have a single DASH manifest or HLS playlist containing multiple codecs, and let the player handle it. Does this help? This helps! Thanks!
gharchive/issue
2022-07-15T11:34:58
2025-04-01T06:45:46.502516
{ "authors": [ "joeyparrish", "kiran-dcunha" ], "repo": "shaka-project/shaka-player", "url": "https://github.com/shaka-project/shaka-player/issues/4355", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1850802834
How to get the list of resolution qualities? Hello Brothers, I am using Shaka player 4.0.2 to play DRM content. My App is Vue 2 + Nuxt 2 on Microsoft Edge. I wonder if Shaka player has any built-in function/method to get the list of resolution qualities of the current playing content (I mean Shaka can auto detect the list of resolution quality: for example Auto adaptive, 480p, 720p, 1080p,...). We only have 1 stream link for this content which look like: https://example.cdn.net/live_dzones/max.smil/manifest.mpd. I have searched in Internet as well as your docs but I found nothing. Thank you. Hi! It's here getVariantTracks https://shaka-player-demo.appspot.com/docs/api/shaka.Player.html#getVariantTracks It should be able to give you the resolution qualities. @WINOFFRG thanks for your rep. I see it return list of shaka.extern.Track I am confused that this case "track" mean text track/ audio track? Am I wrong? As the method name getVariantTracks you will get list of all combinations of A/V. So if a stream has 3 Video Tracks and 2 Audio Tracks this will return you 6 combinations. To get the language and resolution or meta for that specific track just look at the defintion of shaka.extern.Track. TL;DR It return the combination of both.
gharchive/issue
2023-08-15T02:05:48
2025-04-01T06:45:46.506879
{ "authors": [ "WINOFFRG", "vanminhquangtri" ], "repo": "shaka-project/shaka-player", "url": "https://github.com/shaka-project/shaka-player/issues/5483", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
191353271
Error in SERVER PRERENDERING - ReactOnRails is not defined Hi @justin808, My configuration is heavily based on: https://github.com/shakacode/react-webpack-rails-tutorial/ I'm currently having this issue MESSAGE ERROR in SERVER PRERENDERING Encountered error: "ReferenceError: ReactOnRails is not defined" when I try to run a react component with prerender enable. To provide a bit of context the app works fine without prerender in dev and production, works fine with prerender in dev but not in production. This is part of the error message that I'm getting: MESSAGE ERROR in SERVER PRERENDERING Encountered error: "ReferenceError: ReactOnRails is not defined" when prerendering App with props: {"model":""} js_code was: (function() { var railsContext = {"inMailer":false,"i18nLocale":"en","i18nDefaultLocale":"en","href":"---","location":"---","scheme":"http","host":"---","port":null,"pathname":"---","search":"category=most-popular","httpAcceptLanguage":"en-GB,en-US;q=0.8,en;q=0.6","serverSide":true}; var props = {"model":""}; return ReactOnRails.serverRenderReactComponent({ name: 'App', domNodeId: 'App-react-component-81b00e0f-d0b0-4b44-9255-5d318aaabc60', props: props, trace: false, railsContext: railsContext }); })() my initializer/react_on_rails.rb: ReactOnRails.configure do |config| config.server_bundle_js_file = "server-bundle.js" end my initializer/assets.rb: Rails.application.config.assets.paths << Rails.root.join("app", "assets", "webpack") Rails.application.config.assets.precompile += [ "server-bundle.js" ] my webpack.server.config: // Common webpack configuration for server bundle const webpack = require('webpack'); const path = require('path'); const devBuild = process.env.NODE_ENV !== 'production'; const nodeEnv = devBuild ? 'development' : 'production'; module.exports = { context: __dirname, entry: [ 'babel-polyfill', './app/bundles/App/startup/App' ], output: { filename: 'server-bundle.js', path: '../app/assets/webpack' }, resolve: { extensions: ['', '.js', '.jsx'], alias: { react: path.resolve('./node_modules/react'), 'react-dom': path.resolve('./node_modules/react-dom') } }, plugins: [ new webpack.DefinePlugin({ 'process.env': { NODE_ENV: JSON.stringify(nodeEnv) } }) ], module: { loaders: [ { test: /\.jsx?$/, loader: 'babel-loader', exclude: /node_modules/ } ] } }; my ./app/bundles/App/startup/App: import React from 'react'; import ReactOnRails from 'react-on-rails'; import Container from '../containers/Container'; // _railsContext is the Rails context, providing contextual information for rendering const App = (props, _railsContext) => ( <Container {...props} /> ); ReactOnRails.register({ App }); I'm currently using: react-on-rails@^6.2.1: react@^15.3.1: babel-polyfill@^6.16.0 When I run RAILS_ENV=production rake assets:precompile the public/assets folder contains the server-bundle-[digest].js and I'm rsync'ing the folder to the server, so the sprockets manifest and server-bundle exist in the server. Do you have any idea why this is failing only in production? And/or can you point me to the right direction? Thanks Found the issue. server_bundle_js_file is looking for the file inside app/assets/webpack by default. Didn't know that would happen.
gharchive/issue
2016-11-23T18:49:24
2025-04-01T06:45:46.513175
{ "authors": [ "ivoreis" ], "repo": "shakacode/react_on_rails", "url": "https://github.com/shakacode/react_on_rails/issues/619", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
227725462
Referer prop is overwritten when the value contains '/users/' In my Rails view I have: <%= react_component('ReporterApp', props: @init_props, prerender: false) %> @init_props has a value 'referer'. In certain cases, like when referer contains /users/, the referer prop in the react app contains my rails home page, instead of the value from @init_props. Is there a setting I can use to stop this behavior? @figandbones Please create a simple repro app based on the generator. This sounds like a bug on your side. If you want personalized attention for your app, please consider our Coaching Plan.
gharchive/issue
2017-05-10T15:45:58
2025-04-01T06:45:46.516225
{ "authors": [ "figandbones", "justin808" ], "repo": "shakacode/react_on_rails", "url": "https://github.com/shakacode/react_on_rails/issues/837", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1679421874
🛑 Shakedown Social - Uploads is down In a0ae1b1, Shakedown Social - Uploads (https://files.shakedown.social/accounts/avatars/109/357/389/718/265/462/original/dd80d2c95e48cfd2.jpeg) was down: HTTP code: 403 Response time: 367 ms Resolved: Shakedown Social - Uploads is back up in 53c7b97.
gharchive/issue
2023-04-22T08:41:21
2025-04-01T06:45:46.519128
{ "authors": [ "clifff" ], "repo": "shakedown-social/upptime", "url": "https://github.com/shakedown-social/upptime/issues/1367", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2034141379
🛑 Shakedown Social - Uploads is down In 32f9645, Shakedown Social - Uploads (https://files.shakedown.social/accounts/avatars/109/357/389/718/265/462/original/c83dff7e175cde04.png) was down: HTTP code: 404 Response time: 35 ms Resolved: Shakedown Social - Uploads is back up in ad5e36e after 2 minutes.
gharchive/issue
2023-12-10T00:03:28
2025-04-01T06:45:46.522632
{ "authors": [ "clifff" ], "repo": "shakedown-social/upptime", "url": "https://github.com/shakedown-social/upptime/issues/6071", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1671729408
🛑 Shakedown Social - Uploads is down In 9cae2cb, Shakedown Social - Uploads (https://files.shakedown.social/accounts/avatars/109/357/389/718/265/462/original/dd80d2c95e48cfd2.jpeg) was down: HTTP code: 403 Response time: 352 ms Resolved: Shakedown Social - Uploads is back up in 7c0e78e.
gharchive/issue
2023-04-17T18:18:25
2025-04-01T06:45:46.525464
{ "authors": [ "clifff" ], "repo": "shakedown-social/upptime", "url": "https://github.com/shakedown-social/upptime/issues/781", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
792863231
Adjusting center on a fixture may not be working as intended. Disclaimer: I'm fairly new to Box2D so my knowledge of the engine is based on quite a bit of reading on iForce2D I'm trying to attach a sensor to a body by offsetting the sensors Center in a Fixture as shown in this Box2D guide however when doing so the center-offset is completely ignored and new objects are added to the center of an object, creating something that looks like this: Link to Testbed Sourcecode planck.testbed('GroundSensor', function (testbed) { var Edge = planck.Edge, Box = planck.Box, Vec2 = planck.Vec2, Circle = planck.Circle, RevoluteJoint = planck.RevoluteJoint; var world = new planck.World(Vec2(0, -9.82)); let edge = world.createBody(); edge.createFixture(Edge(Vec2(-20, 0), Vec2(20, 0))); let body = world.createDynamicBody(); body.createFixture(Box(1, 2)); body.setFixedRotation(true); body.setPosition(Vec2(0, 10)); let sensorFixture = body.createFixture(Box(0.6, 0.3), Vec2(0, -2), 0); sensorFixture.setSensor(true); sensorFixture.setUserData('player');https://i.imgur.com/7M8afG6.png return world; }); The way I've found to combat this is using joints, but I feel like this isn't the right way. Link to Testbed Sourcecode planck.testbed('GroundSensorJoint', function (testbed) { var Edge = planck.Edge, Box = planck.Box, Vec2 = planck.Vec2, Circle = planck.Circle, RevoluteJoint = planck.RevoluteJoint; var world = new planck.World(Vec2(0, -9.82)); let edge = world.createBody(); edge.createFixture(Edge(Vec2(-20, 0), Vec2(20, 0))); let body = world.createDynamicBody(); body.createFixture(Box(1, 2)); body.setFixedRotation(true); body.setPosition(Vec2(0, 10)); let sensor = world.createDynamicBody(); let sensorFixture = sensor.createFixture(Box(0.6, 0.3)); sensorFixture.setSensor(true); sensorFixture.setUserData('player'); sensor.setPosition(Vec2(0, 8)); let joint = RevoluteJoint({}, sensor, body, Vec2(0, 8)); world.createJoint(joint); return world; }); Fixtures do not have geometry and you need to position shape, so here is the correct code: body.createFixture(Box(0.6, 0.3, Vec2(0, -2), 0)) Note that I'm passing location and angle to Box instead of createFixture. Fixtures do not have geometry and you need to position shape, so here is the correct code: body.createFixture(Box(0.6, 0.3, Vec2(0, -2), 0)) Note that I'm passing location and angle to Box instead of createFixture. By the way, there is a Discord channel which is more active than here, if you need faster response please feel free to use it. Also you are making great examples, they may help others learn something new with you. By the way, there is a Discord channel which is more active than here, if you need faster response please feel free to use it. Also you are making great examples, they may help others learn something new with you.
gharchive/issue
2021-01-24T18:36:07
2025-04-01T06:45:46.537523
{ "authors": [ "ChristianTucker", "shakiba" ], "repo": "shakiba/planck.js", "url": "https://github.com/shakiba/planck.js/issues/158", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2240458564
THORChain Lending - Token collaterals Overview See https://github.com/shapeshift/web/pull/6676: Note this is an immediate "fix the bug and make this work" fix, and is to be directly follow-up on top of this stack with handling tokens at borrow time too - which is currently unsupported, but we should be prepared for it as things will similarly fail there References and additional details https://github.com/shapeshift/web/pull/6676 Acceptance Criteria Using tokens as collateral is implemented using depositWithExpiry() Need By Date No response Screenshots/Mockups No response Estimated effort No response looks fixed but not closed. and lending going bye bye
gharchive/issue
2024-04-12T16:07:40
2025-04-01T06:45:46.675821
{ "authors": [ "gomesalexandre", "twblack88" ], "repo": "shapeshift/web", "url": "https://github.com/shapeshift/web/issues/6678", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1078823251
fix: all input works regarding the language used Description Please describe your changes Notice [x] Have you followed the guidelines in our Contributing guide? [x] Have you checked to ensure there aren't other open Pull Requests for the same update/change? [x] Do all new and existing tests pass? Does the linter pass? Pull Request Type [x] :bug: Bug fix (Non-breaking Change: Fixes an issue) [ ] :hammer_and_wrench: Chore (Non-breaking Change: Doc updates, pkg upgrades, typos, etc..) [ ] :nail_care: New Feature (Breaking/Non-breaking Change) Issue (if applicable) closes #554 Testing No special yet Screenshots (if applicable) Commenting here from our conversation in discord, but the rates still show up with incorrect localization. This wasn't specifically called out in the ticket, so will add a tip if we can resolve it. If not, we can merge this and track the issue separately. Can you please update the description and briefly explain why this fix works? Added another commit to also change the trading pair value format
gharchive/pull-request
2021-12-13T17:59:56
2025-04-01T06:45:46.681460
{ "authors": [ "0xean", "cjthompson", "lolboysg" ], "repo": "shapeshift/web", "url": "https://github.com/shapeshift/web/pull/564", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1955821499
DEV Cluster ArgoCD ELB getting deletd Every few mins ArgoCD ELB getting deleted haven't we decided to switch to port forward and URL preview feature in C9? We have bigger issues to address at this point. To answer your question on how enable external access to argoCD UX, please follow this doc for ALB and Argo: https://argo-cd.readthedocs.io/en/stable/operator-manual/ingress/#aws-application-load-balancers-albs-and-classic-elb-http-mode
gharchive/issue
2023-10-22T09:03:01
2025-04-01T06:45:46.687077
{ "authors": [ "YogeshBhatiaAWS", "shapirov103" ], "repo": "shapirov103/appmod-blueprints", "url": "https://github.com/shapirov103/appmod-blueprints/issues/26", "license": "MIT-0", "license_type": "permissive", "license_source": "github-api" }
1397298050
Won't compile for macOS projects General Information Project version: Platform/OS version: macOS 11+ IDE version: Xcode 14 Describe the bug Compiling MackdownText in a macOS project fails with the SwiftUIBackports dependency calling UIScreen which is not present under macOS. https://github.com/shaps80/SwiftUIBackports/blob/d9842f24656ff7c6116dc207b60f0b6d302a5a9a/Sources/SwiftUIBackports/Shared/ImageRenderer/Renderer.swift#L7 Steps to reproduce Create a new macOS SwiftUI project. Import MarkdownText. Add a simple `MarkdownText("<some_markdown") to your view. Expected behavior The README.md claims this is supported under macOS but the dependencies are not. Screenshots Additional context Currently looking for an alternate to MarkdownUI for supporting markdown in a macOS project to run under macOS 11 or newer (so unable to take advantage of built in markdown capability) Oh nice catch! Not sure how that slipped through. Closing as this is now resolved
gharchive/issue
2022-10-05T06:25:12
2025-04-01T06:45:46.691945
{ "authors": [ "bartreardon", "shaps80" ], "repo": "shaps80/MarkdownText", "url": "https://github.com/shaps80/MarkdownText/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2488467059
[建议]Disscusion部分Category类型调整 Type 功能建议 Bug 描述 要适应https://share121.github.io/inter_knot/的”分区“部分 Disscusion这里的标签是不是也可以按——GI,SR,ZZZ来分类 截屏或视频 No response 您确认您已经阅读了上方的注意事项,并已知此处并非讨论区。您遇到了问题需要反馈,或有功能建议需要提出。 [X] 我确认。 好的已实现,请访问新域名 inot.top
gharchive/issue
2024-08-27T06:59:30
2025-04-01T06:45:46.713768
{ "authors": [ "Pleasurecruise", "share121" ], "repo": "share121/inter-knot", "url": "https://github.com/share121/inter-knot/issues/3746", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1548215962
Unable to rename I have installed the plugin and cannot find how to rename diagram. I see it in the code but do not see where it is in the UI. Would it be possible to have it display the Name value from the General properties. Thanks, Jay Camunda Modeler system information Version: 5.6.0 Operating System: Windows 10 amd64 Plugins: Multi-diagram model Plug-in, bpmn-js Token Simulation Execution Platform: BPMN - Camunda Platform 8 Hello Jay, Unfortunately I had to disable this feature on modeler 5.x, it's a small issue that we still have to fix. Feel free to contribute if you like! Thank you, Luca
gharchive/issue
2023-01-18T22:23:00
2025-04-01T06:45:46.724524
{ "authors": [ "Ichigo85", "JaySmith" ], "repo": "sharedchains/camunda-modeler-plugin-multidiagram", "url": "https://github.com/sharedchains/camunda-modeler-plugin-multidiagram/issues/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
739213919
Oct 26 punchlist (targets for alpha version so we can go map) Needed before we can map SF and release to testing cohort: [ ] iOS app release (will need to test this out and prob make a few tweaks) [ ] S3 upload workflow, so that users can locate and access uploads [x] Data model tweaks. For each point captured / feature surveyed, include: featureID, surveyID, spanID, timestamp, IMEI. These will be included in the export as objectIDs for the regs/points. [ ] Digitizer: address export bug, add featureID to the UI, enable user to sort by label or by featureID, add spanId and position Id to CurbLR export and add all the IDs listed above ^ to points export [x] Buffer fire hydrant points on backend and pass that span to the digitizer. Radius = 4.6 metres (~15 feet). This is the most common rule in the US/Canada, though there were some outliers with a 10-ft radius. (For reference, cities checked: SF, Portland*, Seattle, NYC, DC*, Baltimore, Denver*, St Louis, Minneapolis*, Vancouver, Toronto*, Montreal*, Calgary, Cincinnati*, Austin, Houston, Miami, Raleigh, Boise, Salt Lake City, Las Vegas*, Los Angeles, Atlanta) [x] Put up a hosted version of the digitizer [ ] Create new APK with the app changes [ ] Testing: general check from start to finish, try uploading a large amount of images, check selection ability in UI when mapping the same street multiple times [x] Overhaul docs with non-technical audience in mind Some bugs I'm seeing: Digitizer can't recall images once a regulation is digitized. The moment there's a template name, the images disappear for that reg and don't come back unless I wipe the regulation data. Digitizer won't export the feed if the dataset contains an overly-long street (eg street length is 100m but I rolled 120m). These features should be getting transformed (scaled proportionally) before hitting the digitizer. They appear on the map just fine, but the whole export fails. (Note: the survey that was breaking the digitizer contained a fire hydrant. I was curious what would happen to the geometry. Maybe that contributed?). Console message: Fire hydrants aren't getting their shstLocationStart and shstLocationEnd properties updated. These should reflect the transformed length/location of the regulation. Update: I can't isolate the event that caused my export to fail. I tried: overly long street overly short street fire hydrant fire hydrant at the 0 mark on a street (in case there was a buffer issue) exporting with no regulations (in case it wasn't caused by the geometries) I'm not sure what triggered the fail or if it was a combination of factors. Survey tested: 2020-11-17T00:30:46.234Z/
gharchive/issue
2020-11-09T17:18:41
2025-04-01T06:45:46.732332
{ "authors": [ "emilyeros", "russbiggs" ], "repo": "sharedstreets/curb-wheel", "url": "https://github.com/sharedstreets/curb-wheel/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
499863941
[help]dbg with lambda As you may know, some times dbg expression may be an proc of lamda showing as below int factorial(int n) { if (dbg(n <= 1)) { return dbg(1); } else { return dbg([&](){ printf("%d\n", n); return n * factorial(n - 1); }()); } } but compiler report an error main.cpp:9:16: error: lambda expression in an unevaluated operand return dbg([&](){ so I make a compromise to make compiler happy int factorial(int n) { if (dbg(n <= 1)) { return dbg(1); } else { auto a = [&](){ printf("%d\n", n); return n * factorial(n - 1); }; return dbg(a()); } } Is there a solution to avoid the stack variable when DBG_MACRO_DISABLE is true Thank you for reporting this. Here is a smaller example to reproduce this: dbg([]() { return 42; }()); This does actually compile and work if DBG_MACRO_DISABLE is set. However, as you noticed, it does not compile if the dbg macro is enabled. I get: error: lambda-expression in unevaluated context only available with ‘-std=c++2a’ or ‘-std=gnu++2a’ The problem is that we use decltype(…) on the expression inside of dbg(…) #define dbg(...) \ dbg_macro::DebugOutput(__FILE__, __LINE__, __func__, #__VA_ARGS__) \ .print(dbg_macro::type_name<decltype(__VA_ARGS__)>(), (__VA_ARGS__)) This leads to the "in unevaluated context" error. This will actually work with C++20, but I don't know how to fix this for older versions. It would be great if we could at least generate a proxy type (dbg_macro::NoTypeAnnotation) if decltype(…) fails.
gharchive/issue
2019-09-29T06:05:11
2025-04-01T06:45:46.747963
{ "authors": [ "sharkdp", "westfly" ], "repo": "sharkdp/dbg-macro", "url": "https://github.com/sharkdp/dbg-macro/issues/62", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2583355091
[Feat]Personal profile page for blogger Description Where he can save his draft post and see old post Please assign me i can work on it with level 3 tag Screenshots No response Checklist [X] I have checked the existing issues [X] I have read the Contributing Guidelines [X] I'm a GSSoC'24-Extd contributor [X] I'm a Hacktoberfest'24 contributor [X] I am willing to work on this issue (optional) @Sawan-Kushwah go ahead @sharmavikas4 I am facing problem in project set up can you please help me @sharmavikas4 can you assign it to me @Sawan-Kushwah have you setup the app in the google developer console correctly.
gharchive/issue
2024-10-12T19:28:12
2025-04-01T06:45:46.751970
{ "authors": [ "Sawan-Kushwah", "raj-adi00", "sharmavikas4" ], "repo": "sharmavikas4/MERN_BLOG", "url": "https://github.com/sharmavikas4/MERN_BLOG/issues/32", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
182271774
Need to polish null char filled strings. DXGI Adapter and DXGI Ouput has its Description for each and the desc has its HW name. That field size is fixed in native world. Therefore, after valid chars, fill with null chars. But SharpDX do not handle these null chars currently. Then, the name string looks like this "name\0\0\0\0". I think, just mashalling the fixed size of memory. Might be, it's not expected result in SharpDX users side such as me. However, if you want preserve native clarity, please add some comments to them. I was so confused by a bug that because of the null chars. Indeed, this is a bug. String marshaling had to be refactored a long time ago (when having to provide SharpDX for Metro/Windows 8) and maybe got break at this time. Pull request are welcome, as I don't have very limited spare time dedicated to SharpDX. @xoofx I'm sorry to hear that, but don't know, how the SharpDX converting C++ code to C# in deeply. Might it works by SharpGen? Known about mapping XMLs are exist but don't know, how to use them such like syntax. Therefore, I didn't change the source code at all. Instead, just added a function to handle it. public static string RefineNullFilledString(string rawString) { System.Text.StringBuilder builder = new System.Text.StringBuilder(rawString.Length); foreach (char c in rawString) { if (c != '\0') { builder.Append(c); } else { break; } } return builder.ToString(); } If you provide some references about your converting system, I'll try fixing this to appreciate your efforts.
gharchive/issue
2016-10-11T14:00:31
2025-04-01T06:45:46.761913
{ "authors": [ "donggas90", "xoofx" ], "repo": "sharpdx/SharpDX", "url": "https://github.com/sharpdx/SharpDX/issues/806", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
335256858
Singular of Axes is Axis I'm not sure how the generation is done or propagated, so lmk if I need to do other steps to make a good pull request or if it'd be easier for someone else to do that. We are no longer auto-generating C# file from C++ headers, so you would have to modify the C# code instead. But as this is a legacy API, it would break existing applications without bump SharpDX to a major version, so I would prefer to keep it as-is for now. It's not critical
gharchive/pull-request
2018-06-25T04:35:04
2025-04-01T06:45:46.763394
{ "authors": [ "Axiverse", "xoofx" ], "repo": "sharpdx/SharpDX", "url": "https://github.com/sharpdx/SharpDX/pull/1043", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
249029572
How to batch associations load? Hi there! How can I batch an association load? Example: query { p1 : product(id: 1) { id name comments { id body } } p2 : product(id: 2) { id name comments { id body } } } I would like to see only two queries in my application in that case, one to load the products and one to load the comments. I tried to use that: resolve -> (product, args, ctx) { ctx[:dataloader][:comment_loader].load_many(product.comment_ids) } But this causes a N+1 because the comment_ids will execute one query per product and one final query to bring all comments. This is possible with graphql-batch using a loader like that: class Loaders::AssociationLoader < GraphQL::Batch::Loader def initialize(model, association) @model = model @association = association.to_s end def perform(owners) ActiveRecord::Associations::Preloader.new.preload(owners, @association) owners.each { |owner| fulfill(owner, owner.public_send(@association)) } end end This is possible because graphql-batch receives all the owners (products) on the perform method and then I can use the native AR way to preload all related comments. I didn't find a way to do a similar approach using dataloader. Any suggestions? Thanks! I figured out, would be nice to have that on README or Wiki: class Loader def initialize @loaders_model = {} @loaders_association = {} end def for(model) @loaders_model[model] ||= Dataloader.new { |ids| model.find(*ids) } end def for_association(model, association) @loaders_association[model] ||= {} @loaders_association[model][association] ||= Dataloader.new { |owners| ActiveRecord::Associations::Preloader.new.preload(owners, association) owners.map(&association) } end end
gharchive/issue
2017-08-09T13:24:33
2025-04-01T06:45:46.807105
{ "authors": [ "sobrinho" ], "repo": "sheerun/dataloader", "url": "https://github.com/sheerun/dataloader/issues/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2523877039
🛑 East CTopUp is down In 3d47d70, East CTopUp (https://east.speedpayplus.com:50506/) was down: HTTP code: 0 Response time: 0 ms Resolved: East CTopUp is back up in 25b6cca after 34 minutes.
gharchive/issue
2024-09-13T04:59:06
2025-04-01T06:45:46.813082
{ "authors": [ "shekar3110a" ], "repo": "shekar3110a/PyroMonitor", "url": "https://github.com/shekar3110a/PyroMonitor/issues/390", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
159598203
made Content-Type and Accept header for POST requests configurable Hi, I needed to have the Content-Type to be something other than the default application/x-www-form-urlencoded so I made that configurable and while I was at it, did the same with Accept. Cheers, Axel Sorry, i am travelling and will take a look on Saturday. Thank you! @AxelBehrens , capital letter options were used for options without arguments, so could you please change -A and -C to be something else. For example, -f looks like a good option for Content-type, just as in curl. And for Accept, may be -m is an ok option, or whatever you think is better. Otherwise, LGTM. Closed through #27 Hi, I'm really sorry for being ignorant regarding this PR, normally that's not my style ;-). Glad you find it useful! Cheers, Axel
gharchive/pull-request
2016-06-10T09:45:42
2025-04-01T06:45:46.816526
{ "authors": [ "AxelBehrens", "shekyan" ], "repo": "shekyan/slowhttptest", "url": "https://github.com/shekyan/slowhttptest/pull/22", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1011077725
V-model of recursive, dynamically created components containing select I am loading an object from my server that has a recursive, drill-down relationship. Similar to: { name: 'Thing 1', selectOptions: [...], relatedTo: { name: 'Thing 2', selectOptions: [...], relatedTo: { name: 'Thing 3', selectOptions: [...], } } } I'm then using a component that recursively calls itself and creates another multiselect if there's another layer of relatedTo. <relationship-select v-if="relationship.relatedTo" :relationship="relationship.relatedTo" /> The component looks like this: <template> <div> <multiselect v-model="selectValue" track-by="relationship.name" label="relationship.name" :options="relationship.selectOptions" :searchable="true" :allow-empty="true" /> <relationship-select v-if="relationship.relatedTo" :relationship="relationship.relatedTo" /> </div> </template> <script> import Multiselect from 'vue-multiselect' export default { components: { Multiselect }, props: { relationship: { type: Object, required: true } }, data () { return { selectValue: this.relationship } } } </script> If I set the v-model to the relationship slot that gets passed, I get an error (as expected) saying Unexpected mutation of "relationship" prop due to creating an anti-pattern. But if I set the v-model to a data attribute local to the component, the select options load but the selected value (the label of the select input) does not populate like this: Is there a way to fill the label/selected value of a vue-multiselect input when loading a pre-defined, recursive object nested in components? The object I was passing to v-model didn't contain the key/value vue-select was looking for.
gharchive/issue
2021-09-29T15:01:42
2025-04-01T06:45:46.822164
{ "authors": [ "JoshuaPerk" ], "repo": "shentao/vue-multiselect", "url": "https://github.com/shentao/vue-multiselect/issues/1492", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2125278775
Select doesn't working Hello, your plugin is not working. I use this "vue-multiselect": "^3.0.0-beta.3", in my package.json. Dropdown is not show, after click. Style display none, not change. Please fixed it. can confirm that display:none option is not being removed after you trying to open multiselect help pls @sergey-horoshko can you either provide a reproducible link, or a git that shows an example of this error? @mattelen https://github.com/sergey-horoshko/multiselect @mattelen https://github.com/sergey-horoshko/multiselect almost completed my replica to reproduce issue. Tnx alot Ok, so as a temp workaround, this only affects Vue 3.4.15+. I've found 3.4.14, it seems to work fine. I'll look into creating a fix for this This is an issue as a result of a Vue Core bug. It is recorded here: https://github.com/vuejs/core/issues/10294. I hope they will jump into this and fix this - please jump into this issue and make a reaction to indicate this is affecting you. In the meantime, I won't update anything and will wait for this issue to be fixed there so the only way to make plugin work - downgrade to 3.4.14 for now? well, sounds fair. Ty alot for your quick response and workaround. Saved my a** Vue has just released 3.4.17 (https://github.com/vuejs/core/blob/main/CHANGELOG.md). I've installed it on sergey-horoshko's repo, and it no longer faults. So to fix this issue, install Vue version 3.4.17+ Thanks everyone for contributing and figuring out what is going on here The issue concerns Vue version 3.2.45 also
gharchive/issue
2024-02-08T14:12:50
2025-04-01T06:45:46.829586
{ "authors": [ "LepuavkoDV", "abdouthetif", "mattelen", "romamaslennikov", "sergey-horoshko" ], "repo": "shentao/vue-multiselect", "url": "https://github.com/shentao/vue-multiselect/issues/1752", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
833315645
建议序列模型din、disn等可增加有数据集的案例 Is your feature request related to a problem? Please describe. A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] Describe the solution you'd like A clear and concise description of what you want to happen. Describe alternatives you've considered A clear and concise description of any alternative solutions or features you've considered. Additional context Add any other context or screenshots about the feature request here. https://github.com/shenweichen/DSIN
gharchive/issue
2021-03-17T01:18:02
2025-04-01T06:45:46.845129
{ "authors": [ "lonngxiang", "zanshuxun" ], "repo": "shenweichen/DeepCTR", "url": "https://github.com/shenweichen/DeepCTR/issues/342", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
181271724
Fix invalid lifetime name Not sure if this crate is abandoned or not, but it's going to be broken by an upcoming bugfix in rustc. This PR fixes the lifetime name error that wasn't caught by early versions of rustc. See https://github.com/rust-lang/rust/issues/36892 and rust-lang/rust#36894 for more details. Thanks! abandoned or not, Nope, just not under active development. I'm still investigating ways to make everything performant and still safe! Thanks! Could you also publish the updated version to crates.io so it doesn't show up as a regression again?
gharchive/pull-request
2016-10-05T21:33:20
2025-04-01T06:45:46.849850
{ "authors": [ "petrochenkov", "shepmaster" ], "repo": "shepmaster/sxd-xpath", "url": "https://github.com/shepmaster/sxd-xpath/pull/74", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
963198173
Zarinpal unverified method is not implemented Detailed description Zarinpal unverified method is not implemented Context reference to this method in zarinpal website: https://docs.zarinpal.com/paymentGateway/other/#unverified this method is not a member of the standard payment interface. okay but it can be implemented in zarinpal directly because its needed for using third party gateways because normal gateways will refund the money when the payment is not verified bit third party gateways will keep the money and we most verify it ourselves I can make a pull request if you think there is a place for this method okay, please change the Zarinpal gateway and send a pull request. we don't need to expose a new method called unverified!!! please create and private method and call it when verification fails. I close this issue due to being a long time inactive.
gharchive/issue
2021-08-07T09:52:59
2025-04-01T06:45:46.901865
{ "authors": [ "erfantkerfan", "khanzadimahdi" ], "repo": "shetabit/multipay", "url": "https://github.com/shetabit/multipay/issues/102", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1334671975
🛑 Piped Youtube - Personal is down In 9242094, Piped Youtube - Personal ($PIPED_PERSONAL) was down: HTTP code: 0 Response time: 0 ms Resolved: Piped Youtube - Personal is back up in 79de75e.
gharchive/issue
2022-08-10T13:56:42
2025-04-01T06:45:46.960323
{ "authors": [ "shimuldn" ], "repo": "shimuldn/upptime-personal", "url": "https://github.com/shimuldn/upptime-personal/issues/80", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1089387342
update member (dorsa) Dorsa's picture Hi @dzeinali , seems that you did not update the code, could you please do that as well? You just need to change your photo img path at https://github.com/shinjiwlab/shinjiwlab.github.io/blob/source/_pages/members.md
gharchive/pull-request
2021-12-27T17:49:07
2025-04-01T06:45:46.963300
{ "authors": [ "dzeinali", "ftshijt" ], "repo": "shinjiwlab/shinjiwlab.github.io", "url": "https://github.com/shinjiwlab/shinjiwlab.github.io/pull/54", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
381564395
Date cannot be cleared I tried this.set('dateValue',null); to clear the value in the datepickr but it doesnt work. During debugging i found that this line is executed but it makes no change. There is no error log./ The following is my template code. {{ember-flatpickr date=(readonly dateValue) onChange=(action "onChange") placeholder="Pick date" }} @sibijohn72 I think you meant to post this issue in ember-flatpickr, this is the Ship Shape website repo. Also, this definitely works, as we're doing it in our example app. Please look there, and if you are still experiencing issues let me know https://github.com/shipshapecode/ember-flatpickr/blob/c2e50facdcb89783adf9355bbc5d248fbe6d7ac8/tests/dummy/app/controllers/docs/components/ember-flatpickr.js#L29
gharchive/issue
2018-11-16T11:40:32
2025-04-01T06:45:46.970819
{ "authors": [ "rwwagner90", "sibijohn72" ], "repo": "shipshapecode/shipshape.io", "url": "https://github.com/shipshapecode/shipshape.io/issues/122", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1666075657
Add platform to install-controller-kind target Changes On M1/M2 macOS machines, it can happen that the KinD setup uses arm64 node images and ko wants to match the system OS and architecture. However, the base images Shipwright relies on do not support a platform like darwin/arm64. Add --platform=linux/$(GO_ARCH) to enforce that ko build Linux images with the respective current system architecture. Submitter Checklist [ ] Includes tests if functionality changed/was added [ ] Includes docs if changes are user-facing [X] Set a kind label on this PR [ ] Release notes block has been filled in, or marked NONE See the contributor guide for details on coding conventions, github and prow interactions, and the code review process. Release Notes NONE Works as long as we do not support anything else than Linux. Probably good enough. Maybe an easy alternative: you use the GO_OS variable here and substitute darwin with linux. Does that work ? It seems we already have that "substitution" of sorts. I added GO_OS in the command.
gharchive/pull-request
2023-04-13T09:39:36
2025-04-01T06:45:46.975420
{ "authors": [ "HeavyWombat" ], "repo": "shipwright-io/build", "url": "https://github.com/shipwright-io/build/pull/1262", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
898479407
SHIP Proposal Process Moving and codifying the enhancement proposal process in shipwright-io/build so that it can be shared across all Shipwright projects. Relevant community documentaiton updated with links to the SHIP guidelines. Removed references to OKD and simplified the ship template. /lgtm
gharchive/pull-request
2021-05-21T20:45:42
2025-04-01T06:45:46.977098
{ "authors": [ "adambkaplan", "sbose78" ], "repo": "shipwright-io/community", "url": "https://github.com/shipwright-io/community/pull/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
117990461
Stats are not displayed I just installed shipyard/shipyard:latest (where can I find the exact version?) and the container statistics are not displayed until I resize the browser window. The javascript console has this error many times: Invalid value for <g> attribute transform="translate(NaN,5)" what version of docker On Friday, November 20, 2015, Antoine Pultier notifications@github.com wrote: I just installed shipyard/shipyard:latest (where can I find the exact version?) and the container statistics are not displayed until I resize the browser window. The javascript console has this error many times: Invalid value for attribute transform="translate(NaN,5)" — Reply to this email directly or view it on GitHub https://github.com/shipyard/shipyard/issues/675. Docker version 1.9.0, build 76d6bc9 I have the same version. Hmm what browser? I have this problem on both Firefox and Chrome. Le 22 nov. 2015 6:57 PM, "Evan Hazlett" notifications@github.com a écrit : Hmm what browser? — Reply to this email directly or view it on GitHub https://github.com/shipyard/shipyard/issues/675#issuecomment-158778149. the same problem~ Please answer! thanks! For me also same problem Some use full info Docker Info Client: Version: 1.9.0 API version: 1.21 Go version: go1.4.2 Git commit: 76d6bc9 Built: Tue Nov 3 17:43:42 UTC 2015 OS/Arch: linux/amd64 Server: Version: 1.9.0 API version: 1.21 Go version: go1.4.2 Git commit: 76d6bc9 Built: Tue Nov 3 17:43:42 UTC 2015 OS/Arch: linux/amd64 Browsers Chrome and Firefox Problem solved. The reason:docker 1.9 stats and shipyard some incompatible。 The solution: vim /static/app/containers/stats.controller.js The modified content: addNetworkUsage(timestamp, node.networks.eth0.rx_bytes / 1048576, node.networks.eth0.tx_bytes / 1048576) success! done! Docker 1.9 with docker remote API 1.21 has a API changed. GET /containers/(id)/stats will return networking information respectively for each interface. In the shipyard/controller/static/app/containers/stats.controller.js #L186 the second and third parameter should modify to: node.networks.<interface name>.rx_bytes and node.networks.<interface name>. tx_bytes This is fixed with my PR #688 Not fixed yet, using the latest shipyard the stats are still broken even I see the changes made to /static/app/containers/stats.controller.js
gharchive/issue
2015-11-20T08:32:45
2025-04-01T06:45:46.989637
{ "authors": [ "John-Lin", "baub", "donmimo", "ehazlett", "pavankumarkatakam", "yellowiscool", "zittix", "zzfzqq123" ], "repo": "shipyard/shipyard", "url": "https://github.com/shipyard/shipyard/issues/675", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
192794615
Ping registry fails due to authorization I tried adding the Gitlab CI registry and it fails the logs are: 401 Unauthorized error saving registry: Cannot ping registry The response to server {"errors":[{"code":"UNAUTHORIZED","message":"authentication required","detail":null}]} It seems to me it should ignore the 401 error, after all it hasn't tried authorizing yet. +1 Same here +1
gharchive/issue
2016-12-01T09:32:46
2025-04-01T06:45:46.991792
{ "authors": [ "blckct", "rveede", "sbruggeman", "turnerguo" ], "repo": "shipyard/shipyard", "url": "https://github.com/shipyard/shipyard/issues/917", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1368760982
Crash When Hovering Over Tarot Cards in JEI 0.65-alpha I booted up the alpha from CurseForge to play around with. I looked up the cards in JEI, and upon hovering over one of them my game crashed leaving an exception error. The crash log should be attached to here but please let me know if you need any more info. crash-2022-09-10_20.10.29-client.txt cannot be cast to class io.github.apace100.apoli.util.PowerGrantingItem this seems like a bug coming from apace100.apoli.util.PowerGrantingItem, i'm not sure what mod it is but you might wanna check with them. i use jei in my dev environment so i can confirm it doesnt crash when hovering Alright! I'll check it out, thank you for clarifying! did a little more messing around and i think i found issue, ill see if i can fix it today. has to do with the tarot decks item and other mods that modify tooltips and stuff i believe https://github.com/shiroroku/TarotCards/commit/436615dbad8b0e6508c08d5bbc4164b879172abd newest version should fix the crash, lemme know if it doesnt
gharchive/issue
2022-09-11T00:13:39
2025-04-01T06:45:47.000968
{ "authors": [ "AceTheEldritchKing", "shiroroku" ], "repo": "shiroroku/TarotCards", "url": "https://github.com/shiroroku/TarotCards/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1044341783
Update mongodb@7.3 Build mongodb@7.3 :beers: @BrewTestBot has triggered a merge.
gharchive/pull-request
2021-11-04T04:04:26
2025-04-01T06:45:47.002203
{ "authors": [ "shivammathur" ], "repo": "shivammathur/homebrew-extensions", "url": "https://github.com/shivammathur/homebrew-extensions/pull/1250", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1118205792
Libraries not loaded for php@5.6 Using php@5.6 with brew not working more. I got multiple errors like this: dyld[12266]: Library not loaded: /usr/local/opt/icu4c/lib/libicui18n.68.dylib Referenced from: /usr/local/Cellar/php@5.6/5.6.40.reinstall/bin/php Seems I have updated versions of following libraries to version 69 and php@5.6 requires v 68. /usr/local/opt/icu4c/lib/libicuio.68.dylib /usr/local/opt/icu4c/lib/libicudata.68.dylib /usr/local/opt/icu4c/lib/libicuuc.68.dylib /usr/local/opt/icu4c/lib/libicui18n.68.dylib How to fix this? The problems started after error dyld: Library not loaded: /usr/local/opt/tidy-html5/lib/libtidy.5.dylib which I "resolved" with symlinking the existing updated path. All formulae in the tap were updated to use icu4c 69.1 in Oct last year. Run brew update and reinstall php@5.6 along with its dependencies. That should fix it. brew update brew reinstall $(brew deps shivammathur/php/php@5.6) shivammathur/php/php@5.6 I did it but theres problem with libtidy again. From log ... ==> Reinstalling shivammathur/php/php@5.6 ==> Pouring php@5.6--5.6.40.big_sur.bottle.8.tar.gz ==> /usr/local/Cellar/php@5.6/5.6.40/bin/pear config-set php_ini /usr/local/etc/php/5.6/php.ini system Last 15 lines from /Users/richardtekel/Library/Logs/Homebrew/php@5.6/post_install.01.pear: 2022-01-29 14:45:15 +0100 /usr/local/Cellar/php@5.6/5.6.40/bin/pear config-set php_ini /usr/local/etc/php/5.6/php.ini system dyld[11729]: Library not loaded: /usr/local/opt/tidy-html5/lib/libtidy.5.dylib Referenced from: /usr/local/Cellar/php@5.6/5.6.40/bin/php Reason: tried: '/usr/local/opt/tidy-html5/lib/libtidy.5.dylib' (no such file), '/usr/local/lib/libtidy.5.dylib' (no such file), '/usr/lib/libtidy.5.dylib' (no such file), '/usr/local/Cellar/tidy-html5/5.8.0/lib/libtidy.5.dylib' (no such file), '/usr/local/lib/libtidy.5.dylib' (no such file), '/usr/lib/libtidy.5.dylib' (no such file) Warning: The post-install step did not complete successfully ... Thanks. I also needed brew reinstall tidy-html5 as mentioned in #404.
gharchive/issue
2022-01-29T12:42:17
2025-04-01T06:45:47.006124
{ "authors": [ "rikiless", "shivammathur" ], "repo": "shivammathur/homebrew-php", "url": "https://github.com/shivammathur/homebrew-php/issues/765", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2367371172
Hard to read lualine inactive tabs I'm attempting to understand the hues of your palette and subtly manipulate them to better suit my needs. Yet, I've encountered another issue I'm unable to resolve. The colour scheme in lualine makes the inactive tabs hard to read. Could you guide me on how to adjust this? pushed out an update to increase the contrast a bit on those. but to answer your question: lualine is a bit of a special case because the colours for those are defined by a lualine theme. found here: https://github.com/sho-87/kanagawa-paper.nvim/blob/master/lua/lualine/themes/kanagawa-paper.lua the inactive highlights are controlled by this block. a, b, and c are your regular lualine a,b,c sections: kanagawa_paper.inactive = { a = { bg = theme.ui.bg_m3, fg = theme.ui.fg_gray }, b = { bg = theme.ui.bg_m3, fg = theme.ui.fg_gray, gui = "bold" }, c = { bg = theme.ui.bg_m3, fg = theme.ui.fg_gray }, } so if you want to override those colours, you can create a custom theme based on the included theme, but assign different highlight groups to the sections you want. this is just the same as customizing a theme as per the lualine docs. For example, this will change the bg and fg of your a sections for inactive elements: local M = { "nvim-lualine/lualine.nvim", opts = function() local kanagawa_paper = require("lualine.themes.kanagawa-paper") kanagawa_paper.inactive.a.fg = "#ff0000" kanagawa_paper.inactive.a.bg = "#ffffff" return { options = { theme = kanagawa_paper, ... } Besides the special case of lualine, heres how you would adjust colours in general... Lets say you want to change the bg colour of floating windows, which is defined in the theme file here. You have a couple of options: Option 1: retarget the theme color In your config for this plugin, you can set ui.float.bg to anything you want. That will change the colour for floating window backgrounds. Example: return { "sho-87/kanagawa-paper.nvim", opts = { colors = { theme = { ui = { float = { bg = "#ff0000", }, }, }, }, }, } Option 2: retarget the palette color This is more extreme. Notice that by default ui.float.bg is assigned to palette.sumiInk4. What you can do is redefine the value of palette.sumiInk4. This will change the value of the base colour, so any other place that uses sumiInk4 will also be changed: return { "sho-87/kanagawa-paper.nvim", opts = { transparent = false, colors = { palette = { sumiInk4 = "#ff0000", }, }, }, } Note If you use transparency mode, using either option above might not seem to do anything (especially in the case of bg colors). This is because of how transparency mode works, which just sets bg color to none. Transparency mode can be a bit finicky in general because of how different plugins do transparency Perfect! Thank you kindly for elaborative explanation :pray: Maybe next time I will be able to send a PR instead :-)
gharchive/issue
2024-06-22T00:03:21
2025-04-01T06:45:47.036413
{ "authors": [ "plague-doctor", "sho-87" ], "repo": "sho-87/kanagawa-paper.nvim", "url": "https://github.com/sho-87/kanagawa-paper.nvim/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
102366239
Fix issue 2766. Changed all 'initialize' methods to more descriptive … …names to prevent clash with Ruby. Left Tapkee methods named initialize. Changes Unknown when pulling e3d1d9e6aac2033ba593cbd0859e8eb6225e62d7 on erip:fix_2766 into ** on shogun-toolbox:develop**.
gharchive/pull-request
2015-08-21T12:14:40
2025-04-01T06:45:47.038560
{ "authors": [ "coveralls", "erip" ], "repo": "shogun-toolbox/shogun", "url": "https://github.com/shogun-toolbox/shogun/pull/2895", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
576807233
Serialize JSON dates with milliseconds 1. Why is this change necessary? All the database date fields store milliseconds, but the API transfers the dates only with seconds. It should also transfer the milliseconds. 2. What does this change do, exactly? It changes the date format of all JSON date fields from s.th. like 2020-03-03T16:19:22+00:00 to 2020-03-03T16:19:22.034+00:00 This change change is non-breaking for almost all parsers, but especially the JS call of new Date() with this date format works the same. 3. Describe each step to reproduce the issue or behaviour. 4. Please link to the relevant issues (if any). 5. Checklist [ ] I have written tests and verified that they fail without my change [x] I have squashed any insignificant commits [x] I have written or adjusted the documentation according to my changes [x] This change has comments for package types, values, functions, and non-obvious lines of code [x] I have read the contribution requirements and fulfil them. There are also classes/entities that implement their own jsonSerialize() method if they have some custom serialization logic, if we have agreed to a format to use in json representation please check that it is used consistently I fixed that. There are also classes/entities that implement their own jsonSerialize() method if they have some custom serialization logic, if we have agreed to a format to use in json representation please check that it is used consistently I fixed that. Please have a look at the following JS-Snippet. I took it from a browser console that did run on a German PC: That means the following: If a date-string passed to new Date() does not have a time zone it is interpreted with the local time zone of the browser. That means the actual timed that is parsed is dependent on the time zone of the computer that is running the browser. It would then differ from country to country. If the date-string contains the time zone, it is interpreted as the same time everywhere on the world. Fully independent of the time zone of the browser. Only in the last case the time transferred via JSON would be consistent on every browser and would match the time actual time saved on the server. Omitting the time zone in the JSON would lead to many time offset errors. Hello, thank you for creating this pull request. I have opened an issue on our Issue Tracker for you. See the issue link: https://issues.shopware.com/issues/NEXT-9608 Please use this issue to track the state of your pull request. Thanks for contributing :-) @windaishi I've no idea, why this won't auto close... @pweyck because it was not merged to the default branch (which is 6.2) but master instead
gharchive/pull-request
2020-03-06T09:34:25
2025-04-01T06:45:47.073227
{ "authors": [ "mitelg", "pweyck", "shopwareBot", "windaishi" ], "repo": "shopware/platform", "url": "https://github.com/shopware/platform/pull/623", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
267691769
Expose eventemitter as global 1. Why is this change necessary? The EventEmitter is a useful pattern already included in shopware. Expose it as a global to open it for reusability for other developers. 2. What does this change do, exactly? Create window.EventEmitter 3. Describe each step to reproduce the issue or behaviour. console.log(EventEmitter) 4. Please link to the relevant issues (if any). 5. Which documentation changes (if any) need to be made because of this PR? Possibly add it to the DevDocs 6. Checklist [x] I have squashed any insignificant commits [x] This change has comments for package types, values, functions, and non-obvious lines of code [x] I have read the contribution requirements and fulfil them. Hey @screeny05 thanks for your contribution. I really like the idea to have the event emitter exposed globally. I'm just concerned about the expose key e.g. window.EventEmitter. It was rumored that a W3C Event Emitter will be coming soon (haven't looked in the docs to verify it though). Therefore I would like to suggest that you'll expose the event emitter with the state manager which is our own domain. When there's the possibility that we can break standard APIs I would like to avoid it beforehand. Best regards Good point. Has been done. @screeny05 thank you for your contrib! Could you please add documentation to https://github.com/shopware/devdocs/blob/master/source/designers-guide/javascript-statemanager-and-pluginbase/index.md as well? Hello, thank you for creating this pull request. I have opened an issue on our Issue Tracker for you. See the issue link: https://issues.shopware.com/issues/SW-20314 Please use this issue to track the state of your pull request. thanks for your contribution @screeny05 :+1:
gharchive/pull-request
2017-10-23T14:33:16
2025-04-01T06:45:47.079265
{ "authors": [ "klarstil", "mitelg", "screeny05", "shopwareBot" ], "repo": "shopware/shopware", "url": "https://github.com/shopware/shopware/pull/1332", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
270905382
Allow to change the searchFuzzyCheck-SQL 1. Why is this change necessary? Add a new filter to allow extending of searchFuzzyCheck-SQL. This is useful in some cases to include additional fields in this search (e.g. suppliernumber) 2. What does this change do, exactly? Adds a new filter. 3. Describe each step to reproduce the issue or behaviour. 4. Please link to the relevant issues (if any). 5. Which documentation changes (if any) need to be made because of this PR? 6. Checklist [ ] I have written tests and verified that they fail without my change [x] I have squashed any insignificant commits [x] This change has comments for package types, values, functions, and non-obvious lines of code [x] I have read the contribution requirements and fulfil them. Hey @sebastianlenz, please extract the searchFuzzyCheck function into a separated service, which can be decorated or replaced. Hey @sebastianlenz, this PR was closed due to inactivity. If this change is still important to you, feel free to create a new pull request. For more information about our contribution guidelines, see https://developers.shopware.com/contributing/contribution-guideline/.
gharchive/pull-request
2017-11-03T08:25:39
2025-04-01T06:45:47.083273
{ "authors": [ "sebastianlenz", "shopwareBot" ], "repo": "shopware/shopware", "url": "https://github.com/shopware/shopware/pull/1365", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
287100692
Display no link and remove-button for modus 3 = Shopping basket discount 1. Why is this change necessary? In offcanvas: a link with articleID=0 and the unnessary remove-button are displayed 2. What does this change do, exactly? It removes both 3. Describe each step to reproduce the issue or behaviour. Apply Shopping basket discount for customer group and check offcanvas 4. Please link to the relevant issues (if any). 5. Which documentation changes (if any) need to be made because of this PR? 6. Checklist [ ] I have written tests and verified that they fail without my change [x] I have squashed any insignificant commits [ ] This change has comments for package types, values, functions, and non-obvious lines of code [x] I have read the contribution requirements and fulfil them. Hi @tinect, thank you for your contribution! 💙 We apologize it took so long to be merged! 🙏
gharchive/pull-request
2018-01-09T14:07:52
2025-04-01T06:45:47.086497
{ "authors": [ "shopwareBot", "tinect" ], "repo": "shopware/shopware", "url": "https://github.com/shopware/shopware/pull/1460", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
696753179
Add (state) filter sidebar to order list view Description The order list view can only be filtered by affiliate code and campaign code which is not enough for most users. We already implemented a sidebar-order-filter (by states) in our "orders" tab of the product detail view in our Pickware ERP Starter plugin: Hence we could implement this feature for the order list view as well. Example Take a look at the order list view (/admin/sw/order/index) and compare it to the "orders" tab of any product in our demo shop that has Pickware ERP Starter installed: https://sw6demo.pickware.de/admin There are already some issues about this: https://issues.shopware.com/issues/NEXT-10480 https://issues.shopware.com/issues/NEXT-9261 https://issues.shopware.com/issues/NEXT-7375 I just updated the issue title and screenshot to match the current state of showpare/platform. on the last boostday there was a similar topic for the product list the outcome was to build a general filter sidebar which could be used in all modules https://github.com/shopware/platform/issues/880 (for second boostday: https://github.com/shopwareBoostDay/platform/issues/55 https://github.com/shopware/platform/pull/948 to build a general filter sidebar Are you suggesting that the current components (sw-sidebar, sw-sidebar-item with some select fields) should be combined in a general "filter sidebar" component that uses a unified interface (props, events) in all list views while the actual filter elements (select fields) vary from view to view? yeah basically. like I already stated here: https://github.com/shopware/platform/issues/880#issuecomment-627400506 I'm not a frontend guy, so I don't know how easy/hard this is :smile: Nothing is impossible 🌈 🙌 Hi, thanks for the really great contribution idea. We plan to implement filters for the product, customer and order module in the near future. I just want to avoid that we put in double work here. Many greetings, Marcel Allright! I will close this issue then.
gharchive/issue
2020-09-09T12:00:06
2025-04-01T06:45:47.095482
{ "authors": [ "MarcelSchmaeing", "hanneswernery", "mitelg" ], "repo": "shopwareBoostDay/platform", "url": "https://github.com/shopwareBoostDay/platform/issues/107", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
703456021
NEXT-9023 - Fix ListingPage pagination arrows always visible #44 What does this change do, exactly? Hide prev arrows on first page and next arrows on last page in pagination.html.twig Describe each step to reproduce the issue or behaviour. Go to first page on homepage list Got to last page on homepage list Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you all sign our Contributor License Agreement before we can accept your contribution.1 out of 2 committers have signed the CLA.:white_check_mark: art4webs:x: Michał JędraszczykMichał Jędraszczyk seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2020-09-17T10:15:46
2025-04-01T06:45:47.100057
{ "authors": [ "CLAassistant", "art4webs" ], "repo": "shopwareBoostDay/platform", "url": "https://github.com/shopwareBoostDay/platform/pull/144", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
723124162
fix(docs): readTextFile / readTextFileSync throws when reading directory I believe this description was erroneous. Correct me if I'm wrong. Ups, wrong repo.
gharchive/pull-request
2020-10-16T11:06:38
2025-04-01T06:45:47.103277
{ "authors": [ "vwkd" ], "repo": "shortdiv/deno", "url": "https://github.com/shortdiv/deno/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1474644831
I get these errors after releasing the app. flutter run --release Errors E/flutter ( 5864): [ERROR:flutter/shell/common/shell.cc(93)] Dart Error: Dart_LookupLibrary: library 'package:telephony/telephony.dart' not found. E/flutter ( 5864): [ERROR:flutter/runtime/dart_isolate.cc(668)] Could not resolve main entrypoint function. E/flutter ( 5864): [ERROR:flutter/runtime/dart_isolate.cc(167)] Could not run the run main Dart entrypoint. E/flutter ( 5864): [ERROR:flutter/runtime/runtime_controller.cc(385)] Could not create root isolate. E/flutter ( 5864): [ERROR:flutter/shell/common/shell.cc(604)] Could not launch engine with configuration. My Code import 'package:telephony/telephony.dart'; import 'package:workmanager/workmanager.dart'; @pragma('vm:entry-point') Future backgroundMessageHandler(SmsMessage smsMessage) async { //Handle background message log("==== Background Messages ==== $smsMessage"); // CustomerCardService.sendEmail(smsMessage.body ?? ''); Workmanager().registerOneOffTask('Send Email', 'send-email', inputData: { 'message': smsMessage.body, }); return Future<void>.value(); } void main() { final telephony = Telephony.instance; telephony.listenIncomingSms( onNewMessage: (SmsMessage smsMessage) { CustomerCardService.sendEmail(smsMessage.body ?? ''); }, onBackgroundMessage: backgroundMessageHandler, listenInBackground: true, ); ... } Did you find any solution? I am facing same issue. I am trying solution mentioned in https://github.com/shounakmulay/Telephony/issues/165 telephony: git: url: https://github.com/definev/Telephony.git ref: develop change is adding @pragma('vm:entry-point') to telephony.dart method Not sure if and when it will merged to original package.
gharchive/issue
2022-12-04T09:34:19
2025-04-01T06:45:47.127503
{ "authors": [ "binemmanuel", "rahulmaindargi" ], "repo": "shounakmulay/Telephony", "url": "https://github.com/shounakmulay/Telephony/issues/169", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2041718777
cargo build fails when executing cargo run -r --bin starklings error[E0107]: struct takes 3 generic arguments but 2 generic arguments were supplied using rustc 1.74.1 (a28077b28 2023-12-04) cargo 1.74.1 (ecb9851af 2023-10-18) 12 | pub type Map<K, V> = indexmap::IndexMap<K, V>; | ^^^^^^^^ - - supplied 2 generic arguments | | | expected 3 generic arguments Hi @nicolasbelfis, not really seeing this happen. Could you try updating your cargo and rust?
gharchive/issue
2023-12-14T13:37:01
2025-04-01T06:45:47.131991
{ "authors": [ "nicolasbelfis", "shramee" ], "repo": "shramee/starklings-cairo1", "url": "https://github.com/shramee/starklings-cairo1/issues/180", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2565580929
Feistel Cipher A symmetric structure used in many block cipher algorithms (e.g., DES). It splits the data into two halves and applies a series of operations. Add input and output field so that people can experiment Can you please assign this issue to me? I would like to work over it. Contribute inside Ciphers folder Add how to use, how it works detailed explaination, Reference, Use case, Also add space and time complexity any updates? Yup, i am nearly done with the work, will be creating the merge request by 17th of October.
gharchive/issue
2024-10-04T06:42:57
2025-04-01T06:45:47.138771
{ "authors": [ "SanskariWolf", "shrinidhihegde3" ], "repo": "shrinidhihegde3/Snippify-Hub", "url": "https://github.com/shrinidhihegde3/Snippify-Hub/issues/64", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1674363575
Send the excel file via Email Is there a way to send the downloaded file via email? I using this in our WordPress and I just need to send it via wp_mail use actual version pls setDefaultFont('Roboto') ->setDefaultFontSize(12) ->saveAs('temp.xlsx'); wp_mail( 'sigfred.chamo@gmail.com', 'test xlsx', 'see attachments', [], 'temp.xlsx' ); $data = json_decode($_POST['array2excel'], false); \Shuchkin\SimpleXLSXGen::fromArray($data) ->setDefaultFont('Roboto') ->setDefaultFontSize(12) ->downloadAs($title . '.xlsx'); return; } ?> this is the array2excel file. I used ajax and there's an error 403. I want download it and sent it in email. without the email part its working. kindly help me may be current folder is read only, check linux rights or replace 'temp.xlsx' to get_temp_dir() . 'temp.xlsx' in saveAs and wp_mail `<?php include 'wp-load.php'; if (isset($_POST['array2excel'])) { $title = $_POST['title']; $plugin_dir = '/home/content/public_html/wp-content/plugins/worders/public/lib/simplexlsgen/autoload.php'; require_once ( $plugin_dir ); $data = json_decode($_POST['array2excel'], false); \Shuchkin\SimpleXLSXGen::fromArray($data) ->setDefaultFont('Roboto') ->setDefaultFontSize(12) ->saveAs($title . '.xlsx'); $file = get_temp_dir() . $title . '.xlsx'; wp_mail( 'sigfred.chamo@gmail.com', 'test xlsx', 'see attachments', [], $file ); return; } ?>` This is the updated code but it still returns Error: 403. `$data = [[1, 2, 3]]; \Shuchkin\SimpleXLSXGen::fromArray($data) ->setDefaultFont('Roboto') ->setDefaultFontSize(12) ->saveAs('temp.xlsx'); echo get_temp_dir();` https://contentwarehouse.net/array2excel.php I already directly put the code there. Please help me. I don't see the temp.xlsx. Where can I see the file created when it saveAs? I got it. Thanks
gharchive/pull-request
2023-04-19T07:43:20
2025-04-01T06:45:47.177669
{ "authors": [ "RakiZtahX", "shuchkin" ], "repo": "shuchkin/simplexlsxgen", "url": "https://github.com/shuchkin/simplexlsxgen/pull/111", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1654412634
ATSAM4SD16 chip ID's changed The chip ID's for the ATSAM4SD16 series are incorrect. See the datasheet, page 561. See #169.
gharchive/pull-request
2023-04-04T18:56:39
2025-04-01T06:45:47.183052
{ "authors": [ "llengel" ], "repo": "shumatech/BOSSA", "url": "https://github.com/shumatech/BOSSA/pull/170", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
524235937
ConditionをKeyword入力に対応 WHAT Context Encoderでは階層的に対話をエンコードしている. これを単純なRNNに変更. TODO 全てconfigで指定できるように. 以前のモデルも使えるようにしておく. [ ] 階層モデルの取り消し enc_cell の入力をキーワードに bi_sent_cell をなくす [ ] input floor なし対応 [ ] metaタグ の取り消し ContextEncoderとの互換性を考慮しながら実装すると むちゃくちゃ面倒臭いので、一旦無視して実装する. 以下ブランチで対応 https://github.com/shunyooo/kg-cvae-generator/tree/update-keyword-base 入力を削ると、 入力形式の変更 コードの難解さ の意味がなくなってしまう感があるので、 以下のVAE実装にConditionalを付け足す方向でやる https://github.com/shunyooo/Sentence-VAE
gharchive/issue
2019-11-18T09:20:11
2025-04-01T06:45:47.194964
{ "authors": [ "shunyooo" ], "repo": "shunyooo/kg-cvae-generator", "url": "https://github.com/shunyooo/kg-cvae-generator/issues/3", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
308293025
remove local copy of vendor/d3 Removed the local git syncd copy of d3.js since it is already a dependency in both the package.json and bower.json Updated examples to use the copy from node_modules to maintain consistency with the unit tests Noticed at work that a local copy of d3 was included in the git repo and not being used. d3.v3 and d3.v2 haven't been updated in years and are missing d3 version information. It didn't make sense to me why they were checked into the repo as d3 is a required dependency in both the package and bower json files. I removed the local vendor folder which knocked a couple hundred kb off the repo and updated the examples to reference the dependent version of the library. Not sure whether it's better to reference the node_modules/d3.js or pull for cdnjs like some of the examples do for jquery and angular. Coverage remained the same at 62.953% when pulling 42c7d038ff1f8ea027fbc6eb223cb41b502c9da0 on sebasmurphy:removeD3 into 592fa14b30b785e0e53a618c148d250724eea67c on shutterstock:master.
gharchive/pull-request
2018-03-24T19:44:14
2025-04-01T06:45:47.206870
{ "authors": [ "coveralls", "sebasmurphy" ], "repo": "shutterstock/rickshaw", "url": "https://github.com/shutterstock/rickshaw/pull/613", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1996456500
feat(cargo-shuttle): generate manpage Add the ability to generate man page for cargo-shuttle. Description of change Summary Adds the feature to generate man page for cargo-shuttle. Uses separate generate sub-commands shell and manpage to generate shell completions and man page respectively (breaking API change). No longer uses an environment variable for the output path of shell completions ((breaking API change). Details Introduces a new sub-command manpage for the generate sub-command to generate man page with clap-mangen. To verify the changes, we can run the following from the project root shuttle: cargo run --package cargo-shuttle -- generate manpage This generates the man page in the current working directory by default. We can set a custom directory by setting the OUT_DIR environment variable: This generates the man page and renders it to the standard output. You can redirect the output to a .1 man page file and read it with man: cargo run --package cargo-shuttle -- generate manpage > cargo-shuttle.1 man -l cargo-shuttle.1 Brings both shell completions and man page generation under separate generate sub-commands. For shell completions, this means now you must generate shell completions with the following command: cargo run --package cargo-shuttle -- generate shell SHELL_NAME Instead of the previous API: cargo run --package cargo-shuttle -- generate --shell SHELL_NAME The output option for shell completion no longer relies on the value of the OUTPUT environment variable. If you want a custom path, use the --output option with generate shell: cargo run --package cargo-shuttle -- generate shell -h Generate shell completions Usage: cargo-shuttle generate shell [OPTIONS] <SHELL> Arguments: <SHELL> The shell to generate shell completion for [possible values: bash, elvish, fish, powershell, zsh] Options: -o, --output <OUTPUT> Output to a file (stdout by default) --working-directory <WORKING_DIRECTORY> Specify the working directory [default: .] [aliases: wd] --name <NAME> Specify the name of the project (overrides crate name) -h, --help Print help Closes #1377. How has this been tested? (if applicable) Please see the change description above. I've added the changes and updated the PR description. Apologies for any mistakes. It works, but the manpage only contains the global args and references other manpages for each subcommand (not very useful in our case), and I can find any way to generate proper manpages for subcommands or make the default manpage contain the manuals for all subcommands 🤔 Yeah you're right; it doesn't have much :sweat_smile: . I think to have manual for the sub-commands as well, we need to iterate over them using get_subcommands. And since we're outputting to stdout, I wonder how we can do separate man pages for the sub-commands. As for dumping everything into one manual, maybe we can do this: fn generate_manpage(&self) -> Result<CommandOutcome> { let app = Command::command(); let output = std::io::stdout(); let mut output_handle = output.lock(); Man::new(app.clone()).render(&mut output_handle)?; for subcommand in app.get_subcommands().cloned() { Man::new(subcommand.clone()).render(&mut output_handle)?; if subcommand.has_subcommands() { for sb in subcommand.get_subcommands().cloned() { Man::new(sb).render(&mut output_handle)?; } } } Ok(CommandOutcome::Ok) } Just wanted to discuss the approach without bloating the commit history. :sweat_smile: I quickly tested changing the input Command to Man to be a subcommand, but the result was an empty manpage, so not sure what was happening there. If you find a way to get everything into one manual, go for it. Cool. I see that a lot of parts are repeated (title and footer for each manpage), so we could customize which parts of the manpage are printed for each section to make it more unified. They do have render_* methods for custom rendering: https://docs.rs/clap_mangen/latest/clap_mangen/struct.Man.html#method.render_title We should have less repetitions now.
gharchive/pull-request
2023-11-16T09:41:14
2025-04-01T06:45:47.217787
{ "authors": [ "jonaro00", "selectiveduplicate" ], "repo": "shuttle-hq/shuttle", "url": "https://github.com/shuttle-hq/shuttle/pull/1388", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1552368397
(WIP) support running with SML/NJ Refactoring a few things and creating the appropriate .cm files so that it's possible to use SML/NJ to run smlfmt (instead of requiring building through mlton). It's now possible to run through SML/NJ, for example: $ sml src/smlfmt.cm Standard ML of New Jersey (64-bit) v110.99.3 [built: Thu Jul 28 00:35:16 2022] ... - Main.showSMLFilePreview "test.sml"; ---- test.sml ---- val x = 5 val y = 6 -------- val it = () : unit The function Main.showSMLFilePreview is really just a demo to show that this works. We can think more about functionality to add down the line.
gharchive/pull-request
2023-01-23T01:03:33
2025-04-01T06:45:47.265397
{ "authors": [ "shwestrick" ], "repo": "shwestrick/smlfmt", "url": "https://github.com/shwestrick/smlfmt/pull/80", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2399181064
Passing external props into NodeViewRenderer i have a use case where i need some external context passed into a NodeViewRenderer instance. i think i just need to modify the SvelteNodeViewRenderer to take in an additional props object. Would be happy to submit a pr if you guys would like! thanks, Alex Would be happy to submit a pr if you guys would like! That would be great, but before that can you share the expectation as code(example) here once. I think this is now easier than ever to implement, after merging the migration to Svelte 5, as you can just pass a deep reactive object to the component.
gharchive/issue
2024-07-09T20:43:39
2025-04-01T06:45:47.309600
{ "authors": [ "rChaoz", "sibiraj-s", "thesiti92" ], "repo": "sibiraj-s/svelte-tiptap", "url": "https://github.com/sibiraj-s/svelte-tiptap/issues/51", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }