id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
628832543
Improve validator logging PR Description Introduce validator event logging to log when validators perform duties. This can be enabled with the --log-include-validator-duties-enabled option and is off by default (it would be very noisy when running a lot of validators). Improve the logging output when validators fail to perform duties. Fixed Issue(s) fixes #2019 @bgravenorst This one introduces a new CLI option. Defaults to off but useful if you want logs showing what the validators are doing (probably quite noisy if you run a lot of validators)
gharchive/pull-request
2020-06-02T01:29:45
2025-04-01T04:32:57.804690
{ "authors": [ "ajsutton" ], "repo": "PegaSysEng/teku", "url": "https://github.com/PegaSysEng/teku/pull/2024", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2354174222
Make shallow circuit demo Catalyst-Compatible Context: Convert Shallow Circuit Demo into a Catalyst-compatiable one. Description of the Change: Mainly replace jax.jit with qjit. However, it introduces errors related to functools.partial (#814 and #815). Right now I circumvent the errors by avoiding using functools.partial. Will open the other PRs to fix the issues. Related GitHub Issues: [sc-65901] Just a note that codecov/project coverage failure usually means (a) the feature branch's new feature is not covered by a new test and/or (b) the feature branch is out of sync with main branch.
gharchive/pull-request
2024-06-14T22:08:05
2025-04-01T04:32:57.818938
{ "authors": [ "paul0403", "tzunghanjuang" ], "repo": "PennyLaneAI/catalyst", "url": "https://github.com/PennyLaneAI/catalyst/pull/816", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
793599162
#560 Adds constructors with priority and fixed debug statments to use… closes #560 👍 you guys are too quick for me 👍 you guys are too quick for me
gharchive/pull-request
2021-01-25T17:56:39
2025-04-01T04:32:57.861782
{ "authors": [ "ndorin", "ngenovese11" ], "repo": "PepperDash/Essentials", "url": "https://github.com/PepperDash/Essentials/pull/561", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
144683702
S3 migration This includes some Dropzone view refactoring too. :) Current coverage is 87.94% Merging #451 into master will increase coverage by +0.01% as of f55a987 @@ master #451 diff @@ ====================================== Files 232 232 Stmts 10408 10375 -33 Branches 0 0 Methods ====================================== - Hit 9152 9124 -28 Partial 0 0 + Missed 1256 1251 -5 Review entire Coverage Diff as of f55a987 Powered by Codecov. Updated on successful CI builds. :+1: looks good to me :) p.s. refactoring stuff looks great, I hadn't realized how much redundant code was redundant
gharchive/pull-request
2016-03-30T18:40:39
2025-04-01T04:32:57.889699
{ "authors": [ "beaugunderson", "codecov-io", "madprime" ], "repo": "PersonalGenomesOrg/open-humans", "url": "https://github.com/PersonalGenomesOrg/open-humans/pull/451", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
440992798
Fixed compilation for MSVC C++17 compiles in VisualStudio 2019 there might be a lots of warnings about not using secure functions _CRT_SECURE_NO_WARNINGS defined will suppress that warning #4 Note: _CRT_SECURE_NO_WARNING is only needed for tests due to MS extensions. We checked the tests were with both /c++17 and /latest under MSVC 19.20 @PeterSommerlad By the way, is test_compile_time_warning_nodiscard_scope_guards intended to fail right now? It seems to fail in general (Not just MSVC) I beieve.
gharchive/pull-request
2019-05-07T02:12:23
2025-04-01T04:32:57.893723
{ "authors": [ "AetherShane", "JoeyGrajciar" ], "repo": "PeterSommerlad/scope17", "url": "https://github.com/PeterSommerlad/scope17/pull/6", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
1903814590
vignette code fails to execute properly I am trying to run the vignette code (the Pennsylvania Warbler one) but had a couple issues running the code as provided. First, the workflow$addStructured command failed when the data was provided as a spatial points dataframe, but worked when I changed it to an sf object. Second, when trying to actually create the model it ran for a while but then ended with the following error message: Initializing model Error in initializeModel$addBias(datasetNames = Workflow$.__enclos_env__$private$biasNames) : Dataset provided not available. In addition: Warning message: In saveRDS(object = PSDMsMOdel, file = paste0(modDirectory, "/", : 'package:stats' may not be available when loading I'm not sure why the error about the biasNames pops up as the ebird dataset name is set earlier in the code. If I run it with out the biasNames option, the model runs but then I have a different problem - namely, the models aren't retrurned properly (or don't actually run). Models <- sdmWorkflow(Workflow = workflow) lapply(Models, summary) I get a NULL object for the Models returned when I run the code above (when it actually completes a run after removing the biasNames option). Maybe I'm missing some steps in the vignette, so any help in troubleshooting would be great. Session Info: R version 4.3.0 (2023-04-21 ucrt) Platform: x86_64-w64-mingw32/x64 (64-bit) Running under: Windows 10 x64 (build 19045) Matrix products: default attached base packages: [1] stats graphics grDevices utils datasets methods base other attached packages: [1] USAboundaries_0.4.0 intSDM_2.0.0 terra_1.7-29 ggplot2_3.4.3 PointedSDMs_1.3 [6] R6_2.5.1 inlabru_2.9.0 fmesher_0.1.2 sf_1.0-12 loaded via a namespace (and not attached): [1] DBI_1.1.3 geodata_0.5-8 s2_1.1.4 remotes_2.4.2 [5] readxl_1.4.2 rlang_1.1.0 magrittr_2.0.3 e1071_1.7-13 [9] tidyterra_0.4.0 compiler_4.3.0 callr_3.7.3 vctrs_0.6.2 [13] stringr_1.5.0 rgbif_3.7.7 httpcode_0.3.0 wk_0.8.0 [17] pkgconfig_2.0.3 crayon_1.5.2 taxize_0.9.100 fastmap_1.1.1 [21] labeling_0.4.3 lwgeom_0.2-13 giscoR_0.3.5 utf8_1.2.3 [25] rmarkdown_2.21 ps_1.7.5 MatrixModels_0.5-1 INLA_23.09.09 [29] purrr_1.0.1 xfun_0.39 embarcadero_1.2.0.1003 jsonlite_1.8.4 [33] reshape_0.8.9 uuid_1.1-1 prettyunits_1.1.1 parallel_4.3.0 [37] R.devices_2.17.1 stringi_1.7.12 lubridate_1.9.2 cellranger_1.1.0 [41] stars_0.6-3 Rcpp_1.0.10 iterators_1.0.14 knitr_1.42 [45] triebeard_0.4.1 zoo_1.8-12 base64enc_0.1-3 R.utils_2.12.2 [49] Metrics_0.1.4 splines_4.3.0 Matrix_1.5-4 timechange_0.2.0 [53] tidyselect_1.2.0 rstudioapi_0.14 abind_1.4-5 yaml_2.3.7 [57] codetools_0.2-19 curl_5.0.0 processx_3.8.1 pkgbuild_1.4.0 [61] lattice_0.21-8 tibble_3.2.1 plyr_1.8.8 withr_2.5.0 [65] USAboundariesData_0.4.0 evaluate_0.20 desc_1.4.2 units_0.8-1 [69] proxy_0.4-27 xml2_1.3.3 pillar_1.9.0 whisker_0.4.1 [73] blockCV_3.1-3 KernSmooth_2.23-20 foreach_1.5.2 generics_0.1.3 [77] rprojroot_2.0.3 sp_1.6-0 munsell_0.5.0 scales_1.2.1 [81] rgdal_1.6-6 class_7.3-21 glue_1.6.2 lazyeval_0.2.2 [85] tools_4.3.0 data.table_1.14.8 dbarts_0.9-23 grid_4.3.0 [89] bold_1.2.0 tidyr_1.3.0 ape_5.7-1 urltools_1.7.3 [93] tidyverse_2.0.0 colorspace_2.1-0 nlme_3.1-162 raster_3.6-20 [97] conditionz_0.1.0 cli_3.6.1 fansi_1.0.4 dplyr_1.1.2 [101] gtable_0.3.4 R.methodsS3_1.8.2 oai_0.4.0 digest_0.6.31 [105] classInt_0.4-9 crul_1.3 farver_2.1.1 htmltools_0.5.5 [109] R.oo_1.25.0 lifecycle_1.0.3 dismo_1.3-9 httr_1.4.5 Hey, I noticed that issue not too long ago and managed to fix it. It should work if you use the github version of the package rather than the CRAN version. devtools::install_github('PhilipMostert/intSDM) Hmm, i tried it again with the new version but I'm still having issues with the model output - the output is still NULL after rerunning this morning. However, it did run with the biasNames option this time, while it didn't before. Maybe I am missing something else? Is there any particular issue that you get when running the models? There was an issue in the last line when printing the summaries, but it shouldn't have returned NULL. Have you by any chance changed the Save argument instartWorkflow? I see from your first comment that it was trying to save the output of the model. If set to TRUE then the function will try and save the output of the model to the chosen directory instead of returning the object. Well, that makes a lot of sense! The files saved properly it looks like and I can load them from the saved folders. Looks like it is all good! Thanks for your help. also just wanted to add I'm excited to try it out with my own data, looks like a really cool package!
gharchive/issue
2023-09-19T22:32:16
2025-04-01T04:32:57.957477
{ "authors": [ "PhilipMostert", "king0708" ], "repo": "PhilipMostert/intSDM", "url": "https://github.com/PhilipMostert/intSDM/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
388385160
Make sure releases are signed https://www.electron.build/code-signing once we bring that to ethereum org I can port what we have on ethereum/mist. It's basically, just set some proper env variables.
gharchive/issue
2018-12-06T20:29:25
2025-04-01T04:32:57.959498
{ "authors": [ "PhilippLgh", "evertonfraga" ], "repo": "PhilippLgh/mist-shell", "url": "https://github.com/PhilippLgh/mist-shell/issues/4", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1645858660
More data Swype will properly release their instruction fine tuning dataset real soon: https://huggingface.co/datasets/swype/instruct-102.4k And there is also: https://github.com/nomic-ai/gpt4all I forgot this one, damn https://github.com/mobarski/alpaca-libre Thank you for the reminder. We will collect and format these as soon as possible.
gharchive/issue
2023-03-29T13:50:44
2025-04-01T04:32:57.977628
{ "authors": [ "KnutJaegersberg", "PhoebusSi" ], "repo": "PhoebusSi/Alpaca-CoT", "url": "https://github.com/PhoebusSi/Alpaca-CoT/issues/15", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
532933916
FileValidator won't open in Win10. Hello, I wouldl ike to ask for a short support. I downloaded the "win-photonfilevalidator-2.0 (2)" file. Installed and started. However the program started and it is showned on the tray, but the window not appear on the screen. Do you know what I did wrong? Thank you Problem solved. i had a 2nd monitor which was not turned on. I disconnected it from the PC and the window appeared on the first monitor.
gharchive/issue
2019-12-04T21:05:14
2025-04-01T04:32:57.984715
{ "authors": [ "kenikol" ], "repo": "Photonsters/PhotonFileValidator", "url": "https://github.com/Photonsters/PhotonFileValidator/issues/41", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
666620431
Index Sid in SubPost table for fast retrieval for sub filter posts Why doesn't phuks use any indexing on foreign key field like pid in subcomment table for fast queries/retrieval of data. Thanks We do, but I've forgotten to add a migration for this after we switched from our old .sql dump to migrations for database setup We do, but I've forgotten to add a migration for this after we switched from our old .sql dump to migrations for database setup Does it index parentcid in SubPostComment table? Does it require to index parentcid of SubPostComment. Comment retrieval is using pid of SubPostComment. So does it need to index parentcid. Yes. Yes. Can you show the database table schema. I am testing in local development. I am confused what to index and what not to. Just double checked and the migrations are already creating most of the required indexes. I just pushed a migration with three more indexes that shouldn't impact everyday use a lot but should help anyway. @migrator.create_model class SubPostComment(pw.Model): cid = pw.CharField(max_length=40, primary_key=True) content = pw.TextField(null=True) lastedit = pw.DateTimeField(null=True) parentcid = pw.ForeignKeyField(backref='subpostcomment_set', column_name='parentcid', field='cid', model='self', null=True) pid = Just double checked and the migrations are already creating most of the required indexes. I just pushed a migration with three more indexes that shouldn't impact everyday use a lot but should help anyway. @migrator.create_model class SubPostComment(pw.Model): cid = pw.CharField(max_length=40, primary_key=True) parentcid = pw.ForeignKeyField(backref='subpostcomment_set', column_name='parentcid', field='cid', model='self', null=True) pid = pw.ForeignKeyField(backref='subpostcomment_set', column_name='pid', field='pid', model=migrator.orm['sub_post'], null=True) Where is the indexing for pid? Does peewee migration create indexing automatically for foreignkey? Yes
gharchive/issue
2020-07-27T22:20:49
2025-04-01T04:32:57.994099
{ "authors": [ "Polsaker", "axiomxd" ], "repo": "Phuks-co/throat", "url": "https://github.com/Phuks-co/throat/issues/136", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
298183213
Automatically load Spreadsheet Implement loading an online version of responses on startup. Probably will have to export the spreadsheet as csv or similar... TODO: Direct GDocs loading? GitHub releases? Personal server? After careful consideration I think the best idea would be to create tags on GitHub and upload curated CSV exports of the GDocs spreadsheet there. The application can then load all available tags from GitHub on startup and if it's newer than the local one it downloads the latest CSV and saves it somewhere to disk. The CSV should be loaded through some sort of helper class, to avoid code repetition in different Behaviour classes. I'll see what I can do about inputting from the CSV. I was thinking about doing semi-colon delimiters so the program wouldn't be disrupted by punctuation (Assuming we can keep our text "semi-colon free"). I'm not very familiar with how GitHub and how loading tags from it into the program would work however. Do you have any reference material I can study up on as far as that goes? @PiMaker One could take a look at this: https://github.com/octokit/octokit.net Seems like that would make interfacing with GitHub easier. I will create a sample tag and upload a CSV for testing. Semicolon-Seperated sounds good, that's still called a CSV though. Awesome, i'll take a look at that in the morning! And I am indeed @PiMaker Okay got a good idea for building this, mind assigning me to the issue so I can receive any updates from this thread? @PiMaker Alright, I have no clue how to assign you the issue, it doesn't let me from the dropdown I get... Anyway, you should receive updates to this thread anyway, because you are mentioned in it? Once you have a prototype you can create a pull request from a fork of yours, that should allow me to merge your code into the main repo. I thought about also adding in a general update functionality. Although I feel like that should be a seperate part anyway. Have you focused your prototype idea more on parsing the CSV or downloading updates from GitHub? Because if you want we could split up that task, so I can take on one or the other as well. Okay, maybe there's a way to assign myself! I'm very unfamiliar with this platform so if I get updates then it sounds good! And i'm focusing currently on the parsing the CSV to start with, if you could figure out downloading updates from GitHub so we can combine when i'm finished that would be very helpful. Sounds like a plan. For the format of the CSV, if you could somewhat base that on the spreadsheet already on Google Docs that would be very helpful, as it would minimize effort to create a new version from new entries in there. Yeah, what I did is I exported the Google Docs to a CSV file after swapping the delimiter to a semi-colon and I've been testing my code with that. I just now got it working so I can store the code as members in structs. Tomorrow (hopefully) i'll perfect that and figure out how your code works so I can start working on her reading in the behaviours The interesting part is going to be how to integrate my browser process instance detection function to avoid detecting multiple processes in one single browser instance. I think that will have to be a hardcoded edge-case for now, as long as you don't figure out some magical solution. (You'll see once you take a look at my code) Okay so I've got it down a basic thing to be able to read the information into my struct the main issue i'll have now is integrating my C++ into your C# so that I can finish it up by making the function so MonikAI can read it I apologize for entering the thread late. I started created a csv parser the first day I saw the repository but the code was messy and I did not get a chance to improve it until today but I saw that it was already being worked on. I've submitted a PR with the changes that I made but my code is in C#. Is there any reason why the parser needs to be done in C++? It seems like using unmanaged code with the core codebase might make things more complicated to change and extend in the future. @bwroberts48 Is your code posted anywhere? We can work together to convert it to C# in case my code isn't up to par with the rest of the repository (although if that's the case, I don't think you'd want my help!) I linked my parser to the ApplicationBehavior class and from the testing I did everything seems to be working fine. I will take a look at linking it to the WebBrowserBehavior next if there are no issues to resolve. On a side note, the auto updater seems to be broken for me. For some reason, whenever I run MonikAI, the auto updater immediately maxes out the 60 hourly GitHub API requests that I can make. I will investigate this issue further at a later time. @zchaudhry11 Only reason I was writing in C++ was because I don't know C# so your code makes the process a lot easier. I'm a novice to programming so I'll look at your code but I bet it's more elegant than mine I'm not a professional by any means so feel free to make any corrections if something looks problematic or ask any questions if you have any. C# is an amazing programming language and you should give it a shot since there's no better way to learn than diving head first into a new codebase! Merged the pull request from @zchaudhry11 and implemented the remaining few bits myself. Thank you all!
gharchive/issue
2018-02-19T07:34:35
2025-04-01T04:32:58.004163
{ "authors": [ "PiMaker", "bwroberts48", "zchaudhry11" ], "repo": "PiMaker/MonikAI", "url": "https://github.com/PiMaker/MonikAI/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
427291562
VideoFileRenderer I used the same implementation of the VideoFileRenderer you made, it worked but I am having the following issues: I have recorded a 3 seconds video file, its size is about 40 MegaBytes, is that normal, I tried to change the resolution of video captured so that its quality gets reduced, but this did not make any difference. I could not play the video except with external apps like MXPlayer (could not play it with the normal video player that comes embedded with the android) I wrote about that I found that it is of extension .y4m, I tried to convert it programmatically into mp4 or something playable by the native VideoView or even the video player that comes with the android but could not. This is not an issue: how to do the same with audio, i.e. what is the VideoFileRenderer for audio I really appreciate your effort in webrtc, Thanks in advance. At first, I didn't commit any actual code into this repo, all the source code commit history are the same with the official repo, the real content of this repo is the sdk/android_gradle folder, which is a collection of build scripts. Although I contribute several patches into the official webrtc repo :) Regarding your questions: if you read VideoFileRenderer's source code, you will find answers of your questions :)
gharchive/issue
2019-03-30T15:04:48
2025-04-01T04:32:58.009379
{ "authors": [ "Piasy", "cppgarbagecollector" ], "repo": "Piasy/webrtc", "url": "https://github.com/Piasy/webrtc/issues/7", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1051310930
✨ Add raw_content property to UserMessage This will remove text formatting (ex Italics, Bold, Code Blocks). This will ame it easier to deal with text. I suppose I can take a look at this, do emoji's need to removed as well? I don't think so because I think it would be good if you can search for an emoji in a cleaned string I think it should be called clean_content, because raw_content tells me nothing has changed to the content I think it should be called clean_content, because raw_content tells me nothing has changed to the content Sure! It'll make more sense for dpy devs too. Just putting my grain of salt. I personally wouldn't put the burden on cleaning the text -- removing the content aforementioned in the comment thread -- on the library but rather in something else, like user code. Okay, maybe if this comes often enough, why not make something external that is standalone. I would rather have a property or a method that will nullify effects of certain mentions like @everyone or @here, replace different mentions[^1] by their textual counterpart. For example, <#0123456789> becomes #my-channel-has-been-mentioned. If cleaning the message text will be useful at multiple places in the library, why not. If it is convenient for the end user so they don't have headaches with regex, also why not ! [^1]: Roles, channels, users, whatever is mentionable, you name it! #237 Should it also convert emojis like :eyes: to unicode?
gharchive/issue
2021-11-11T19:54:19
2025-04-01T04:32:58.036931
{ "authors": [ "Lunarmagpie", "Sly-Little-Fox", "beastmatser", "lapin-b" ], "repo": "Pincer-org/Pincer", "url": "https://github.com/Pincer-org/Pincer/issues/192", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1944737449
fix(tidb): fix script integrationtest.sh Correct start pd cmd params. Specify using store tikv for testing. /hold Currently, running with store tikv will result in the following errors. ERRO[0663] 8 tests failed ERRO[0663] run test [executor/admin] err: sql:admin checksum table cache_admin_table_with_index_test;: failed to run query "admin checksum table cache_admin_table_with_index_test;" around line 66, we need(172): admin checksum table cache_admin_table_with_index_test; Db_name Table_name Checksum_crc64_xor Total_kvs Total_bytes executor__admin cache_admin_table_with_index_test 0 2 2 but got(172): admin checksum table cache_admin_table_with_index_test; Db_name Table_name Checksum_crc64_xor Total_kvs Total_bytes executor__admin cache_admin_table_with_index_test 0 0 0 ERRO[0663] run test [executor/executor] err: sql:change pump to node_state ='paused' for node_id 'pump1';: failed to run query "change pump to node_state ='paused' for node_id 'pump1';" around line 693, we need(132): change pump to node_state ='paused' for node_id 'pump1'; Error 1105 (HY000): URL scheme must be http, https, unix, or unixs: change but got(132): change pump to node_state ='paused' for node_id 'pump1'; Error 1105 (HY000): node pump, id pump1 from etcd 127.0.0.1:2379 not found ERRO[0663] run test [expression/builtin] err: sql:select * from tb5 where cast(a as unsigned int)=0;: failed to run query "select * from tb5 where cast(a as unsigned int)=0;" around line 280, we need(156): select * from tb5 where cast(a as unsigned int)=0; a b Level Code Message Warning 1690 constant 1.844674407370955e+20 overflows bigint select * from tb5 whe but got(156): select * from tb5 where cast(a as unsigned int)=0; a b Level Code Message Warning 1690 evaluation failed: constant 184467440737095500000 overflows LongLong ERRO[0663] run test [expression/charset_and_collation] err: sql:admin recover index t a;: run "admin recover index t a;" at line 359 err Error 1105 (HY000): [components/tidb_query_executors/src/table_scan_executor.rs:425]: Data is corrupted, missing data for NOT NULL column (offset = 0) ERRO[0663] run test [expression/issues] err: sql:select * from t where field('A', a collate utf8mb4_general_ci, b) > 1;: failed to run query "select * from t where field('A', a collate utf8mb4_general_ci, b) > 1;" around line 1445, we need(79): select * from t where field('A', a collate utf8mb4_general_ci, b) > 1; a b sele but got(79): select * from t where field('A', a collate utf8mb4_general_ci, b) > 1; a b a A ERRO[0663] run test [planner/cascades/integration] err: sql:select /*+ HASH_AGG() */ a, count(distinct a) from t;: failed to run query "select /*+ HASH_AGG() */ a, count(distinct a) from t;" around line 135, we need(78): select /*+ HASH_AGG() */ a, count(distinct a) from t; a count(distinct a) 1 2 but got(78): select /*+ HASH_AGG() */ a, count(distinct a) from t; a count(distinct a) 2 2 ERRO[0663] run test [planner/core/plan_cache] err: sql:select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934;: failed to run query "select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934;" around line 209, we need(263): select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934; Error 1690 (22003): DECIMAL value is out of range but got(263): select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934; Error 1690 (22003): value is out of range in '' ERRO[0663] run test [session/vars] err: sql:SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable';: failed to run query "SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable';" around line 3, we need(98): SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable'; variable_value SET G but got(98): SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable'; variable_value true /unhold
gharchive/pull-request
2023-10-16T09:07:51
2025-04-01T04:32:58.045284
{ "authors": [ "purelind" ], "repo": "PingCAP-QE/ci", "url": "https://github.com/PingCAP-QE/ci/pull/2519", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2014636052
chore(apps/prod/tekton/setup): bump tekton operator to v0.61.0 With components updated: hub "v1.9.0" dashboard "v0.28.0" pipeline "v0.39.0" chains "v0.11.0" Signed-off-by: wuhuizuo wuhuizuo@126.com I have already done a preliminary review for you, and I hope to help you do a better job. This pull request updates the Tekton operator and several other components to the newest versions. There are no potential problems found within this pull request. Suggestion: It is always a good idea to test the updated components before merging the changes into the main branch. If it failed I will revert it. /approve
gharchive/pull-request
2023-11-28T14:56:14
2025-04-01T04:32:58.048874
{ "authors": [ "wuhuizuo" ], "repo": "PingCAP-QE/ee-ops", "url": "https://github.com/PingCAP-QE/ee-ops/pull/804", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2023159936
feat(charts): add chart git-cdn We will deploy it to proxy git clone globally to save CI/CD time and brandwidth. It's based open source project git_cdn. Signed-off-by: wuhuizuo wuhuizuo@126.com /approve
gharchive/pull-request
2023-12-04T06:57:30
2025-04-01T04:32:58.050730
{ "authors": [ "wuhuizuo" ], "repo": "PingCAP-QE/ee-ops", "url": "https://github.com/PingCAP-QE/ee-ops/pull/819", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2596141522
Customization of wave function for TileReflection The sharp, clear-cut edges of the tiles in TileReflection are achieved through the tangent function, which approaches infinite on one side and minus infinite on the other side, resulting in very sudden jumps. But what if we chose a different function? For example if we chose sine, the image would be bulged in a tile pattern but it wouldn't have clear edges. Other functions would behave in different, interesting ways. This is cool, definitely has some interesting effects! I'm mostly unsure about the naming of the Wave Type parameter and its menu options, which are pretty technical and might not make much sense to a user without an understanding of what the effect is doing internally. Ideally there could be some names that describe what behaviour is produced? @cameronwhite what names would you suggest? I haven't thought much about names for the new modes, but for example you might label the existing mode as Tile Shape -> Curved Edges or something like that? The current mode looks like the image is reflected in a mirror made up of tiles that have some curvature along their edges. Ideally we could have some labels for the new modes that give some idea of the general look of the effect @cameronwhite if we focus on the visual effect and not the wave function as such, we'd have to think about this differently, so I reduced the number of options to two to start with (with more tile types coming in the future). Also, I gave the enum the name of TileType and the values are CleavedEdges (which is what we already have) and CurvedEdges (the new option). English is not my first language, so if you can think of better names, that would be cool. @cameronwhite I implemented your suggestions 👍
gharchive/pull-request
2024-10-18T01:27:33
2025-04-01T04:32:58.058870
{ "authors": [ "Lehonti", "cameronwhite" ], "repo": "PintaProject/Pinta", "url": "https://github.com/PintaProject/Pinta/pull/1051", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
15739453
Cosmetic bugfixes Created new branch for these commits as we decided to look again at the changes for Pinta 1.6 (https://github.com/PintaProject/Pinta/pull/61). I needed this in order to be able to continue working on master without mixing up things so much. Closing this due to lack of time to look into it
gharchive/pull-request
2013-06-19T11:21:31
2025-04-01T04:32:58.060330
{ "authors": [ "ericksson" ], "repo": "PintaProject/Pinta", "url": "https://github.com/PintaProject/Pinta/pull/63", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
131314427
Cant compile it This shit cannot be compiled So don't use it. Stupid and ignorant issues like this are not tolerated.
gharchive/issue
2016-02-04T11:07:53
2025-04-01T04:32:58.077823
{ "authors": [ "finnishhacker", "ludo237" ], "repo": "Pirate-Crew/IPTV-Extreme", "url": "https://github.com/Pirate-Crew/IPTV-Extreme/issues/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2000842386
Unable to install the component Hello, I've just tried to install the pirateweather component but got no entities after installing. In the log file I have the following error message: Logger: homeassistant.setup Source: setup.py:379 Unable to prepare setup for platform pirateweather.weather: Platform not found (cannot import name 'WeatherEntityFeature' from 'homeassistant.components.weather' (/***/homeassistant/lib/python3.10/site-packages/homeassistant/components/weather/init.py)). Anybody knows a fix for this problem? Thank you very much. I'll tag @alexander0042 so he can take a look at this but some helpful information to know is: Which version of the integration are installing and what version of HA are you running? Hello, Home Assistant Version: 2023.7.3 Pirate-Weather: Version 1.3.2 Hi, thanks for opening this issue, and apologies for the issue here. The hiccup is that you're running the pre-release version of Pirate Weather, but your version of HA is tool old. However, HACS should be able to prevent this, so let me take a look at the manifest to see what went wrong here! Argh, just realized I've installed HA Core - will this even work in core? The pre-release version of the integration requires HA version 2023.10 so you'd need to update your version of HA to run that version. The release version 1.2.3 will work with your HA version so have you tried installing that and seeing if it works? I'm not a HA expert but I'll tag @alexander0042 again as he might know more information on if the integration would work in Core. I've updated HA to the latest version and it's working - thanks! Glad you were able to get it working. Since this issue has been solved I'm going to close this issue.
gharchive/issue
2023-11-19T13:04:27
2025-04-01T04:32:58.083162
{ "authors": [ "alexander0042", "cloneofghosts", "km8742" ], "repo": "Pirate-Weather/pirate-weather-ha", "url": "https://github.com/Pirate-Weather/pirate-weather-ha/issues/143", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
311535732
Avoid panic when blending two pixels with zero alpha. I ran into a situation where trying to overlay two rgba images led to a panic. The root cause was a division by zero. Here's a minimal reproducible example program run against image = "0.18.0": extern crate image; use image::{ImageBuffer, Rgba, RgbaImage}; use image::imageops::overlay; fn main() { let mut a: RgbaImage = ImageBuffer::new(1, 1); a.put_pixel(0, 0, Rgba([0, 0, 0, 0])); let mut b: RgbaImage = ImageBuffer::new(1, 1); b.put_pixel(0, 0, Rgba([0, 0, 0, 0])); overlay(&mut a, &b, 0, 0); } I don't have a ton of graphics experience, so I'm not sure if there's a "correct" way to blend two completely transparent pixels in terms of retaining the color components. This patch just uses the blended colors and leaves alpha at zero. This patch is the simplest thing I could think of to fix, but I'd be happy to change to something more idiomatic (num::CheckedDiv trait and explicit handling of the None condition?) It seems that the final_alpha is set to 1.0? Should this not be zero? Doh, I wasn't thinking and was only testing against images with completely opaque pixels. I'll amend the commit. That said --- is there any reason to do anything with the color values at all when blending two transparent pixels? If not, then presumably it'd be more efficient to return a transparent white/black pixel rather than doing calculations on the color components. I think blending is good, because it preserves the information in the image to neighboring pixels. For example, if you are super-sampling textures, white/black pixels can cause trouble. Thanks for the quick reply! I did some reading, and I don't think this kind of compositing is well defined when both pixels have zero alpha. The referenced stack overflow answer and SVG spec, defines compositing in terms of "premultiplied" color components (as defined in the original Porter and Duff paper). But since alpha is in the domain [0, 1], let alpha_final = bg_a + fg_a - bg_a * fg_a; must always be greater than the alpha of either pixel. Which means that it can only be zero when both pixel's alpha are zero. But in that case, the "premultiplication" step let (bg_r_a, bg_g_a, bg_b_a) = (bg_r * bg_a, bg_g * bg_a, bg_b * bg_a); let (fg_r_a, fg_g_a, fg_b_a) = (fg_r * fg_a, fg_g * fg_a, fg_b * fg_a); will set all of the color components to zero. Or am I missing something in my logic there? Premultiplied color components do not conserve the color information in pixels when making them transparent. This is a different way of encoding images, which is not used here. The code I quoted wasn't hypothetical, it's from the RGBA blend implementation and has the comment "We premultiply our channels bu their alpha, as this makes it easier to calculate". Or are you suggesting that we switch to a different blending algorithm in the case when both source alphas are zero? @lynaghk I messed this up, sorry. I was not aware of premultiplying colors being used here. Not sure if this is the right algorithm. @bvssvni No worries. I'm not too familiar w/ the tradeoffs for color blending and graphics stuff like this, so I'm not sure what the appropriate thing to do would be --- all I wanted from my changes was to avoid a panic. If you do find another way through and want a hand, let me know. But no worries if it'd be easier to just do the changes yourself too. Another issue that there is no conversion to sRGB before blending. This algorithm is very opinionated. However, that is a design decision that could be dealt with in a later version. Can you add a unit test that blends two transparent pixels and see what happens to the alpha channel? At least the resulting alpha channel should be zero. I think it should simply return instead of setting the alpha to 1.0. Closing in favor of https://github.com/PistonDevelopers/image/pull/759 Thanks for pointing this out! Sorry for the delay on unit tests, I was away from the Internet. Returning makes sense, given that the blend operation mutates. Thanks for the fix and your work on this library!
gharchive/pull-request
2018-04-05T09:27:49
2025-04-01T04:32:58.095973
{ "authors": [ "bvssvni", "lynaghk" ], "repo": "PistonDevelopers/image", "url": "https://github.com/PistonDevelopers/image/pull/744", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1081735411
yarn start error yarn install success,but yarn start with this error: Attempted import error: 'CanvasSlicImg' is not exported from '../../pixels-metaverse'. yarn install success,but yarn start with this error: Attempted import error: 'CanvasSlicImg' is not exported from '../../pixels-metaverse'. Sorry, since I need to make some upgrades to the feature to participate in the contest, the code has not been released yet, so the update to the react-pixel-metaverse package has not been released yet, so you may not find the corresponding CanvasSlicImg component. 很抱歉,由于我需要对功能进行一些升级才能参与黑客松比赛,对react-pixel-metaverse包的更新尚未发布,因此您可能找不到相应的CanvasSlicImg组件。 You can now access the link , which was the previous version, but can be used for learning. In addition, the project will be released once before Christmas, and then you can see that the warehouse code is working properly. Thank you for your attention and hope we can make progress together. 当前您可以访问链接 https://github.com/xiangzhengfeng/pixels-metavers,这是以前的版本,但可用于学习。此外,该项目 将在圣诞节前发布一次,到时候您可以看到仓库代码正常工作了。感谢您的关注,希望我们能共同进步。
gharchive/issue
2021-12-16T03:38:40
2025-04-01T04:32:58.132191
{ "authors": [ "lhzwb2008", "qianduanxinlv" ], "repo": "PixelsMetaverse/pixels-metaverse", "url": "https://github.com/PixelsMetaverse/pixels-metaverse/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1864337680
出现这个问题 我有8G的GPU,但是还是出现这样的问题? G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\whisper\timing.py:58: NumbaDeprecationWarning: The 'nopython' keyword argument was not supplied to the 'numba.jit' decorator. The implicit default value for this argument is currently False, but it will be changed to True in Numba 0.59.0. See https://numba.readthedocs.io/en/stable/reference/deprecation.html#deprecation-of-object-mode-fall-back-behaviour-when-using-jit for details. def backtrace(trace: np.ndarray): Use 8 cpu cores for computing Traceback (most recent call last): File "G:\AI\VALL\VALL-E-X-master\VALL-E-X-master\ui.py", line 84, in <module> whisper_model = whisper.load_model("medium").cpu() File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\whisper\__init__.py", line 154, in load_model return model.to(device) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 1145, in to return self._apply(convert) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 797, in _apply module._apply(fn) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 797, in _apply module._apply(fn) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 797, in _apply module._apply(fn) [Previous line repeated 2 more times] File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 820, in _apply param_applied = fn(param) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 1143, in convert return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 4.00 GiB total capacity; 3.05 GiB already allocated; 0 bytes free; 3.06 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF 看起来是在加载whisper(另一个ai模型,用于识别声音中的文字)的时候出现了错误。 目前vallex需要输入一段声音,这段声音中的文字内容,以及要合成的文字内容,输出合成的声音。 如果你没有输入原音频中的文字内容,那么就会调用whisper模型自动识别原音频中的文字内容。这将花费额外的显存。 可以尝试手动输入原音频中的文本内容来避免调用whisper 感谢哥哥答复,我运行 ui时(python launch-ui.py),没有启动成功就出现这个报错。很奇怪 可以尝试修改源代码,删除加载whisper模型的部分 我试试看看~ 可以了,打开ui了。是whisper和程序强GPU了。去掉whisper就可以
gharchive/issue
2023-08-24T03:28:48
2025-04-01T04:32:58.175782
{ "authors": [ "40740", "ZYFDroid" ], "repo": "Plachtaa/VALL-E-X", "url": "https://github.com/Plachtaa/VALL-E-X/issues/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2435373023
chore(appeals): health-check (eviction settings) Adds a timeout (5 mins) for health check eviction Issue ticket number and link Type of change 🧩 [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Documentation update [x] Other (please explain in the description section above) Checklist before requesting a review [ ] I have performed a self-review of my own code [ ] I have double checked this work does not include any hardcoded secrets or passwords [ ] I have made corresponding changes to the documentation [ ] I have provided details on how I have tested my code [ ] I have referenced the ticket number above [ ] I have provided a description of the ticket [ ] I have included unit tests to cover any testable code changes Unsupported argument
gharchive/pull-request
2024-07-29T13:04:57
2025-04-01T04:32:58.179974
{ "authors": [ "fabfab-kc" ], "repo": "Planning-Inspectorate/appeals-back-office", "url": "https://github.com/Planning-Inspectorate/appeals-back-office/pull/334", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1529125494
chore(web/applications): log sessionaccount and AD access token Describe your changes Issue ticket number and link Type of change 🧩 [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Documentation update [ ] Other (please explain in the description section above) Checklist before requesting a review [ ] I have performed a self-review of my own code [ ] I have double checked this work does not include any hardcoded secrets or passwords [ ] I have made corresponding changes to the documentation [ ] I have provided details on how I have tested my code [ ] I have referenced the ticket number above [ ] I have provided a description of the ticket [ ] I have included unit tests to cover any testable code changes :tada: This PR is included in version 1.0.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket: :tada: This PR is included in version 1.0.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-01-11T14:12:18
2025-04-01T04:32:58.186028
{ "authors": [ "AnthonyJWalsh", "lutamKC" ], "repo": "Planning-Inspectorate/back-office", "url": "https://github.com/Planning-Inspectorate/back-office/pull/668", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1933746743
[common] instruction definition Adds opcode identifiers operation identifiers decoded instruction structure definition LGTM
gharchive/pull-request
2023-10-09T20:27:48
2025-04-01T04:32:58.215488
{ "authors": [ "Tako-San", "mmamayka" ], "repo": "Plumbus-Group-Inc/Plumbus-VM", "url": "https://github.com/Plumbus-Group-Inc/Plumbus-VM/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
816390963
Modification of QS after a success Allowance of modifications of QS after a roll, but only if the roll was a success (otherwise, you could work-around by just adding 3FP to the roll) Needed for future implementations anyways, like Athlet. Low prio, if accepted at all. Alternatively, allow to edit the Fertigkeitspunkte after a roll (currently, this works for dice values and modificators, but not for FP.
gharchive/issue
2021-02-25T12:27:06
2025-04-01T04:32:58.218599
{ "authors": [ "mckee14" ], "repo": "Plushtoast/dsa5-foundryVTT", "url": "https://github.com/Plushtoast/dsa5-foundryVTT/issues/139", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1231421223
Fix error recovery mode for SubmitTx As part of #233 we implemented a simple approach to error recovery with the use of reconnecting-websocket. However, there's one special endpoint for which simple replaying is not going to work. We need to check if a transaction has been sent before resubmitting it (otherwise UTXOs may be consumed by TX and we'll get an error when trying to send it for the second time). How to address this problem: maintain request queue as a list of transactions instead of a list of String payloads. Before retrying, use https://ogmios.dev/mini-protocols/local-tx-monitor/ to see if a transaction is in local node mempool. The reason why we are delaying this: the chance to observe the error is low.
gharchive/issue
2022-05-10T16:38:15
2025-04-01T04:32:58.220509
{ "authors": [ "klntsky" ], "repo": "Plutonomicon/cardano-transaction-lib", "url": "https://github.com/Plutonomicon/cardano-transaction-lib/issues/398", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1123817023
[BUG] Unable to reinitialize plugin after getting GDPR consent Hi, first of all, thanks for the plugin, I've been using it in my iOS app that was released recently, and now I'd like to implement GDPR consent, I just realized that your plugin also supports this. I'm almost done debugging it, but I noticed that the plugin won't reinitialize after getting user consent (in case they choose non-personalized ads), I couldn't find "is_personalized" parameter or something like that in your initialization function (https://github.com/Poing-Studios/godot-admob-ios/blob/master/plugin/admob/src/main/admob.mm#L143), so I'm curious how do you forward consent to the Google Mobile Ads SDK? I'm using https://github.com/Shin-NiL/Godot-Android-Admob-Plugin + https://github.com/pschw/GodotConsentPlugin for the Android version of my app and I'm able to reinitialize the admob plugin with the "is_personalized" parameter set to false when users don't consent, Shin-NiL's plugin also configures GADRequest (https://github.com/Shin-NiL/Godot-Android-Admob-Plugin/blob/main/admob-plugin/godotadmob/src/main/java/shinnil/godot/plugin/android/godotadmob/GodotAdMob.java#L115) as described in the admob documentation (https://developers.google.com/admob/ump/android/quick-start#forward-consent / https://developers.google.com/admob/ump/ios/quick-start#forward-consent). I just want to make sure your plugin works the same to comply with GDPR, since I'm thinking of using it in android too instead of Shin-NiL's. Btw, the consent works as expected with "Test Europe User Consent" enabled or using a vpn, but when testing it in my location (Colombia), it still shows a window but only with a continue button and a different text that says something like "ads keep our app free, select allow in the next screen...", then it shows ATT consent as expected (I use another plugin for that), is that intended? because the console prints "Consent Form Available" and then "User consent obtained. Personalization not defined." when I was actually expecting "Consent Form not Available" or "User consent not required. For example, the user is not in the EEA or the UK.". Hope I made myself understood, and please let me know if these are bugs or missing features, I'd like to contribute but idk much about those languages nor creating plugins, anyways, thanks for reading. Plugin v2.0.4 Editor plugin v1.2.4 Godot v3.4.2 Hello, thank you for using our plugin! I couldn't find "is_personalized" parameter or something like that in your initialization function We don't use is_personalized parameter because when I developed this part of GDPR, in the documentation of AdMob I remember it said that it was deprecated to do it this way, so we only use UMP and to talk about whether it is personalized or not, you need to change it within the AdMob application: http://admob.google.com/ I just want to make sure your plugin works the same to comply with GDPR Yes it does, but i have not done much tests because i don't live in Europe. From what I analyzed in the codes, I don't know how "GodotConsentPlugin" will communicate with "Godot-Android-Admob-Plugin" can you tell me how you are using these two Plugins? For example, if in AdMob you defined by code to always be isPersonalized = true, from what I understand, regardless of what the user selects, it will always be personalized, I believe that what the "Forward consent" is used for the game to always demonstrate personalized ads or not I saw that in the code it is considering these 3 consent statuses: enum CONSENT_STATUS { UNKNOWN, NON_PERSONALIZED, PERSONALIZED } But I remember that when I tested as i was developing, regardless of whether the user clicked "Consent" or "Do not consent", AdMob can't know what he clicked, so for me there's no way to know if it was granted or not, but I'll do some tests to guarantee Check the documentation: https://developers.google.com/admob/ump/android/quick-start#present_the_form_if_required There are only four Status: https://developers.google.com/admob/ump/android/api/reference/com/google/android/ump/ConsentInformation.ConsentStatus ConsentStatus.UNKNOWN: Unknown consent status. ConsentStatus.REQUIRED: User consent required but not yet obtained. ConsentStatus.NOT_REQUIRED: User consent not required. For example, the user is not in the EEA or the UK. ConsentStatus.OBTAINED: User consent obtained. Personalization not defined. but when testing it in my location (Colombia), it still shows a window but only with a continue button and a different text that says something like "ads keep our app free, select allow in the next screen..." It should not happen, maybe you forgot to export again? I will test it, by the way i also live on South America. then it shows ATT consent as expected (I use another plugin for that), is that intended? I don't think so because the console prints "Consent Form Available" and then "User consent obtained. Personalization not defined." Oh, maybe as you have accepted before "being" in europe, what you selected was saved, for that you would need to reset the consent using this method: https://github.com/Poing-Studios/godot-admob-ios/blob/master/plugin/admob/src/main/admob.mm#L97 Also: I've done some research now and it doesn't seem to be deprecated to declare "npa=1", they added it to the updated "EU Consent" page maybe it's a good idea to add the "isPersonalized" parameter but I really don't know if it's recommended since AdMob says this on end of page: https://developers.google.com/admob/ump/ios/quick-start#forward-consent If non-personalized ads are requested, the ad request URL currently includes &npa=1. However, note that this is an internal implementation detail of the Google Mobile Ads SDK and is subject to change. If you want to help me contribute to this part of the GDPR, you can join the Discord link: https://github.com/Poing-Studios/godot-admob-android/discussions/63 This is something I really need help with as the AdMob documentation isn't very good and as I don't live in Europe it just makes it more difficult Also: Our Android and iOS Plugin code should be identical, so GDPR should work the same for Android and iOS I don't recommend using multiple Plugins for the "same" purpose because in addition to being bad to maintain the plugins, you will probably have to rewrite code for each operating system, Android and iOS Maybe this old issue can help you in any way: https://github.com/Poing-Studios/godot-admob-android/issues/57 can you tell me how you are using these two Plugins? So this is how I use "GodotConsentPlugin" and reinitialize "Godot-Android-Admob-Plugin": https://github.com/bouncymarble/bouncymarble.github.io/blob/main/ConsentPlugin.gd Btw, I had to add a neutral age screen that asks the user for their age before asking for GDPR consent (Google Play rejected my app the first time because consent shouldn't be shown to children under 13). I saw that in the code it is considering these 3 consent statuses Yeah I was wondering why the statuses are different, also the form interface is very different as you can see in the demo video (https://youtu.be/PJ2H8ZK8O_w), but I just realized they actually use "Personalized Ad Consent SDK" which is now deprecated (https://developers.google.com/admob/android/eu-consent/) instead of UMP SDK, so I think I must migrate to your plugin. so we only use UMP and to talk about whether it is personalized or not, you need to change it within the AdMob application: http://admob.google.com/ Ok, should I set it to non-personalized or keep it personalized (default)? Oh, maybe as you have accepted before "being" in europe, what you selected was saved Hmm you might be right, maybe I forgot to reset it. I just need to make sure the user is not in europe (with the signal "consent_info_update_success" or "consent_status_changed") so I don't show them a "data usage" button from my game settings to update their consent. you can join the Discord link Thanks for the link. I don't recommend using multiple Plugins for the "same" purpose because in addition to being bad to maintain the plugins, you will probably have to rewrite code for each operating system, Android and iOS Yeah I've been trying to merge my Android and iOS projects with your plugin so it's easier for me to release updates. Maybe this old issue can help you in any way: Poing-Studios/godot-admob-android#57 Well, I've had no issues updating the consent for european users, just added a "data usage" button in my game settings panel with this code: func _on_TB_UpdateGDPR_button_up(): MobileAds.reset_consent_state() MobileAds.request_user_consent() Although it still shows the button for non european users, but maybe it's due to the above mentioned, so I need to test it again. Thank you so much for your help. So this is how I use "GodotConsentPlugin" and reinitialize "Godot-Android-Admob-Plugin": https://github.com/bouncymarble/bouncymarble.github.io/blob/main/ConsentPlugin.gd I think this will not work, i just tested here on my phone and the buttons "Consent" and "Do not Consent" have the same result: Get the ConsentStatus.OBTAINED which means this: https://developers.google.com/admob/ump/android/api/reference/com/google/android/ump/ConsentInformation.ConsentStatus#public-static-final-int-obtained. I had to add a neutral age screen that asks the user for their age before asking for GDPR consent (Google Play rejected my app the first time because consent shouldn't be shown to children under 13). I didn't know that, maybe we can include this in the plugin or create a new plugin to display this part of getting the user's age and thus displaying the form or not also the form interface is very different as you can see in the demo video (https://youtu.be/PJ2H8ZK8O_w) Yes, i think this is a deprecated version Ok, should I set it to non-personalized or keep it personalized (default)? Unfortunately I don't know the answer to this question :( But where you can configure this is here: Although it still shows the button for non european users Remember to unmark this button: If you mark this, whenever the user is or not on EEA, it will show the Form Closing this as there's no need to reinitialize the plugin, it seems the UMP handles user choice and somehow forwards it to the Google Mobile Ads SDK. Although there's this issue when "don't consent" button is pressed (or "manage options" and then "confirm choices" without selecting any consent option), ads won't load due to error code 3 (ERROR_CODE_NO_FILL), but this has nothing to do with the plugin, so the only way to load ads is by resetting and updating consent. More info about what admob requires to load either personalized or non-personalized ads: https://support.google.com/admob/answer/9760862?hl=en&ref_topic=9756841 Also, here are the answers to some questions about UMP and GDPR that @gumaciel and I asked on the Google Mobile Ads SDK Technical Forum: https://groups.google.com/g/google-admob-ads-sdk/c/xfLEdy06pyc the console prints "Consent Form Available" and then "User consent obtained. Personalization not defined." when I was actually expecting "Consent Form not Available" or "User consent not required. For example, the user is not in the EEA or the UK." Regarding this issue, it turns out I didn't forget to unmark "Test Europe User Consent", and what is actually triggering this message "Consent Form Available" is the IDFA form that is required on iOS, no matter location. On Android it returns "Consent Form not Available" as expected if not in the EEA. Btw, I had another issue when trying to start my app with no internet connection, the plugin tries to initialize and returns status 0 which I thought was ok, but @gumaciel tested it too and always gets status 1 (MobileAds.AdMobSettings.INITIALIZATION_STATUS.READY). The thing is when I turn on internet, initialize() keeps returning status 0, it seems like it can only be called once. Anyways, what I had to do is make sure there's internet connection to initialize the plugin. For this I use Timer and HTTPRequest nodes, here's the code in case anyone finds it useful: func _on_Timer_CheckInternet_timeout(): check_internet() func check_internet(): if http_request.get_http_client_status() != 0: return http_request.request("https://google.com") func _on_HTTPRequest_request_completed(result, response_code, _headers, _body): if result == 0 and response_code == 200: if not MobileAds.get_is_initialized(): MobileAds.ready_to_initialize()
gharchive/issue
2022-02-04T05:13:28
2025-04-01T04:32:58.257890
{ "authors": [ "bouncymarble", "gumaciel" ], "repo": "Poing-Studios/godot-admob-ios", "url": "https://github.com/Poing-Studios/godot-admob-ios/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
167625551
Error when starting program Error: A JavaScript error occurred in the main process Uncaught Exception: Error: spawn python ENOENT at exports._errnoException (util.js:949:11) at Process.ChildProcess._handle.onexit (internal/child_process.js:182:32) at onErrorNT (internal/child_process.js:348:16) at _combinedTickCallback (internal/process/next_tick.js:74:11) at process._tickCallback (internal/process/next_tick.js:98:9) Operating System: OSX El Capitan Same error when starting. OS : Windows 7 Same error, windows 10 x64, latest commit 5f96174. screenshot shows fixes applied to this issue in past are present
gharchive/issue
2016-07-26T14:44:38
2025-04-01T04:32:58.278509
{ "authors": [ "JVenberg", "pyr0ball", "tzull" ], "repo": "PokemonGoF/PokemonGo-Bot-Desktop", "url": "https://github.com/PokemonGoF/PokemonGo-Bot-Desktop/issues/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
167725296
KeyError: 'fort_id' KeyError: 'fort_id' After #1039 change i update my dev version now my problem is that . Other Information OS: Ubuntu Git Commit: 9c99cc0de7e5b0c35762cae9b657c495d32e3ade same here ` [23:14:59] Starting PokemonGo Bot.... [23:14:59] Scanning area for objects.... [23:14:59] Walking from (51.500575, -0.127703) to [51.500575, -0.127703] 0.00km [████████████████████████████████████████] 100% [23:15:01] Attempting to evolve all pokemons ... [23:15:04] Something rustles nearby! Traceback (most recent call last): File "./pokecli.py", line 246, in main() File "./pokecli.py", line 237, in main bot.take_step() File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/init.py", line 46, in take_step self.work_on_cell(cell, location) File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/init.py", line 189, in work_on_cell if self.catch_pokemon(pokemon) == PokemonCatchWorker.NO_POKEBALLS: File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/init.py", line 337, in catch_pokemon return_value = worker.work() File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/cell_workers/pokemon_catch_worker.py", line 31, in work response_dict = self.create_encounter_api_call() File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/cell_workers/pokemon_catch_worker.py", line 400, in create_encounter_api_call fort_id = self.pokemon['fort_id'] KeyError: 'fort_id' ` I think who change it on other Issues #1045 #1046 maybe have same problem Commit just been reverted, please repull. #1065 @aslafy-z Thank you when i change to head 1056 working ;) @aslafy-z :( after visit pokestore now this problem :( @Zgrkaralar This is another isue due to your outdated config file. Recreate it from configs/config.json.example Now problem is this Please open another issue, and someone close this one please. Commit has been reverted, please repull. #1065
gharchive/issue
2016-07-26T22:16:05
2025-04-01T04:32:58.287119
{ "authors": [ "Zgrkaralar", "aslafy-z", "surfaace", "und3rdg" ], "repo": "PokemonGoF/PokemonGo-Bot", "url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/1067", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
170194360
Error in bot startup Hello , is experiencing an error when booting the bot, do not know what to do, please help me . The following image: Please ask this question in the #help section on slack. This is not an issue with the bot.
gharchive/issue
2016-08-09T15:20:18
2025-04-01T04:32:58.288637
{ "authors": [ "RedViper9", "piai46" ], "repo": "PokemonGoF/PokemonGo-Bot", "url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/3302", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
167171249
Update included yaml import, but missing in requirements.txt Expected Behavior Program runs Actual Behavior ImportError: No module named yaml Steps to Reproduce *Clone repo *Install *Run Other Information OS: Mac OSX Git Commit: 90225f7133ede69f10167e5ae816391023de0219 Was introduced in 49d3ef6 duplicate and already fixed in #355
gharchive/issue
2016-07-23T06:43:27
2025-04-01T04:32:58.290645
{ "authors": [ "eiselems" ], "repo": "PokemonGoF/PokemonGo-Bot", "url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/356", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
167182812
infinite loop if Out of pokeballs [#] Something rustles nearby! [#] A Wild Pidgey appeared! [CP 25] [x] Out of pokeballs... [#] A Wild Zubat appeared! [CP 167] [x] Out of pokeballs... [#] A Wild Pidgey appeared! [CP 25] [x] Out of pokeballs... [#] A Wild Zubat appeared! [CP 167] [x] Out of pokeballs... [#] Something rustles nearby! [#] A Wild Pidgey appeared! [CP 25] [x] Out of pokeballs... [#] A Wild Zubat appeared! [CP 167] [x] Out of pokeballs... [#] A Wild Pidgey appeared! [CP 25] [x] Out of pokeballs... [#] A Wild Zubat appeared! [CP 167] [x] Out of pokeballs... [#] Something rustles nearby! [#] A Wild Pidgey appeared! [CP 25] [x] Out of pokeballs... [#] A Wild Zubat appeared! [CP 167] [x] Out of pokeballs... [#] A Wild Pidgey appeared! [CP 25] [x] Out of pokeballs... [#] A Wild Zubat appeared! [CP 167] [x] Out of pokeballs... This should be fixed with the newest dev branch.
gharchive/issue
2016-07-23T12:46:06
2025-04-01T04:32:58.295026
{ "authors": [ "Hatsunyan", "s1x" ], "repo": "PokemonGoF/PokemonGo-Bot", "url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/427", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1510196923
A FAQ would be awesome Reviewing the discord chat reveals that many of the same questions are asked and answered over and over. A linkable FAQ would be great Heres a start FAQ Is Pokemon Unity still in development? Yes Is the Unity part of this project being worked on? No Is there a more up to date version and runnable version of this? Not that we know How do I start developing my own Pokemon game? Thats completely up to you. The maintained Pokemon code in this repository is actually NOT related to Unity at all. This means someone can take the code here and start their own Pokemon game in any way that supports loading C# DLL libraries. Unity is just 1 way :) What relevance do the folders other than the Pokemon Unity folder have? The repo contains several projects. Each folder (probably) represents one of those projects. The Pokemon Unity folder is just where the Unity related files are found TODO: add breakdown of contents of each parent folder Unity FAQ Why is my screen pink? Unity metafiles were broken sometime in the past. Because of this, images in (probably) all scenes are corrupted and show up pink How do I learn Unity? There are several great tutorials on YouTube to learn Unity. To name a few Brackeys Code Monkey Jason Weimann Unity Sebastian Lague Game Maker's Toolkit Why isnt the Unity side of things being worked on? There is only 1 developer currently active on this project. Their focus is on the non-Unity Pokemon framework Why is my screen pink? The answer can be a pretty lengthy one, but it's just a unity issue with the .meta files. The why, the how, and the solution is discussed here: https://discord.com/channels/285560546244427777/285565647507685379/1046810010132435067 Why isnt the Unity side of things being worked on? I wouldnt say it's because i'm the only dev, but that we dont have any "Unity-developers", it's way more appealing to write programming code, than to struggle with getting unity to behave nice. Those that are comfortable with unity, are mostly working on their own private projects... It's not easy to manage a unity project in git, as there's lots of problems that occur just on the .meta files, if the team isnt in sync or on the same page, in terms of development style/practice, and communication (overlapping edits in the same scene, across different commits can cause major problems when a conflict emerges). At the end of the day... the project is geared more towards being an API framework, than a game kit (a playable unity demo/sample used as a foundational resource to start you game from). How to download? This is a Github question, not really specific to project. But since a lot of people new to the project are also new to github, i recommend just reading this guide and following the steps. Better than downloading a zip, and the frustrations that come after from not receiving bug fixes and updates. https://docs.github.com/en/desktop/contributing-and-collaborating-using-github-desktop/adding-and-cloning-repositories/cloning-a-repository-from-github-to-github-desktop or... https://docs.github.com/en/repositories/creating-and-managing-repositories/cloning-a-repository Downloading zips aren't always the answers because some files don't get moved over into zip There are settings on some projects on repository to limit what will and will not go into a "zip" download. Though I don't believe I'm using those settings... When you clone a project, it bypasses those settings and gives you a one-for-one copy of the project and all of its contents Literally, "cloning" instead of downloading a "package" This was an answer i gave to a user in discord chat. But when you "clone" the project, the project lives on your computer as well as on the git cloud it came from... so if a problem is "fixed" on the server, you dont need to download the entire project over again just for the files that were corrected (also, you dont need to worry about your changes be lost or deleted). If you make any edits and want to share it with the project for others to download, then this is also the same steps needed in order to do that. I downloaded the Github, like instructed above, but how do i play it? I want to experience a pokemon game inside unity... This isnt a game kit, nor a pokemon clone. It's a resources for your game to implement pokemon features (like a plug-in or API). The resources are downloaded assets you can add to your existing or new project, and shortcut the need to have to program all the pokemon logic yourself from zero. This is a transferrable drag-and-drop library, and for the time being, there really isnt any flagship game to showcase or demo the content in usage. (Unless you wish to be the first, and link back to us here, for us to tag you... no one contributing back to this repository is a "game developers" :shrug: ) What about the youtube video that i saw, where the game was running in unity? How do i download and play that? That's found in the release tab, as an .exe download... https://github.com/PokemonUnity/PokemonUnity/releases you can grab the project that was created by the original author of this project repository. But it's dated back to 2015, and has a lot of incomplete code that we're attempting to resolve by rewriting and redoing the code source from the ground up. So if you want the latest changes and fixes, there is no "game" to play; and yes... you will download a git repository with a unity project that may contain a lot of errors (preventing you from playing any scene).
gharchive/issue
2022-12-24T21:52:57
2025-04-01T04:32:58.319506
{ "authors": [ "Xerner", "herbertmilhomme" ], "repo": "PokemonUnity/PokemonUnity", "url": "https://github.com/PokemonUnity/PokemonUnity/issues/129", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
393752346
[Sample app Kotlin] separate sample-kotlin module Creates a separate sample-kotlin module. As a start, ScanActivity and DeviceActivity are implemented, with parts of functionalities outcommented to make for an easier review. Also all layouts have been copied over from the Java sample and cleaned up a little bit. I have added Work In Progress to the title as I have seen that there are some TODOs in the code. Feel free to rename when you finish Have a Happy New Year! :) In this case I will put it on top of my Open Source priority list One last thing — I have just released version 1.8.0 which is now using androidx. I have migrated the java sample app already. Could you also update the kotlin one? From now on I do not expect more such changes
gharchive/pull-request
2018-12-23T14:03:50
2025-04-01T04:32:58.362151
{ "authors": [ "dariuszseweryn", "lukaszkalnik" ], "repo": "Polidea/RxAndroidBle", "url": "https://github.com/Polidea/RxAndroidBle/pull/520", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1069058593
Pack load error icon Including a small little icon after pressing Reload Packs if any of the packs failed to load would make it easy to see if you need to check the console. Additionally, a check mark could indicate no pack errors. I plan on redoing the majority of that functionality, as I don't like the tabs feature. Thoughts?
gharchive/issue
2021-12-02T02:23:37
2025-04-01T04:32:58.393470
{ "authors": [ "Astrashh", "solonovamax" ], "repo": "PolyhedralDev/BiomeTool", "url": "https://github.com/PolyhedralDev/BiomeTool/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
702390706
[lit-html] Clarify behavior of classMap The test for classMap say that dynamic classes can not override and remove static classes, but this is only true on the first render. Subsequent renders will remove static classes. This is clearly broken, but what is the right behavior? To allow overriding or not? If we're going to change this, 2.0 is the time. I think the most important thing is that classMap can remove the classes it previously set. I don't think the behavior when you have a conflict between a static class and a classMap class is important. We should just add docs that you shouldn't do that. We can't change this behavior w/o potentially breaking people though, so I do want to settle on what we want, even if we recommend against class clashes like this.
gharchive/issue
2020-09-16T02:07:11
2025-04-01T04:32:58.413077
{ "authors": [ "justinfagnani", "sorvell" ], "repo": "Polymer/lit-html", "url": "https://github.com/Polymer/lit-html/issues/1278", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
144036224
Add carbon elements Let's add carbon elements to the catalog. Should close https://github.com/PolymerElements/carbon-route/issues/57 This looks good locally. I needed to stamp v0.8.1 of carbon-route to get our updated docs (v0.8.0 had invalid markdown). The demo looks pretty good but I'd like to wait for https://github.com/PolymerElements/carbon-route/issues/53 to land before we push this out though, so that there's a useful URL bar in the demo. After that lands (and we stamp v0.8.2) I'd say that this is good to go. great! ^ cc @tjsavage Sounds good - let's wait for it to land then push. Ok, it's landed and released as v0.8.2, LGTM!
gharchive/pull-request
2016-03-28T17:45:46
2025-04-01T04:32:58.417596
{ "authors": [ "blasten", "rictic", "tjsavage" ], "repo": "Polymer/polymer-element-catalog", "url": "https://github.com/Polymer/polymer-element-catalog/pull/264", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
197519963
Polymer 2.0 - Dblclick is not bubbling up http://codepen.io/anon/pen/ObYaqX?editors=1000 If you look at the example you'll notice a blue box with black box inside The blue box is listening to the dblclick event, however when you click the black box the dblclick event does not trickle up to the blue box... any idea why? (If you replace dblclick with click you'll see it works that way, which is odd) Here's a slightly updated example that works x-browser: http://codepen.io/anon/pen/wJBJyd?editors=1000. Closing since this no longer seems to repro on the example. Please re-open if the problem resurfaces.
gharchive/issue
2016-12-26T00:17:03
2025-04-01T04:32:58.420488
{ "authors": [ "afroradiohead", "sorvell" ], "repo": "Polymer/polymer", "url": "https://github.com/Polymer/polymer/issues/4234", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
170665613
Do notify once per property at most If you have multiple behaviors defining the same property, all setting notify to true, Polymer just blindly added multiple effects. This resulted in multiple DOM events sent upwards. The fix is to just check if there is already a notify effect registered for a given property, before we register the effect. Reference issue #3418. Fixes #1791. Fixes PolymerElements/paper-input#66 I hit this bug that this solves with paper-input on-value-change firing twice for every key press...whatever became of this since it has not been merged? Closing per https://github.com/Polymer/polymer/issues/1791#issuecomment-321436929
gharchive/pull-request
2016-08-11T14:53:11
2025-04-01T04:32:58.422992
{ "authors": [ "TimvdLippe", "dman777", "kaste" ], "repo": "Polymer/polymer", "url": "https://github.com/Polymer/polymer/pull/3856", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
146141329
1.14.9 fails on Windows Older version 1.14.8 works, but new version 1.14.9 returns: ERROR finding /Work/QLine/WebClients/AppPortal/elements/bower_components/bower_components/bower_components/bower_components/web-animations-js/web-animations-next-lite.min.js There is a package neon-animation which references web-animations-js. +1 Getting the same issue +1 Same issue on Mac +1 Same issue on ubuntu Same here. In my case I have a local element in project\component folder (not in bower_components) that links ../bower_components/paper-dialog/paper-dialog.html and it ends up looking for project\component\bower_components\paper-dialog\import\neon-animation\import\neon-animation\import\web-animations-js\web-animations-next-lite.min.js This is a really bad since gulp-vulcanize will automatically get the latest 1.14.x version of this tool. Working on it Unpublished v1.14.9 until I can get to the bottom of the breakage.
gharchive/issue
2016-04-05T23:49:02
2025-04-01T04:32:58.431745
{ "authors": [ "ElsewhereGames", "azakus", "davidlwatsonjr", "dudabone", "govis", "kwaclaw", "xwsxethan" ], "repo": "Polymer/vulcanize", "url": "https://github.com/Polymer/vulcanize/issues/332", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1975120904
feat: 🎸 add NFT module JIRA Link ✅ Closes: DA-913 Changelog / Description add endpoint to allow management of NFT collections, including issuing, redeeming and fetching NFT details Checklist - [ ] New Feature ? [ ] Updated swagger annotation (if API structure is changed) ? [ ] Unit Test (if possible) ? [ ] Updated the Readme.md (if required) ? :tada: This PR is included in version 4.0.0-alpha.6 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-11-02T22:11:07
2025-04-01T04:32:58.464405
{ "authors": [ "polymath-eric" ], "repo": "PolymeshAssociation/polymesh-rest-api", "url": "https://github.com/PolymeshAssociation/polymesh-rest-api/pull/217", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2573591900
docs: ✏️ update readme + export join creator params Description Update readme + fix doc generator warning Breaking Changes JIRA Link Checklist [ ] Updated the Readme.md (if required) ? /fast-forward :tada: This PR is included in version 26.0.0-alpha.2 :tada: The release is available on: npm package (@alpha dist-tag) GitHub release Your semantic-release bot :package::rocket: :tada: This PR is included in version 26.0.0-beta.1 :tada: The release is available on: npm package (@beta dist-tag) GitHub release Your semantic-release bot :package::rocket: :tada: This PR is included in version 26.0.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-10-08T15:48:03
2025-04-01T04:32:58.472127
{ "authors": [ "polymath-eric", "prashantasdeveloper" ], "repo": "PolymeshAssociation/polymesh-sdk", "url": "https://github.com/PolymeshAssociation/polymesh-sdk/pull/1329", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
962684342
Removes YouTube Premium fullscreen pop-up Rename the tweak to NoYTAnnoyances https://github.com/PoomSmart/NoYTPremium
gharchive/pull-request
2021-08-06T12:16:06
2025-04-01T04:32:58.481165
{ "authors": [ "PoomSmart", "nicknitewolf" ], "repo": "PoomSmart/YouAreThere", "url": "https://github.com/PoomSmart/YouAreThere/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
701286565
nom start errors - please help Failed to compile. ./node_modules/material-ui/styles/withStyles.js Module not found: Can't resolve '@babel/runtime/core-js/map' in '/home/test/downloads/nifty-game/dapp/node_modules/material-ui/styles' /home/test/downloads/nifty-game/dapp/scripts/start.js:11 throw err; ^ Can you help, please? can you fixed? that mean material-ui is not in the map in babel..
gharchive/issue
2020-09-14T17:23:33
2025-04-01T04:32:58.509373
{ "authors": [ "JoseMariani", "rocheston" ], "repo": "PortalNetwork/nifty-game", "url": "https://github.com/PortalNetwork/nifty-game/issues/73", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2126726853
fix: respect cache max age from configs Problem: max_age passed in configs is not getting respected. Due to this, the sent max age is not applied to cache sometimes. Example config: { "cache": {"mode": "simple", "maxAge": 60} } Using the above config, cache should be expired after 60 seconds. Should it be max_age? Yes. Fixed the example.
gharchive/issue
2024-02-09T09:02:33
2025-04-01T04:32:58.511373
{ "authors": [ "VisargD", "vrushankportkey" ], "repo": "Portkey-AI/gateway", "url": "https://github.com/Portkey-AI/gateway/issues/201", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1105198247
Min version of Posthog to run? What's the minimum version of Posthog this plugin runs on? When I tried on Posthog 1.29, I saw this ERROR in the plugin log: Task "Export historical events" not found for plugin "Migrator 3000 (Beta)" with config id [object Object] I have confirmed this should work on PostHog 1.29.0. Can you send me a screenshot of your system status on the instance you've enabled this plugin in? Sorry, I misspoke, I was on 1.27.0. I can try to upgrade to 1.29.0 or 1.29.1 as I've been having issues upgrading to 1.30.0 on Heroku. On 1.27.0 here's my system status: 1.29.1 is working for me! Cheers!
gharchive/issue
2022-01-16T21:57:49
2025-04-01T04:32:58.515898
{ "authors": [ "avk", "yakkomajuri" ], "repo": "PostHog/posthog-plugin-migrator3000", "url": "https://github.com/PostHog/posthog-plugin-migrator3000/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1305121279
Add param for tx value col, let user data override database, & bug fix Make the value column in tx data tables a parameter. This lets users pick firm/non-firm or custom names. Let user tx data override the database table Add docstring to agg_transmission_constraints table Add error messages Expand testing Codecov Report Merging #198 (51184ee) into develop (cc7da97) will decrease coverage by 0.65%. The diff coverage is 54.78%. @@ Coverage Diff @@ ## develop #198 +/- ## =========================================== - Coverage 71.58% 70.93% -0.65% =========================================== Files 16 16 Lines 3258 3354 +96 =========================================== + Hits 2332 2379 +47 - Misses 926 975 +49 Impacted Files Coverage Δ ...owergenome/run_powergenome_multiple_outputs_cli.py 0.00% <0.00%> (ø) powergenome/GenX.py 56.32% <8.89%> (-5.80%) :arrow_down: powergenome/generators.py 82.62% <69.70%> (-0.22%) :arrow_down: powergenome/util.py 69.52% <76.32%> (+0.79%) :arrow_up: powergenome/transmission.py 88.54% <80.95%> (+0.38%) :arrow_up: powergenome/eia_opendata.py 92.86% <100.00%> (+0.15%) :arrow_up: powergenome/load_profiles.py 78.20% <100.00%> (+1.03%) :arrow_up: powergenome/nrelatb.py 89.15% <100.00%> (-0.05%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 6ee7432...51184ee. Read the comment docs.
gharchive/pull-request
2022-07-14T18:00:20
2025-04-01T04:32:58.603341
{ "authors": [ "codecov-commenter", "gschivley" ], "repo": "PowerGenome/PowerGenome", "url": "https://github.com/PowerGenome/PowerGenome/pull/198", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
811206566
Update request - Determine which groups a user is a member of in a logon script Summary of the update request Link to the post in the old blog: https://devblogs.microsoft.com/scripting/how-can-i-determine-which-groups-a-user-belongs-to/ Description of what needs to be changed: Needs to show implicit remoting of AD cmdlets from the DC, then use get-ADPrincipalGroupMembership to get the group memberships. Use whoami.exe .upn, to get the date, then look the user up using the UPN. PR #45 adds this to the blog, so I am closing this issue.
gharchive/issue
2021-02-18T15:30:58
2025-04-01T04:32:58.606048
{ "authors": [ "doctordns" ], "repo": "PowerShell/Community-Blog", "url": "https://github.com/PowerShell/Community-Blog/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
715489049
Error in last two SuppressMessageAttribute examples in readme.MD In Suppressing rules in README.md, the last two examples omit the CheckID parameter. When used like this, the constructor for SuppressMessageAttribute throws an exception Eg. when using the example provided under 'Suppress violations in all the functions' in a script: [Diagnostics.CodeAnalysis.SuppressMessageAttribute('PSAvoidUsingWriteHost', Scope='Function', Target='*')] Param() Write-Output 'Hello world' When executing this script, PowerShell throws this error: Cannot find an overload for ".ctor" and the argument count: "1". The example starting on line 302 should read: [Diagnostics.CodeAnalysis.SuppressMessageAttribute('PSAvoidUsingWriteHost', '', Scope='Function', Target='*')] Param() and on line 307 [Diagnostics.CodeAnalysis.SuppressMessageAttribute('PSAvoidUsingWriteHost', '', Scope='Function', Target='start-b*')] Param() Thanks @hairyken you are correct, we will make this fix
gharchive/issue
2020-10-06T09:10:57
2025-04-01T04:32:58.634458
{ "authors": [ "SydneyhSmith", "hairyken" ], "repo": "PowerShell/PSScriptAnalyzer", "url": "https://github.com/PowerShell/PSScriptAnalyzer/issues/1592", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
610716632
[Feature Request] Better display of UNC paths Summary of the new feature/enhancement Currently, when the CWD is a UNC path, it is displayed like so PS Microsoft.PowerShell.Core\FileSystem::\\wsl$\Ubuntu\home\stan\projects\personal\yarntest> Would it be possible to display it as a normal path instead? like so PS \\wsl$\Ubuntu\home\stan\projects\personal\yarntest> Or is that something my profile broke? Proposed technical implementation details (optional) Not sure how complex this would be to make behave as you expect, but you can always use the New-PSDrive cmdlet to add a simpler address to the path, e.g.: New-PSDrive -Provider FileSystem -Root \\wsl$\Ubuntu\ -Name WSL Set-Location WSL: Well my request was mostly to improve the default since it does look a bit "wrong". Something you can add to your profile: function prompt { if ($null -eq $PWD.Drive -and $PWD.Provider.Name -eq 'FileSystem') { return "PS $($PWD.ProviderPath)> " } return "PS $($PWD.Path)> " } Maybe worth adding as a default. No worries, @thlac is working on a fix already. Bad bot bad bot bad bot
gharchive/issue
2020-05-01T11:58:20
2025-04-01T04:32:58.644798
{ "authors": [ "SeeminglyScience", "Stanzilla", "vexx32" ], "repo": "PowerShell/PowerShell", "url": "https://github.com/PowerShell/PowerShell/issues/12544", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
170208284
Remove dead code in StopJob.cs (line 164) Jason found this through code inspection. This code will never run because of the previous if statement and in fact the code is duplicated in the HandleStopJobCompleted() where it is correct. So we can simply remove the code at line 164. var parentJob = job as ContainerParentJob; if (parentJob != null && parentJob.ExecutionError.Count > 0) { foreach ( var e in parentJob.ExecutionError.Where( e => e.FullyQualifiedErrorId == "ContainerParentJobStopError")) { WriteError(e); } } If we have a "code cleanup" label at some point, this would fall under it. Hi @powershell, I just submit a PR (#5091) for this issue. Hmm, I seem to be missing something with this issue and the associated PR. HandleStopJobCompleted appears to be only used for Job2 instances and not Job instances. While the code is duplicated, removing it for Job instances doesn't appear to be correct.
gharchive/issue
2016-08-09T16:13:26
2025-04-01T04:32:58.646950
{ "authors": [ "PaulHigin", "dantraMSFT", "joeyaiello", "travisty-" ], "repo": "PowerShell/PowerShell", "url": "https://github.com/PowerShell/PowerShell/issues/1696", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1293148872
"Failed to create CoreCLR, HRESULT: 0x80070008" when using latest Prerequisites [X] Write a descriptive title. [X] Make sure you are able to repro it on the latest released version [X] Search the existing issues. [X] Refer to the FAQ. [X] Refer to Differences between Windows PowerShell 5.1 and PowerShell. Steps to reproduce In my build script I have a step that spawns a docker container based on the powershell:latest image that fails with the error in the title. Here is the command executed on the build server docker run --rm -e AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY -e COMMIT_SHA1=$CIRCLE_SHA1 --network project_default mcr.microsoft.com/powershell pwsh -command ' Install-Module -Name AWS.Tools.Installer -Force Install-AWSToolsModule AWS.Tools.S3 -Force -CleanUp Invoke-WebRequest -Uri http://entrypoint-api/docs/v1.0/schema.json -OutFile /tmp/schema.json Write-S3Object -File /tmp/schema.json -BucketName "some_bucket" -Key "some_key/$env:COMMIT_SHA1.json" -AccessKey $env:AWS_ACCESS_KEY_ID -SecretKey $env:AWS_SECRET_ACCESS_KEY ' The command has worked without any issue until few days ago. Then we started getting the error specified above: "Failed to create CoreCLR, HRESULT: 0x80070008" Unable to find image 'mcr.microsoft.com/powershell:latest' locally latest: Pulling from powershell 018f9d1d: Pulling fs layer Digest: sha256:844fb73db69ed0bbe82d159d4ba50aea06ef4d42bda766c52f2928683f5c8d09 Status: Downloaded newer image for mcr.microsoft.com/powershell:latest Failed to create CoreCLR, HRESULT: 0x80070008 Exited with code exit status 137 CircleCI received exit code 137 After some testing, I noticed that the issue is only happening when using latest which is pointing at ubuntu-22.04. If I use 7.2.0-debian-10, 7.2-debian-10, 7.2-alpine-3.14, alpine-3.14, ubuntu-22.04 it works without any problem. In issue #13166 it was suggested to set COMPlus_EnableDiagnostics to 0. This didn't help. Expected behavior Being able to run a container with latest image of powershell. Actual behavior Container can't start Error details Unable to find image 'mcr.microsoft.com/powershell:latest' locally latest: Pulling from powershell 018f9d1d: Pulling fs layer Digest: sha256:844fb73db69ed0bbe82d159d4ba50aea06ef4d42bda766c52f2928683f5c8d09 Status: Downloaded newer image for mcr.microsoft.com/powershell:latest Failed to create CoreCLR, HRESULT: 0x80070008 Exited with code exit status 137 CircleCI received exit code 137 Environment data Don't have it. Visuals No response The tag ubuntu-22.04 doesn't work for me either. (no surprise given that latest points to the same version) Status: Downloaded newer image for mcr.microsoft.com/powershell:ubuntu-22.04 mcr.microsoft.com/powershell:ubuntu-22.04 Failed to create CoreCLR, HRESULT: 0x80070008 Process exited with code 137 What version of powershell are you using? I'll look into replicating this my side and resolving soon! @anamnavi the faulty command is launched in a mcr.microsoft.com/powershell container. Basically it's a step of a job in a CircleCI build. The job uses mcr.microsoft.com/powershell as executor and itself spawns another container based off the same image. In our tests we didn't need to change the outer container but just the one we were using to execute those commands. I'm not the OP, but I'm hitting the same issue (with a simpler script, but the same container). My host is a relatively ancient Debian box running docker 19.03. The host does have ~3GB free RAM and running that command on the host just reports a max long value, as does running it inside the container. Running the full powershell command on a Windows 10 host runs without issues, so it does seem to be something host-specific. Using a different powershell container also works on my ancient host. @anamnavi Here is what I got in the host container root@d1e24fe67e4c:~/project# docker run --rm mcr.microsoft.com/powershell Failed to create CoreCLR, HRESULT: 0x80070008 root@d1e24fe67e4c:~/project# docker-compose up -d [...] root@d1e24fe67e4c:~/project# /bin/sh -c 'grep hierarchical_memory_limit /sys/fs/cgroup/memory/memory.stat' hierarchical_memory_limit 4320133120 root@d1e24fe67e4c:~/project# docker-compose down [...] root@d1e24fe67e4c:~/project# /bin/sh -c 'grep hierarchical_memory_limit /sys/fs/cgroup/memory/memory.stat' root@d1e24fe67e4c:~/project# hierarchical_memory_limit 4320133120 @anamnavi Here it is log.txt The log shows a failure loading /opt/microsoft/powershell/7-lts/libcoreclr.so which is the CoreCLR itself. @Kralizek Digging into the error HRESULT: 0x80070008 points an issue with a syscall being blocked due to SECCOMP in an container environment. For example, https://github.com/dotnet/runtime/issues/1634 shows that this error happened when .NET called sched_setaffinity to set thread's CPU affinity mask in a Snap confined environment that doesn't have the process-control interface. You will want to read this article: https://www.linuxserver.io/blog/anatomy-of-an-issue, which points out a similar problem due to using old versions of Docker engine. Their resolution may be helpful to you: https://docs.linuxserver.io/faq#jammy. @daxian-dbw thanks for the reply. Do you have an idea why older images and images based off different distributions like alpine don't suffer of the same problem? I'm not sure, but article I referenced to has the following that may be an explanation: The problem is that if a new syscall gets introduced but Docker doesn't have it included in the profile then your containers can't use it and if, as in this case, the syscall replaces an older one, new distros using it will break previously functional containers. Does either of the resolutions suggested at https://docs.linuxserver.io/faq#jammy work for you? @daxian-dbw I'll try when I get back to work. Currently on vacation deploy: resources: limits: pids: 1 accidently had the same pids twice in my docker-compose.yml, causing this error.
gharchive/issue
2022-07-04T13:05:44
2025-04-01T04:32:58.662894
{ "authors": [ "FelixSFD", "Kralizek", "Pentadome", "TravisEz13", "anamnavi", "daxian-dbw", "uecasm" ], "repo": "PowerShell/PowerShell", "url": "https://github.com/PowerShell/PowerShell/issues/17624", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
340355311
PSCore - Where is the PowerShell Core GA version(s)? On the releases page, I'm looking for the PowerShell Core 6.0 GA version. But, I can't find it. So, which of the non-Preview are GA releases. Or, should I assume that all non-Preview are actually GA's? I think this should be address better in the Release Notes. My organization only support GA release in their products. Preview releases are marked as preview, everything else is GA. But I agree, it is not very easy to find out as one has to scroll down quite a bit. The latest GA release is 6.0.2 here Maybe it would be better to display the latest GA always at the very top or at least at 2nd place? We have links on main page https://github.com/PowerShell/PowerShell#get-powershell @iSazonov I see now. Of course... "Stable" = "GA". Understood! :) @iSazonov Forgot to mention, my main point is more about the "Release Notes" not mention when the product release reached GA status. I think it would be a "nice" to include in the notes. :) I believe this could be added in release check list. @iSazonov, That would be Great!! :) @SteveL-MSFT Is the request about "GA in Release Notes" addressed? Can we close the issue? I believe this can be closed
gharchive/issue
2018-07-11T18:27:40
2025-04-01T04:32:58.668922
{ "authors": [ "MaximoTrinidad", "SteveL-MSFT", "bergmeister", "iSazonov" ], "repo": "PowerShell/PowerShell", "url": "https://github.com/PowerShell/PowerShell/issues/7271", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
378875524
Start-Transcript sporadically loses exe output About half the time there is not output result in the $logFile even though on the console there was plently of output. The other half of the time there is output in the $logFile. It seems very random on if the output works or not. Actually I would say a little less than half the time the helm ouput does NOT appear in the $logFile. I'm using PowerShell core 6.1 on Windows 10. Start-Transcript -Force -IncludeInvocationHeader -Path $logFile try { helm upgrade --install --wait --force --debug -f ./helm/zookeeper/values.yaml platform-zookeeper ./helm/zookeeper } finally { Stop-Transcript } This also happen with kubectl commands as well. This appears to be a similar issue. http://web.archive.org/web/20130608164153/http://connect.microsoft.com/PowerShell/feedback/details/315875/unable-to-capture-all-session-output-into-a-transcript Do you get a random output in transcript file run by run?
gharchive/issue
2018-11-08T19:09:38
2025-04-01T04:32:58.671775
{ "authors": [ "AceHack", "iSazonov" ], "repo": "PowerShell/PowerShell", "url": "https://github.com/PowerShell/PowerShell/issues/8215", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
218263174
Incorrect Stringbuilder usage StringBuilder doesn't have a constructor that takes a char. In this case, the char is only used to inadvertently define capacity when the intent is to populate the StringBuilder with a newline. Addresses https://github.com/PowerShell/PowerShell/issues/3457 Looked through the code for StringBuilder and only found this one occurrence of incorrect usage I'm closing this PR as it is a duplicate of #3458
gharchive/pull-request
2017-03-30T17:01:00
2025-04-01T04:32:58.673755
{ "authors": [ "SteveL-MSFT", "daxian-dbw" ], "repo": "PowerShell/PowerShell", "url": "https://github.com/PowerShell/PowerShell/pull/3461", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
214486198
Register-PSRepository in non-interactive mode? Team, I'm using Register-PSRepository for automation on internal build system. On PS 5.1. the cmdlet prompts and I get the following message(from the build logs): Exception calling "ShouldContinue" with "2" argument(s): "Windows PowerShell is in NonInteractive mode. Read and Prompt functionality is not available." The cmdlet doesn't have -Force/-Confirm parameters and it doesn't respect the $ConfirmPreference variable. Is there a workaround for this? Thanks, Nedko @bmanikm That command doesn't seem to download nuget.exe. In my environment, I have been running that command, but I don't get the error about non-interactive mode until I get to the publish-module step, which fails because it's trying to prompt about downloading nuget.exe. @godefroi Correct, Install-PackageProvider only installs NuGet provider, and it doesn't download nuget.exe. NuGet.exe is required only during Publish-Module or Publish-Script operations. Please take a look at https://github.com/PowerShell/PowerShellGet/blob/development/tools/build.psm1#L72 to understand the logic for bootstrapping NuGet.exe in automation scripts. This is extremely unsatisfying. @bmanikm's suggestion results in downloading NuGet in every iteration/build/deployment! We end up needing something like this in our deployment scripts? if(!(Get-PackageProvider NuGet)) { Install-PackageProvider Nuget -ForceBootstrap -Force } It's a bit frustrating that I need to write so many lines for that, and then I have the same problem when installing modules. In automation scripts, please use the below simple command to bootstrap the NuGet provider. This bootstrap the NuGet provider only if it is not available on the local machine. Get-PackageProvider NuGet -ForceBootstrap In non-interactive mode this step requires additional privileges, at least account under which is running this commandlet has to have SeAssignPrimaryTokenPrivileges and SeServiceLogonRight. I've just extended this script by adding SeAssignPrimaryTokenPrivilege options. Currently Register-PSRepository works fine for me from our Chef recipes. This should be reopened @bmanikm - I have similar use case with Docker (installing pswh on docker image behind proxies) and it would be very welcomed Here is my Recipe how to upgrade PowerShellGet under the docker Write-Host Write-Host "Updating PowerShellGet ..." -ForegroundColor Green # Original way to install PowerShellGet does not work under docker or without user confirms #Install-Module PowerShellGet -Repository PSGallery -Force -Confirm -AllowClobber -Scope CurrentUser # Here is the alternative install script $nugetPackageProvider = $null $nugetPackageProvider = Get-PackageProvider NuGet -ErrorAction SilentlyContinue if (!$nugetPackageProvider) { $nugetPackageProvider = Install-PackageProvider Nuget -ForceBootstrap -Force Write-Host "Installed NuGet package provider version $($nugetPackageProvider.Version)" } Write-Host "Installing PowerShellGet module..." $userModulesPath = "$([Environment]::GetFolderPath("MyDocuments"))/WindowsPowerShell/Modules" New-Item -ItemType Directory -Path $userModulesPath -Force | Out-Null Save-Module PowerShellGet -Path $userModulesPath Remove-Module PowerShellGet, PackageManagement -Force Import-Module PowerShellGet $psGetModule = Get-Module PowerShellGet $installedVersion = $psGetModule.Version I was having a similar issue when trying to run a PowerShell script in my release pipeline that needed access to a module in our private Azure DevOps Artifacts feed, so I was attempting to do: Register-PSRepository -Name MyRepoName -SourceLocation 'https://pkgs.dev.azure.com/[MyOrganization]/_packaging/[MyFeed]/nuget/v2' -InstallationPolicy Trusted Install-Package MyPrivatePowerShellModule -InstallUpdate -AllowClobber However, when the script runs on the MS hosted agent, it ends up in an infinite wait loop outputting: "C:\windows\System32\WindowsPowerShell\v1.0\powershell.exe" -NoLogo -NoProfile -NonInteractive -ExecutionPolicy Unrestricted -Command ". 'D:\a_temp\527651b0-116e-46aa-a599-141b4dfc9899.ps1'" Registering PSRepository [Minimal] [CredentialProvider]DeviceFlow: https://pkgs.dev.azure.com/[MyOrganization]/_packaging/[MyFeed]/nuget/v2 [CredentialProvider]To sign in, use a web browser to open the page https://microsoft.com/devicelogin and enter the code EB64QSSD6 to authenticate.##[section]Starting: PowerShell Script I came across this feedback post though and was able to get it working with the following code (ensuring $PersonalAccessToken contains a PAT with permissions to read from the provider): $pat = ConvertTo-SecureString '$PersonalAccessToken' -AsPlainText -Force $credential = New-Object System.Management.Automation.PSCredential 'Username@DoesNotMatter.com', $pat Install-PackageProvider NuGet -Scope CurrentUser -Force > $null if ($null -eq (Get-PSRepository -Name 'MyRepoName' -ErrorAction SilentlyContinue)) { Register-PSRepository -Name 'MyRepoName' -SourceLocation 'https://pkgs.dev.azure.com/[MyOrganization]/_packaging/[MyFeed]/nuget/v2' -InstallationPolicy Trusted -Credential $credential } Install-Module -Name 'MyPrivatePowerShellModule' -Repository 'MyRepoName' -Force -AllowClobber -Credential $credential Import-Module -Name 'MyPrivatePowerShellModule' -Force The trick was to first do Install-PackageProvider NuGet -Scope CurrentUser -Force, and then also to use Personal Access Token credentials. I hope this helps anybody else who stumbles across this issue.
gharchive/issue
2017-03-15T18:35:03
2025-04-01T04:32:58.692777
{ "authors": [ "Jaykul", "Vit0hA", "bmanikm", "deadlydog", "dmitriyse", "dmpe", "f0rt", "godefroi" ], "repo": "PowerShell/PowerShellGet", "url": "https://github.com/PowerShell/PowerShellGet/issues/101", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
171257845
update-markdownhelp should update the help file name in each of the cmdlets MD and also update the GUID in the module landing page Steps to reproduce Create MDs from a MAML file that doesn't have a standard help file name. Run an update on MDs with the module installed Expected behavior Cmdlet MDs should have the external help file name to be updated with original xml help file name from the module. Also, the landing page should have the GUID updated if not already there. Actual behavior The external help file name in the Cmdlets MD remains to be the same as the MAML file that was used to convert to MDs. This poses a problem when we cab the file and release the updated content changes will not take into effect. v0.6.1 @sankethka could you split this into two issues? These are separate asks. resolved in 0e7749df
gharchive/issue
2016-08-15T20:45:44
2025-04-01T04:32:58.708165
{ "authors": [ "jongeller", "sankethka" ], "repo": "PowerShell/platyPS", "url": "https://github.com/PowerShell/platyPS/issues/193", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
163964292
BLOCKER: PSSA Errors Need to Be Fixed Errors from PS script analyzer are causing the tests to fail like this. This resource module cannot be released until these errors are fixed or suppressed and the tests are passing. #12
gharchive/issue
2016-07-05T22:55:25
2025-04-01T04:32:58.712032
{ "authors": [ "Indhukrishna", "kwirkykat" ], "repo": "PowerShell/xAzurePack", "url": "https://github.com/PowerShell/xAzurePack/issues/13", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1943743232
Superhero Penance I image not showing Hi, I found that some superhero images are not showing, like Penance I , Atom and Goliath. Please assign it to me under hacktoberfest 2023 label. I can work on it. Found that the Superhero API has some Heros' images missing. Server is responding with a 404.error message. I don't know how to work around this situation now. Please assign it to someone who can work on it.
gharchive/issue
2023-10-15T06:22:39
2025-04-01T04:32:58.722554
{ "authors": [ "Shybymh" ], "repo": "Prajwal0225/My-Favorite-SuperHero", "url": "https://github.com/Prajwal0225/My-Favorite-SuperHero/issues/187", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
912731849
Adding the small screen fonts view and changing vote rank Well detailed description of the change : I worked on the changing the position of the icons on small screens and also I changed the vote rank so if it is 0 it does not say - 0 Context of the change : - Why is this change required ? Does it solve a problem ? (please link the issue) Type of change : [ ] Bug fix [ ] New feature [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) Preview (Screenshots) : If it is possible, please link screenshots of your changes preview ! Checklist: [ ] I have read the CONTRIBUTING document. [ ] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [ ] All new and existing tests passed. Reviewers @PraveenMalethia @PraveenMalethia I dont know if I changed the vote rank to not say -0 the correct way. There might be a simpler way but this works.
gharchive/pull-request
2021-06-06T08:59:52
2025-04-01T04:32:58.731904
{ "authors": [ "srikarkandula" ], "repo": "PraveenMalethia/mumble-vue", "url": "https://github.com/PraveenMalethia/mumble-vue/pull/42", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1311597055
Updated prefect_dbt.cloud.jobs import Updated prefect_dbt.cloud.jobs import to reflect accurate name for prefect_dbt.cloud.jobs import() Summary Relevant Issue(s) Checklist [ ] Summarized PR's changes in CHANGELOG.md Thanks! The tests are fixed in https://github.com/PrefectHQ/prefect-dbt/pull/22 so that needs to go in first.
gharchive/pull-request
2022-07-20T17:49:14
2025-04-01T04:32:58.782903
{ "authors": [ "ahuang11", "belasobral93" ], "repo": "PrefectHQ/prefect-dbt", "url": "https://github.com/PrefectHQ/prefect-dbt/pull/31", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2287833274
Remove orphan tabs following deletion of old controllers Questions Answers Description? While upgrading to PrestaShop v9, remove tabs who lost their parent relation with the previous row deletion. this cleans the table and fixes an issue reported while opening the Stock page. Type? bug fix BC breaks? Nope Deprecations? Nope Fixed ticket? Fixes https://github.com/PrestaShop/PrestaShop/issues/35879 Sponsor company PrestaShopCorp How to test? Upgrade from PS 8 to 9, then open the Stock page. It should load without error. Hi @Quetzacoalt91 Thank you for your PR, I tested it on : 8.1.5 to 9.0.0, that works without any exception but when I try to upgrade from 8.0.5 to 9.0.0, that what I have : recording(202).webm Waiting for your correction Seen with @AureRita: Because the stack trace relates to a module it is interesting to see if the error can be reproduced on the dev branch as well.
gharchive/pull-request
2024-05-09T14:38:07
2025-04-01T04:32:58.973322
{ "authors": [ "AureRita", "Quetzacoalt91" ], "repo": "PrestaShop/autoupgrade", "url": "https://github.com/PrestaShop/autoupgrade/pull/698", "license": "AFL-3.0", "license_type": "permissive", "license_source": "github-api" }
1749438023
Deploy kanbanbot v1 The first kanbanbot feature are done. We have to deploy the main branch using the automatic deploiement based en github flow and docker. Kanban bot deployed
gharchive/issue
2023-06-09T08:46:47
2025-04-01T04:32:58.978042
{ "authors": [ "WahbiPS", "lartist" ], "repo": "PrestaShop/kanbanbot", "url": "https://github.com/PrestaShop/kanbanbot/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
691413822
show staff affiliates on affiliates page #192 see checklist on #192 for a list of changes on this PR. Codecov Report Merging #196 into develop will increase coverage by 0.00%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## develop #196 +/- ## ======================================== Coverage 98.08% 98.09% ======================================== Files 82 82 Lines 2767 2778 +11 ======================================== + Hits 2714 2725 +11 Misses 53 53
gharchive/pull-request
2020-09-02T20:54:29
2025-04-01T04:32:58.992543
{ "authors": [ "codecov-commenter", "thatbudakguy" ], "repo": "Princeton-CDH/cdh-web", "url": "https://github.com/Princeton-CDH/cdh-web/pull/196", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1398308609
Slamp checks Merge checking passes into master The main functionality is not yet ready. Please checkpick the repl related updates and only PR those.
gharchive/pull-request
2022-10-05T20:00:26
2025-04-01T04:32:59.001781
{ "authors": [ "vgene", "yebinchon" ], "repo": "PrincetonUniversity/cpf", "url": "https://github.com/PrincetonUniversity/cpf/pull/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
374646700
NEON configuration language support NEON is a configuration language similar to YAML, with the notable addition of so-called "NEON entities". NEON is used mainly by the Nette PHP framework. See #2140.
gharchive/pull-request
2018-10-27T13:45:05
2025-04-01T04:32:59.002756
{ "authors": [ "RunDevelopment", "jahudka" ], "repo": "PrismJS/prism", "url": "https://github.com/PrismJS/prism/pull/1600", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2090922710
Add background to main-page and improve its CSS I visited the website and think the css of front page could be improved and background could be added. Please assign it to me under JWOC 24 @Priyanshi662 please assign me under JWOC 24. @Priyanshi662 thank you so much. @Priyanshi662
gharchive/issue
2024-01-19T16:52:14
2025-04-01T04:32:59.021620
{ "authors": [ "KhushiAgarwal22" ], "repo": "Priyanshi662/FunFusion", "url": "https://github.com/Priyanshi662/FunFusion/issues/57", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1997810628
Input value goes back to initial values after reopening the program Code: - I_x =? {48} in^4 This is the original value the I_x was set, when the program is executed and new values are replaced the calculations are done for the updated values, however when I close and reopen the program the I_x value goes back to 48 instead of last value. It would be useful if it's replaced by the last user input value. Hi! Did you try to save the file with the new values? You need to save it to persist the new values for the next time. Yeah! it works, thank you :)
gharchive/issue
2023-11-16T21:23:56
2025-04-01T04:32:59.051602
{ "authors": [ "Proektsoftbg", "strsthapa" ], "repo": "Proektsoftbg/Calcpad", "url": "https://github.com/Proektsoftbg/Calcpad/issues/243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1582996934
Update Wiki for Stripe Need to update Wiki for Stripe Take out Apple Pay for Checkout https://github.com/ProgramEquity/amplify/blob/main/server/routes/api/checkout.js From @masmei https://stripe.com/docs/api/payment_intents/object#payment_intent_object-setup_future_usage ^^ this is on saving customer data We can store data from payments object in our DB @manishapriya94 could you please elaborate what we have to do for this issue? From what I gather, we just have to remove references to Apple Pay from the wiki page and rename it?
gharchive/issue
2023-02-13T20:13:22
2025-04-01T04:32:59.055739
{ "authors": [ "iofall", "manishapriya94" ], "repo": "ProgramEquity/amplify", "url": "https://github.com/ProgramEquity/amplify/issues/492", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1411832362
CVE-2015-5237 Grype scan reported CVE-2015-5237. Please investigate and update .grype.yaml. A false-positive scan result from grype. mc team confirmed that mc is using the latest version of protobuf which is versioned v1.28.1. This new versioning number made grype think that an older/vulnerable protobuf is used in mc. Closing as not an issue and updating grype ignore list.
gharchive/issue
2022-10-17T16:00:32
2025-04-01T04:32:59.074407
{ "authors": [ "mocsharp" ], "repo": "Project-MONAI/monai-deploy-workflow-manager", "url": "https://github.com/Project-MONAI/monai-deploy-workflow-manager/issues/515", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2697742374
Return information about specified user in report Portal: api spec change API deprecation: https://reportportal.io/docs/api/service-api/versions/5.11/get-user-using-get/ API new: https://reportportal.io/docs/api/service-api/get-user Connected issue: https://github.com/ProjectAlita/projectalita.github.io/issues/512
gharchive/pull-request
2024-11-27T09:02:36
2025-04-01T04:32:59.090796
{ "authors": [ "plageat" ], "repo": "ProjectAlita/application-tools", "url": "https://github.com/ProjectAlita/application-tools/pull/84", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1521449795
Dan/2023/01/remove unused functions Addresses the enumerated unused functions in https://github.com/spearbit-audits/review-seaport-1.2/issues/97. Note that @Saw-mon-and-Natalie left a todo two days ago to add others. Update about additional enumerations in the above Spearbit Issue that haven't yet been added to this PR: LowLevelHelpers.sol#L25 LowLevelHelpers.sol#L112 ZoneInteraction.sol#L227 Update about additional enumerations in the above Spearbit Issue that haven't yet been added to this PR: LowLevelHelpers.sol#L25 LowLevelHelpers.sol#L112 ZoneInteraction.sol#L227 Those were all removed in other PRs, I believe. True. These were in the changes happening during the early stages of the audit: https://github.com/ProjectOpenSea/seaport/commit/4de24dfd64b56a966d348405cedf3c4ac7e4450e Closing as we decided to keep all the helpers in PointerLibraries even if they're not all currently being used.
gharchive/pull-request
2023-01-05T21:27:22
2025-04-01T04:32:59.116222
{ "authors": [ "0age", "DJViau", "JustDravee" ], "repo": "ProjectOpenSea/seaport", "url": "https://github.com/ProjectOpenSea/seaport/pull/838", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1903687006
Fix typo in README.md right at the end Godot as written as "GoDot", which is incorrect There's also one in the requirements.txt file at "tokenzie"(should be "tokenize") and "confiugration" https://github.com/ProjectUnifree/unifree/blob/8a5456fb10b24954c3e4b308154cc8ef985a728a/configs/godot.yaml#L3 also calls it "GoDot".
gharchive/pull-request
2023-09-19T20:34:19
2025-04-01T04:32:59.136406
{ "authors": [ "NotHyper-474", "aragubas", "gsteinLTU" ], "repo": "ProjectUnifree/unifree", "url": "https://github.com/ProjectUnifree/unifree/pull/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1952194600
Make Watchy_GSR compatible? First of all, I love your watchface! I just wish it was compatible with Watchy_GSR. Do you know about Watchy_GSR? (https://github.com/GuruSR/Watchy_GSR) I've seen many people in the watchy discord interested in your watchface, and a few that were interested in it being converted to Watchy GSR compatibility. If I could code myself, I would try to convert it, but I figured you would be the one that knows your code best! Just thought maybe you didn't know about Watchy_GSR and thought I would let you know there is some interest in it's conversion. Feel free to close this "issue" that isn't really an issue, just wanted to communicate this idea to you. This would be awesome
gharchive/issue
2023-10-19T12:55:35
2025-04-01T04:32:59.138608
{ "authors": [ "Izaic", "grnqrtr" ], "repo": "Prokuon/watchy-starfield", "url": "https://github.com/Prokuon/watchy-starfield/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
438877827
Scrolling down the mod list shows invisible items The list seems to be empty after 2nd page but the items are still clickable. Have about 50 mods installed now. After searching with pattern scrolling not works anymore. it seems to only show about 20 mods for me, how about you Knito? @Hellzbellz123 Same for me. Thanks for seconding me. I really thought that only happens to me. yeah it seems to be a bug, if you happen to remember mod names you can filter the list to show ones past the 20 limiut Exactly.
gharchive/issue
2019-04-30T16:54:21
2025-04-01T04:32:59.149232
{ "authors": [ "Hellzbellz123", "Knito58" ], "repo": "Prospector/ModMenu", "url": "https://github.com/Prospector/ModMenu/issues/34", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
443660254
RawFileReader update? I’m not sure which version of Thermo’s RawFileReader is currently in use by ProteoWizard (I assume 4.0.26), but they do have 2 newer versions (4.0.89 and 5.0.6) available on their sharepoint site, with release notes mentioning unspecified “bug fixes in the libraries” and a change to a single set of binaries (compared to the 3 previous sets for each of Windows, MacOS, and Linux). Update is merged, I don't see any need for this to remain open.
gharchive/issue
2019-05-14T00:42:39
2025-04-01T04:32:59.157808
{ "authors": [ "FarmGeek4Life" ], "repo": "ProteoWizard/pwiz", "url": "https://github.com/ProteoWizard/pwiz/issues/541", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
238397463
Don't allow a user to use localhost as a connection address Tons of issues with people using localhost and then being unable to connect to the SQL socket, require using an IP which avoids that issue. this is only force localhost to 127.0.0.1 in assing config or block this parameter? (question for.. know?) Technically if it is set by a plugin I don't think there is much you can do there, though making a huge warning stating you must use the external ip instead of localhost on the docs should make a difference. This is for panel DB connections, not plugins. For now I've put a warning on the environment setup for database connections, and the command will also test connections, so I think I'm going to close this as 8722571 addresses it for now.
gharchive/issue
2017-06-25T18:34:23
2025-04-01T04:32:59.354539
{ "authors": [ "DaneEveritt", "Doc94", "Polarcraft" ], "repo": "Pterodactyl/Panel", "url": "https://github.com/Pterodactyl/Panel/issues/518", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2322752816
Accessing quarto metadata programmatically via QuartoNotebookWorker Now that we have the expand extension pipeline, I was thinking that downstream packages might want to expand differently depending on, for example, the output format that quarto renders to. I think that's different from the loading hooks we currently have. So I wonder how we could make the full quarto metadata accessible, like you can access it from knitr with knitr::pandoc_to(), as one example. Stored in https://github.com/PumasAI/QuartoNotebookRunner.jl/blob/85f65102f63c2c4eac98c5cb51eed14fc2948535/src/QuartoNotebookWorker/src/NotebookState.jl#L8, so we can make an official API for that I guess. Yes, although there's currently no open way to access anything QNW-related, is there? The extension methods hook in at package load time, you couldn't easily reach that from the notebook itself. And I think we were holding off on adding any API into the notebook workspace due to it not being available in a REPL. you couldn't easily reach that from the notebook itself. And I think we were holding off on adding any API into the notebook workspace due to it not being available in a REPL. Would only be accessed in 3rd-parth package extensions, not directly in a notebook, e.g. module PackageNameQuartoNotebookWorkerExt import PackageName import QuartoNotebookWorker function QuartoNotebookWorker.expand(::PackageName.Foo) opts = QuartoNotebookWorker.options() # hypothetical public API, used at runtime. # ... end end Right, ok, so it should stay that way then. One could not develop such functionality in scripts is the only drawback I can see. One could not develop such functionality directly within notebooks is the only drawback I can see. Nope, you can't. Revise ideally should just work with this fine and functionality can just be developed directly in the extension file. If that doesn't work then we should make it work. Yes for "serious" development that's true, it just adds overhead if you want to quickly try out things. I would have liked an "escape hatch" of some kind. Actually, maybe that's already possible by requiring the right UUID manually. Just thinking out loud here, I think the extension is in general the right place to do this kind of stuff. Actually, maybe that's already possible by requiring the right UUID manually. Given the way the LOAD_PATH is done you should just be able to always do using QuartoNotebookWorker in a cell as an escape hatch if needed. (Worked for me just now).
gharchive/issue
2024-05-29T08:53:26
2025-04-01T04:32:59.379974
{ "authors": [ "MichaelHatherly", "jkrumbiegel" ], "repo": "PumasAI/QuartoNotebookRunner.jl", "url": "https://github.com/PumasAI/QuartoNotebookRunner.jl/issues/139", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
430995219
No attempts for a long time I'm having this issue inly for exact user, cause I see that for some other usernames this bruter works well I was facing the same issue, then I found out that my other server is bruting the same account. so just terminated other one (guess multiple login attempts cause errors) They use the same proxies, so it’s not a good idea to use multiple servers to attack your account
gharchive/issue
2019-04-09T14:20:08
2025-04-01T04:32:59.387093
{ "authors": [ "Pure-L0G1C", "ali3shah", "sshkrv" ], "repo": "Pure-L0G1C/Instagram", "url": "https://github.com/Pure-L0G1C/Instagram/issues/88", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1636843715
update redirects for opengov changes Add redirects to the CN mkdocs file for the opengov related changes Based on these PRS: https://github.com/PureStake/moonbeam-mkdocs/pull/105/files https://github.com/PureStake/moonbeam-mkdocs/pull/106/files Approved both OpenGov related PRs
gharchive/pull-request
2023-03-23T04:47:30
2025-04-01T04:32:59.388820
{ "authors": [ "albertov19", "eshaben" ], "repo": "PureStake/moonbeam-mkdocs", "url": "https://github.com/PureStake/moonbeam-mkdocs/pull/112", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
1007171196
Android build error when including in pubspec.yaml This plugin causes a build error on Android when simply being included in pubspec.yaml, without any other changes being made to my project: Execution failed for task ':app:checkDebugDuplicateClasses'. > A failure occurred while executing com.android.build.gradle.internal.tasks.CheckDuplicatesRunnable > Duplicate class org.threeten.bp.Clock found in modules jetified-threetenbp-1.4-no-tzdb.2-no-tzdb (org.threeten:threetenbp:1.4.2) and jetified-threetenbp-1.4.2 (org.threeten:threetenbp:1.4.2) > ... (hundreds of similar errors) My minSdkVersion is 24, above the 21 required in the README file. Hi, it's a problem with your gradle config, try with the suggestions here or here or here and let me know if it works. Simone
gharchive/issue
2021-09-25T19:03:04
2025-04-01T04:32:59.397158
{ "authors": [ "palkerecsenyi", "sstasi95" ], "repo": "PurpleSoftSrl/sumup_flutter_plugin", "url": "https://github.com/PurpleSoftSrl/sumup_flutter_plugin/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1425432426
🛑 Gitea is down In 27c1e0e, Gitea (https://git.pxlbuzzard.com) was down: HTTP code: 503 Response time: 103 ms Resolved: Gitea is back up in 620150f.
gharchive/issue
2022-10-27T10:50:14
2025-04-01T04:32:59.411755
{ "authors": [ "PxlBuzzard" ], "repo": "PxlBuzzard/upptime", "url": "https://github.com/PxlBuzzard/upptime/issues/685", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
707561817
Syntax Issue File "filemovergui.py", line 45 ".wmv",".mov",".qt",".rm",]"CodePerfectPlus", ^ SyntaxError: invalid syntax When I run python3 filemovergui.py Thanks, @Raavtube For pointing out Bug. It's now fixed.
gharchive/issue
2020-09-23T17:50:07
2025-04-01T04:32:59.413465
{ "authors": [ "Raavtube", "codePerfectPlus" ], "repo": "Py-Contributors/FileMoverGUI", "url": "https://github.com/Py-Contributors/FileMoverGUI/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
712570224
Find the Missing Number Before jumping into Pr first comment for assign. Problem statement : Find the Missing Number Probelm Link : https://edabit.com/challenge/7YaJhC4terApw5DFa Choose one Cateorgy - Very Easy, Easy, Medium, Hard, Very Hard, Expert question category : Easy Before submitting a PR please Check some details. [ ] fileName formet should be in camelcase example- returnTheNextNumber.py [ ] Choose label for Problem , example- "good first issue", "easy" [ ] In file there should be problem statement and link to problem /* Problem-Task : Create a function that takes an array of numbers between 1 and 10 (excluding one number) and returns the missing number. * Problem Link : https://edabit.com/challenge/7YaJhC4terApw5DFa */ int main() { int bla; return 0; } This repository is vaild for HacktoberFest2020 Read before making PR - How to contribute on Github Good Luck, Can you assign me this issue Can you assign this to me? @pranav230 can u assign this to me Please assign me this. @pranav230 can you assign this to me .... i am working on this issues ,I will create PR soon can you please assign this to me @pranav230 I Added FindMissingNumber.cpp .. please check my code and review my PR PR of assign Person will review.. sorry
gharchive/issue
2020-10-01T07:22:40
2025-04-01T04:32:59.419362
{ "authors": [ "NidhiNivedita6", "VaderX", "aayushi-droid", "pranav230", "yasharth291", "zahinekbal" ], "repo": "Py-Droid/Edabit-Solutions", "url": "https://github.com/Py-Droid/Edabit-Solutions/issues/112", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1347583624
Make options changeable after codec or (de)muxer has been opened Overview Currently, as best I can tell, it is only possible to set options once: when a codec or (de)muxer is opened. There are cases where one might want to change some option afterward. My use case is changing -crf in the libx264 encoder during encoding based on user preference and changing network conditions. Existing FFmpeg API The av_opt_set*() in libavutil/opt.h. av_opt_get*() are probably also prudent to implement. Doxy link: https://ffmpeg.org/doxygen/trunk/group__avoptions.html Expected PyAV API The av_opt functions are very generic, so the easiest route would probably be to guarantee that all relevant classes have self.ptr exposed that is guaranteed to be passable to the av_opt functions. This could be hidden from the user so that they only have to pass a CodecContext, OutputContainer or InputContainer. Example based on aiortc's h264.py: from av.avutil import av_opt_set_int, av_opt_get_int class H264Encoder(Encoder): # [...] def set_crf(self, crf: int) av_opt_set_int(self.codec, "crf", crf) # search_flags could be optional, default to 0 def get_crf(self) -> int: return av_opt_get_int(self.codec, "crf") Errors returned by av_opt_get_int() could be raised as Exceptions, which simplifies the get functions like above. It would also be useful to expose AV_OPT_SEARCH_* and optionally allow setting them on the set and get calls. Here's a little proof-of-concept that works well enough for my purposes. In context.pyx: cdef class CodecContext(object): # Default search_flags to AV_OPT_SEARCH_CHILDREN def opt_set_int(self, name: str, value: int, search_flags: int = 1): ret = lib.av_opt_set_int(self.ptr, name, value, search_flags) if ret != 0: raise RuntimeError(f'av_opt_set_int() = {ret}') def opt_get_int(self, name: str, search_flags: int = 1): cdef int64_t val ret = lib.av_opt_get_int(self.ptr, name, search_flags, &val) if ret != 0: raise RuntimeError(f'av_opt_get_int() = {ret}') return val And of course in avutil.pxd right after av_opt_set_int(): cdef int av_opt_get_int( void *obj, char *name, int search_flags, int64_t *out_val ) It'd be better if all classes inherited some base class from which void* ptr can be gotten, perhaps via an abstract property or something.
gharchive/issue
2022-08-23T08:55:20
2025-04-01T04:32:59.424271
{ "authors": [ "Tjoppen" ], "repo": "PyAV-Org/PyAV", "url": "https://github.com/PyAV-Org/PyAV/issues/1021", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2268572841
feat: support pyclass on tuple enums (closer but not there yet) This is an extension of this PR, which hasn't been updated in two weeks: https://github.com/PyO3/pyo3/pull/4072 Aiming to close this issue: https://github.com/PyO3/pyo3/issues/3748 @newcomertv in that PR added support for tuple enums but only with ._0 ._1 access to elements. This PR attempts to do the following: [x] respond to minor comments from #4072 [x] support [0] [1] access on tuple elements with __getitem__() [ ] with single-element tuples, automatically forward to that element I've got the Rust-side macros working but haven't yet managed to get them wired up on the Python side... Main complication is that the __getitem__() implementation has to return a Box<dyn Any> (as the type won't be known) and I haven't figured out how to deal with that on the other side. Will give it another go soon but putting this up in case anyone has suggestions... Thanks for continuing the work here! These double underscore methods are a bit special in how their code generation works. I think we can reuse // `slot` is `__GETITEM__`/`__LEN__` from `pyo3-macros-backend/src/pymethods` slot.generate_type_slot(variant_cls_type, &spec, &name, ctx) to generate these special definitions and thread them back up into PyClassImplsBuilder::new slots argument. Then in impl_complex_enum_tuple_variant_cls we can just generate the corresponding functions fn __getitem__(slf: PyRef<Self>, idx: usize) -> PyResult<PyObject> {...} fn __len__(slf: PyRef<Self>) -> usize {...} There should be no need to use Box<dyn Any>, because they need to be convertible to a PyObject anyway. Maybe this helps you a bit while continuing. Just for future reference: We might also want to generate __match_args__ to allow tuple pattern matching on the Python side. Closingin favour of #4072 which has picked this up again! Thanks @Icxolu
gharchive/pull-request
2024-04-29T09:41:56
2025-04-01T04:32:59.489305
{ "authors": [ "Icxolu", "carderne" ], "repo": "PyO3/pyo3", "url": "https://github.com/PyO3/pyo3/pull/4135", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
642307411
Do not use PyObject_RichCompareBool to detect invalid comparison Fixes #985 I think this is the right thing to do in the general case, but I suspect that there's a reasonable speed-up associated with using PyObject_RichCompareBool, since most objects actually are equal when the object identity is the same. I have not looked at this issue in any detail, and I haven't tried to do any benchmarks to see if this change even really had a major speed impact, but I'm wondering how much work it would be to either special case certain types where it is known that pointer comparison is valid, like int, str, list, tuple, datetime.datetime, etc, or even something more elaborate using traits (or an extra field in PyAny indicating whether pointer comparison is "safe" for a given ). I would be curious to see how this change affected the speed of comparison of, say, two identical tuples or lists, since for container types this optimization is already made in __eq__: >>> a = [float('nan')] >>> b = [float('nan')] >>> a == b False >>> a == a True >>> b == b True Another possible performance regression would be two lists that are not the same object but consistent of a large number of elements all of which are the same object, like so: >>> OBJECT = "bbbbb" >>> a = [OBJECT] * 10000 >>> b = [OBJECT] * 10000 >>> a == b
gharchive/pull-request
2020-06-20T04:37:01
2025-04-01T04:32:59.492884
{ "authors": [ "kngwyu", "pganssle" ], "repo": "PyO3/pyo3", "url": "https://github.com/PyO3/pyo3/pull/986", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
927549489
Add note to ExclusionContainer.add_raster(...) that anonymous (lambda) functions are not permitted. Change proposed in this Pull Request Doc-string change for ExclusionContainer.add_raster(...) and changes to tests: Anonymous functions are not permitted for the codes=... keyword argument. Description Non-anonymous functions can be used. As an example I've changed the related test (which was for the meantime inactive) from a lambda function to an equivalent formulation using functools.partial and np.greater_equal. Motivation and Context Anonymous functions lead to complications with python's multiprocessing, c.f. https://github.com/PyPSA/pypsa-eur/issues/249 . How Has This Been Tested? n/a Type of change [x] Bug fix (non-breaking change which fixes an issue) [n/a ] New feature (non-breaking change which adds functionality) [n/a] Breaking change (fix or feature that would cause existing functionality to change) Checklist [x] I tested my contribution locally and it seems to work fine. [ ] I locally ran pytest inside the repository and no unexpected problems came up. [x] I have adjusted the docstrings in the code appropriately. [n/a] I have documented the effects of my code changes in the documentation doc/. [n/a] I have added newly introduced dependencies to environment.yaml file. [x] I have added a note to release notes doc/release_notes.rst. [ ] I have used pre-commit run --all to lint/format/check my contribution great!
gharchive/pull-request
2021-06-22T19:06:54
2025-04-01T04:32:59.498506
{ "authors": [ "FabianHofmann", "euronion" ], "repo": "PyPSA/atlite", "url": "https://github.com/PyPSA/atlite/pull/171", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
755745092
intermediate_source/rpc_param_server_tutorial.rst 번역 문서 URL 번역을 하실 / 요청하실 URL을 남겨주세요. (예. https://tutorials.pytorch.kr/beginner/saving_loading_models.html) URL: https://tutorials.pytorch.kr/intermediate/rpc_param_server_tutorial.html (대략적인) 예상 완료 일정 예상하시는 완료 일정이 있으시다면 알려주세요. (예. 1달 내, 12월 중 등) (반드시 지키셔야 하는 일정이 아닙니다 - 일정이 너무 늦어질 경우 다른 번역자를 위해 남겨주세요.) 2주일 내 관련 이슈 현재 번역 요청 / 진행 내역을 보기 위해 각 버전의 메인 이슈를 참조합니다. (특별한 일이 없다면 수정하지 않으셔도 됩니다.) #90 (v1.6) 진행되지 않은지 오래되어 이슈를 닫습니다. 제가 잘못 닫거나, 진행 중이신 경우 이슈를 다시 열어주시고 덧글로 알려주세요.
gharchive/issue
2020-12-03T01:49:15
2025-04-01T04:32:59.557121
{ "authors": [ "9bow", "hongsam123" ], "repo": "PyTorchKorea/tutorials-kr", "url": "https://github.com/PyTorchKorea/tutorials-kr/issues/195", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
921249232
Added nn.Module support for FasterRCNN backbone What does this PR do? Fixes #660 Before submitting [x] Was this discussed/approved via a Github issue? (no need for typos and docs improvements) [x] Did you read the contributor guideline, Pull Request section? [x] Did you make sure your PR does only one thing, instead of bundling different changes together? [x] Did you make sure to update the documentation with your changes? [x] Did you write any new necessary tests? [not needed for typos/docs] [ ] Did you verify new and existing tests pass locally with your changes? [ ] If you made a notable change (that affects users), did you update the CHANGELOG? PR review [x] Is this pull request ready for review? (if not, please submit in draft mode) Anyone in the community is free to review the PR once the tests have passed. If we didn't discuss your PR in Github issues there's a high chance it will not be merged. Did you have fun? Sure! 🙃 @Borda, I made the changes. Let me know what you think! PS: Tests had failed due to the issue below, which I see has been fixed now... This change is causing tests to fail in master. Please check! 3bf65da#diff-81b21ad681f00e8e80279aaf4d5e2c3a304b5a587b79850cff8d4472147f7f0aL92 Hey @abhayraw1 seems there is one failing test left, do you mind checking it out! thanks for your work so far!! Not sure what is wrong here: boxes = target["boxes"] degenerate_boxes = boxes[:, 2:] <= boxes[:, :2] if degenerate_boxes.any(): # print the first degenerate box bb_idx = torch.where(degenerate_boxes.any(dim=1))[0][0] degen_bb: List[float] = boxes[bb_idx].tolist() > raise ValueError("All bounding boxes should have positive height and width." " Found invalid box {} for target at index {}." .format(degen_bb, target_idx)) E ValueError: All bounding boxes should have positive height and width. Found invalid box [137.5, 556.25, 743.75, 556.25] for target at index 0. as I see the [137.5, 556.25] <= [743.75, 556.25] shall be true, but it fails...
gharchive/pull-request
2021-06-15T10:29:01
2025-04-01T04:32:59.563859
{ "authors": [ "Borda", "SeanNaren", "abhayraw1" ], "repo": "PyTorchLightning/lightning-bolts", "url": "https://github.com/PyTorchLightning/lightning-bolts/pull/661", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1054080139
[RFC] Introduce Strategy in favor of TrainingTypePlugin Proposed refactor Part of #9932 Related PR: #10548 Motivation As we are introducing StrategyPlugin in favor of TrainingTypePlugin for this release. @ananthsub raised some really good points for discussion regarding StrategyPlugin. Reference Proposed solution Three main points to be discussed here Drop Plugin from the name of the class It should be under pytorch_lightning/strategy directory Over time, these classes should be renamed as DDPStrategy and so on. Would like your thoughts? @PyTorchLightning/core-contributors cc @borda @justusschock @awaelchli @akihironitta @kaushikb11 @ananthsub Yes, definitely. We agreed on this I think. This was already set in motion by the introduction of the strategy argument in the Trainer. Over time, these classes should be renamed as DDPStrategy and so on. Has to be done for 1.6. We are targeting stable API for these components at 1.6. Should we name all classes without the "Plugin"/"Strategy" suffix? Here is how this would look like: DataParallel, (or just DP) DDP2, DDP, DDPSpawn, DDPFullySharded, DeepSpeed, Horovod, IPU, SingleDevice, SingleTPU, TPUSpawn, TrainingType, Parallel, DDPSharded, DDPSpawnSharded, On naming, Loggers, Profilers, Accelerators, and ClusterEnvironments are suffixed with the component name (e.g. TensorboardLogger, SimpleProfiler, GPUAccelerator, SlurmEnvironment). If we drop the suffix for strategy: Would we also make this change for other components of Lightning so the naming scheme is unified across components (leading to one less thing to think about for users) ? Do you think there's a risk with naming collisions? For instance, the IPU accelerator and IPU strategy could both be IPU classes. As a result, users would likely need to change the import for them to be: from pytorch_lightning.accelerators import IPU as IPUAccelerator from pytorch_lightning.strategy import IPU as IPUStrategy Proposed plan: Rename classes XPlugin to XStrategy Move files from pl/plugins/training_type to pl/strategy or pl/strategies Reroute pl/plugins/training_type to pl/strategy with a deprecation message Step 3) is necessary because of certain usages of these plugins, even when they are not being subclassed. One example is: trainer = Trainer(plugins=DDPPlugin(find_unused_parameters=False)) in which case we don't want to break user's code. More follow-ups will include: Rename TTP registry Rename tests, comments, docstrings that use the TTP terms @awaelchli the step 1 will be breaking change for trainer = Trainer(plugins=DDPPlugin(find_unused_parameters=False)) use case already, as the DDPPlugin won't exist anymore? To keep backward compatibility we will have to copy the folder pl/plugins/training_type to pl/strategy, and only rename Xplugin to Xstrategy in the new folder pl/strategy and kept pl/plugins/training_type unchanged? In the mean time, reroute pl/plugins/training_type to pl/strategy with a deprecation message. Do we want to do this way? We also have to remember to rename the test files and test names that include the word plugin. We also have to remember to rename the test files and test names that include the word plugin. I could take care of it. Closing this issue, as the Strategy transition has been completed! Thank you to everyone involved ❤️
gharchive/issue
2021-11-15T20:40:02
2025-04-01T04:32:59.575210
{ "authors": [ "ananthsub", "awaelchli", "four4fish", "kaushikb11" ], "repo": "PyTorchLightning/pytorch-lightning", "url": "https://github.com/PyTorchLightning/pytorch-lightning/issues/10549", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
709789011
ref: deprecated results obj, added support for simpler comms (1/n) Here we move to deprecate results objects in favor of simpler syntax. Results obj support will remain until 0.10.0 but not 1.0.0. (ie: 0.10.0 is 1.0.0 but backwards compatible for anyone who really needs it). The new pattern is to decouple logging from hooks. That means that step, step_end, epoch_end are independent from logging. example def training_step(...): loss = ... return loss # equivalent def training_step(...): return {'loss': loss} # to log def training_step(...): self.log('anything', x, on_step=True, on_epoch=True) Passing around step results. In the case the user still needs/wants to do something with the output of each batch, the other hooks are still there. def training_step(...): return {'loss': loss, 'random_thing': [1, 'a', Tensor(), ...]} def training_epoch_end(self, training_step_outputs): for d in training_step_outputs: random_thing = d['random_thing'] Hooks and .log are decoupled... user can log from anywhere def training_epoch_end(...): some_new_val = ... self.log('my_new_val', some_new_val) cc @ananthsub @awaelchli @justusschock Hello @williamFalcon! Thanks for opening this PR. In the file pytorch_lightning/trainer/logging.py: Line 70:1: W293 blank line contains whitespace Do see the Hitchhiker's guide to code style cc @ananthsub @awaelchli @justusschock Hi, after updating to version 1.0.4, i think below approach seems to be not working as desired def training_step(...): return {'loss': loss, 'random_thing': [1, 'a', Tensor(), ...]} def training_epoch_end(self, training_step_outputs): for d in training_step_outputs: random_thing = d['random_thing'] training_step_outputs is always empty. On further debugging i find that training_step is never getting called, instead this function training_step_and_backward is called with run_training_batch ok, will fix asap. thanks for reporting! mind replicating it on a boring model? @edenafek (p0) @nazim1021! Thank you for reporting! I created an issue to track this, would be great if you can try to repro using boring model and paste a colab link in the new issue](https://github.com/PyTorchLightning/pytorch-lightning/issues/4452) @williamFalcon @edenlightning sure, of course. Thanks for prompt reply i ll add the colab link to new issue
gharchive/pull-request
2020-09-27T18:18:23
2025-04-01T04:32:59.584537
{ "authors": [ "edenlightning", "nazim1021", "pep8speaks", "williamFalcon" ], "repo": "PyTorchLightning/pytorch-lightning", "url": "https://github.com/PyTorchLightning/pytorch-lightning/pull/3681", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
711657802
Support OmegaConf hparams on LightningModule What does this PR do? Fixes bug in this stacktrace: https://gist.github.com/ananthsub/07f99b2acacb7ef46bf426f0323cc615 The LightningModule in question does this in the init: class MyModule(LightningModule): def __init__(self, cfg=None, visualize_cfg=None, benchmark_cfg=None, **kwargs): super().__init__() self._cfg = cfg self._visualize_cfg = visualize_cfg self._benchmark_cfg = benchmark_cfg self.save_hyperparameters() ... where all the cfgs are OmegaConf containers. It doesn't look like the framepointer logic for save_hyperparameters handles this well. The LightningModule hparams convert OmegaConf hparams to AttributeDict from this conversion: https://github.com/PyTorchLightning/pytorch-lightning/blob/master/pytorch_lightning/core/lightning.py#L1640-L1641 This causes checks in the tensorboard logger to never trigger, because the type(hparams) is type(params) is <class 'pytorch_lightning.utilities.parsing.AttributeDict'>. This results in YAML dumping to file issues, as seen in the stacktrace Before submitting [ ] Was this discussed/approved via a Github issue? (no need for typos and docs improvements) [x] Did you read the contributor guideline, Pull Request section? [x] Did you make sure your PR does only one thing, instead of bundling different changes together? Otherwise, we ask you to create a separate PR for every change. [ ] Did you make sure to update the documentation with your changes? [ ] Did you write any new necessary tests? [ ] Did you verify new and existing tests pass locally with your changes? [ ] If you made a notable change (that affects users), did you update the CHANGELOG? PR review Anyone in the community is free to review the PR once the tests have passed. If we didn't discuss your PR in Github issues there's a high chance it will not be merged. Did you have fun? Make sure you had fun coding 🙃 Abandoning this PR. The workaround was to Merge the OmegaConfs into a single object set self.hparams = <merged omegaconf> Drop the call to save_hyperparameters() Ideally hparams could be set in the init instead
gharchive/pull-request
2020-09-30T06:02:36
2025-04-01T04:32:59.591388
{ "authors": [ "ananthsub" ], "repo": "PyTorchLightning/pytorch-lightning", "url": "https://github.com/PyTorchLightning/pytorch-lightning/pull/3736", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
850171484
[WIP] Move NaN/Inf detection to a separate utilities file What does this PR do? Fixes #6815 Before submitting [ ] Was this discussed/approved via a GitHub issue? (not for typos and docs) [ ] Did you read the contributor guideline, Pull Request section? [ ] Did you make sure your PR does only one thing, instead of bundling different changes together? [ ] Did you make sure to update the documentation with your changes? (if necessary) [ ] Did you write any new necessary tests? (not for typos and docs) [ ] Did you verify new and existing tests pass locally with your changes? [ ] Did you update the CHANGELOG? (not for typos, docs, test updates, or internal minor changes/refactorings) PR review Anyone in the community is free to review the PR once the tests have passed. Before you start reviewing make sure you have read Review guidelines. In short, see the following bullet-list: [ ] Is this pull request ready for review? (if not, please submit in draft mode) [ ] Check that all items from Before submitting are resolved [ ] Make sure the title is self-explanatory and the description concisely explains the PR [ ] Add labels and milestones (and optionally projects) to the PR so it can be classified Did you have fun? Make sure you had fun coding 🙃 Hello @ananthsub! Thanks for opening this PR. In the file pytorch_lightning/trainer/training_tricks.py: Line 37:121: E501 line too long (121 > 120 characters) Do see the Hitchhiker's guide to code style
gharchive/pull-request
2021-04-05T07:54:04
2025-04-01T04:32:59.599505
{ "authors": [ "ananthsub", "pep8speaks" ], "repo": "PyTorchLightning/pytorch-lightning", "url": "https://github.com/PyTorchLightning/pytorch-lightning/pull/6834", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
926246836
SWT2 with db4 error Hi! Not an issue in itself I guess (might be), but I am posting here because so far I cannot find a solution to this, even though I searched StackOverflow and the PYWT forum. I am trying to apply swt2 with a db4 decomposition to an (516, 862) 2D array. The data has been padded to meet the 2**level criteria and the max level has been calculated to be 3 using swt_max_level. My need is to extract the coefficients at each level (up to max_level), however I when the level > 1 I keep receiving the start_level must be less than 1 error, even though I specifically declare it to be 0. Am I missing something? has been padded to meet the 2**level criteria This should work. If I want to do a 3-level decomposition of data that is shape (516, 862) then it needs to be padded to a multiple of 2**3 on each axis. So, the padded shape should be (516, 864) in this case. The following works for me. import numpy as np import pywt x = np.random.randn(516, 862) x_pad = np.pad(x, pad_width=((0, 0), (0, 2)), mode='reflect') y = pywt.swt(x_pad, wavelet='db4', level=3) If you are not padding as above, it is expected that the max level will be one since 862 is divisible by 21, but not 22 If you are still seeing a problem, can you paste the script similar to the example above that reproduces the issue? Ohhhh, nooo, this was so obvious. I padded the signal but to consider the 2*n not the exponent. So much in my face. Code's fine and thanks so much! Sometimes all it takes is a clear mind to point it out to you after spending to much time on it. Great, glad it is working now.
gharchive/issue
2021-06-21T13:58:07
2025-04-01T04:32:59.603104
{ "authors": [ "cavrinceanu", "grlee77" ], "repo": "PyWavelets/pywt", "url": "https://github.com/PyWavelets/pywt/issues/592", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2086371886
Run screenshotbot.io Test screenshotbot.io integration @PycKamil Arnold from Screenshotbot here (saw your accidental PR on swift-snapshot-testing). Feel free to ping me if you're stuck with anything :)
gharchive/pull-request
2024-01-17T14:57:07
2025-04-01T04:32:59.604889
{ "authors": [ "PycKamil", "tdrhq" ], "repo": "PycKamil/swift-snapshot-testing", "url": "https://github.com/PycKamil/swift-snapshot-testing/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
714317879
Update README.md Correct typo Thanks for that
gharchive/pull-request
2020-10-04T13:47:30
2025-04-01T04:32:59.613865
{ "authors": [ "KJonline", "perjury" ], "repo": "Pyhive/HA-Hive-Custom-Component", "url": "https://github.com/Pyhive/HA-Hive-Custom-Component/pull/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2015503793
Gerneated Tests are empty when run on a fresh project. When i delete the pythagora_tests folder first and then generate the tests, it always generates me an almost empty unit-test file with only the lines ` ` ` ` javascript in it. After i leave the existing folder and just delete the unit-test-file, it will run fine again. Typo in the title. Mus say: "Generated" Update: Now i also get it, when the pythagora_tests folder already exists, while expanding tests with npm run clean && npx pythagora --expand-unit-tests --force --path matchesValue.test.ts The result is then: Content of my handwritten tests //Expanded tests using Pythagora: ```javascript Seems a bit non-deterministic here.
gharchive/issue
2023-11-28T23:36:26
2025-04-01T04:32:59.635188
{ "authors": [ "bogeeee" ], "repo": "Pythagora-io/pythagora", "url": "https://github.com/Pythagora-io/pythagora/issues/67", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }