id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
628832543 | Improve validator logging
PR Description
Introduce validator event logging to log when validators perform duties. This can be enabled with the --log-include-validator-duties-enabled option and is off by default (it would be very noisy when running a lot of validators).
Improve the logging output when validators fail to perform duties.
Fixed Issue(s)
fixes #2019
@bgravenorst This one introduces a new CLI option. Defaults to off but useful if you want logs showing what the validators are doing (probably quite noisy if you run a lot of validators)
| gharchive/pull-request | 2020-06-02T01:29:45 | 2025-04-01T04:32:57.804690 | {
"authors": [
"ajsutton"
],
"repo": "PegaSysEng/teku",
"url": "https://github.com/PegaSysEng/teku/pull/2024",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2354174222 | Make shallow circuit demo Catalyst-Compatible
Context:
Convert Shallow Circuit Demo into a Catalyst-compatiable one.
Description of the Change: Mainly replace jax.jit with qjit. However, it introduces errors related to functools.partial (#814 and #815). Right now I circumvent the errors by avoiding using functools.partial. Will open the other PRs to fix the issues.
Related GitHub Issues:
[sc-65901]
Just a note that codecov/project coverage failure usually means (a) the feature branch's new feature is not covered by a new test and/or (b) the feature branch is out of sync with main branch.
| gharchive/pull-request | 2024-06-14T22:08:05 | 2025-04-01T04:32:57.818938 | {
"authors": [
"paul0403",
"tzunghanjuang"
],
"repo": "PennyLaneAI/catalyst",
"url": "https://github.com/PennyLaneAI/catalyst/pull/816",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
793599162 | #560 Adds constructors with priority and fixed debug statments to use…
closes #560
👍 you guys are too quick for me
👍 you guys are too quick for me
| gharchive/pull-request | 2021-01-25T17:56:39 | 2025-04-01T04:32:57.861782 | {
"authors": [
"ndorin",
"ngenovese11"
],
"repo": "PepperDash/Essentials",
"url": "https://github.com/PepperDash/Essentials/pull/561",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
144683702 | S3 migration
This includes some Dropzone view refactoring too. :)
Current coverage is 87.94%
Merging #451 into master will increase coverage by +0.01% as of f55a987
@@ master #451 diff @@
======================================
Files 232 232
Stmts 10408 10375 -33
Branches 0 0
Methods
======================================
- Hit 9152 9124 -28
Partial 0 0
+ Missed 1256 1251 -5
Review entire Coverage Diff as of f55a987
Powered by Codecov. Updated on successful CI builds.
:+1: looks good to me :)
p.s. refactoring stuff looks great, I hadn't realized how much redundant code was redundant
| gharchive/pull-request | 2016-03-30T18:40:39 | 2025-04-01T04:32:57.889699 | {
"authors": [
"beaugunderson",
"codecov-io",
"madprime"
],
"repo": "PersonalGenomesOrg/open-humans",
"url": "https://github.com/PersonalGenomesOrg/open-humans/pull/451",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
440992798 | Fixed compilation for MSVC C++17
compiles in VisualStudio 2019
there might be a lots of warnings about not using secure functions
_CRT_SECURE_NO_WARNINGS defined will suppress that warning
#4
Note: _CRT_SECURE_NO_WARNING is only needed for tests due to MS extensions.
We checked the tests were with both /c++17 and /latest under MSVC 19.20
@PeterSommerlad By the way, is test_compile_time_warning_nodiscard_scope_guards intended to fail right now? It seems to fail in general (Not just MSVC) I beieve.
| gharchive/pull-request | 2019-05-07T02:12:23 | 2025-04-01T04:32:57.893723 | {
"authors": [
"AetherShane",
"JoeyGrajciar"
],
"repo": "PeterSommerlad/scope17",
"url": "https://github.com/PeterSommerlad/scope17/pull/6",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1903814590 | vignette code fails to execute properly
I am trying to run the vignette code (the Pennsylvania Warbler one) but had a couple issues running the code as provided.
First, the workflow$addStructured command failed when the data was provided as a spatial points dataframe, but worked when I changed it to an sf object.
Second, when trying to actually create the model it ran for a while but then ended with the following error message:
Initializing model
Error in initializeModel$addBias(datasetNames = Workflow$.__enclos_env__$private$biasNames) :
Dataset provided not available.
In addition: Warning message:
In saveRDS(object = PSDMsMOdel, file = paste0(modDirectory, "/", :
'package:stats' may not be available when loading
I'm not sure why the error about the biasNames pops up as the ebird dataset name is set earlier in the code. If I run it with out the biasNames option, the model runs but then I have a different problem - namely, the models aren't retrurned properly (or don't actually run).
Models <- sdmWorkflow(Workflow = workflow)
lapply(Models, summary)
I get a NULL object for the Models returned when I run the code above (when it actually completes a run after removing the biasNames option).
Maybe I'm missing some steps in the vignette, so any help in troubleshooting would be great.
Session Info:
R version 4.3.0 (2023-04-21 ucrt)
Platform: x86_64-w64-mingw32/x64 (64-bit)
Running under: Windows 10 x64 (build 19045)
Matrix products: default
attached base packages:
[1] stats graphics grDevices utils datasets methods base
other attached packages:
[1] USAboundaries_0.4.0 intSDM_2.0.0 terra_1.7-29 ggplot2_3.4.3 PointedSDMs_1.3
[6] R6_2.5.1 inlabru_2.9.0 fmesher_0.1.2 sf_1.0-12
loaded via a namespace (and not attached):
[1] DBI_1.1.3 geodata_0.5-8 s2_1.1.4 remotes_2.4.2
[5] readxl_1.4.2 rlang_1.1.0 magrittr_2.0.3 e1071_1.7-13
[9] tidyterra_0.4.0 compiler_4.3.0 callr_3.7.3 vctrs_0.6.2
[13] stringr_1.5.0 rgbif_3.7.7 httpcode_0.3.0 wk_0.8.0
[17] pkgconfig_2.0.3 crayon_1.5.2 taxize_0.9.100 fastmap_1.1.1
[21] labeling_0.4.3 lwgeom_0.2-13 giscoR_0.3.5 utf8_1.2.3
[25] rmarkdown_2.21 ps_1.7.5 MatrixModels_0.5-1 INLA_23.09.09
[29] purrr_1.0.1 xfun_0.39 embarcadero_1.2.0.1003 jsonlite_1.8.4
[33] reshape_0.8.9 uuid_1.1-1 prettyunits_1.1.1 parallel_4.3.0
[37] R.devices_2.17.1 stringi_1.7.12 lubridate_1.9.2 cellranger_1.1.0
[41] stars_0.6-3 Rcpp_1.0.10 iterators_1.0.14 knitr_1.42
[45] triebeard_0.4.1 zoo_1.8-12 base64enc_0.1-3 R.utils_2.12.2
[49] Metrics_0.1.4 splines_4.3.0 Matrix_1.5-4 timechange_0.2.0
[53] tidyselect_1.2.0 rstudioapi_0.14 abind_1.4-5 yaml_2.3.7
[57] codetools_0.2-19 curl_5.0.0 processx_3.8.1 pkgbuild_1.4.0
[61] lattice_0.21-8 tibble_3.2.1 plyr_1.8.8 withr_2.5.0
[65] USAboundariesData_0.4.0 evaluate_0.20 desc_1.4.2 units_0.8-1
[69] proxy_0.4-27 xml2_1.3.3 pillar_1.9.0 whisker_0.4.1
[73] blockCV_3.1-3 KernSmooth_2.23-20 foreach_1.5.2 generics_0.1.3
[77] rprojroot_2.0.3 sp_1.6-0 munsell_0.5.0 scales_1.2.1
[81] rgdal_1.6-6 class_7.3-21 glue_1.6.2 lazyeval_0.2.2
[85] tools_4.3.0 data.table_1.14.8 dbarts_0.9-23 grid_4.3.0
[89] bold_1.2.0 tidyr_1.3.0 ape_5.7-1 urltools_1.7.3
[93] tidyverse_2.0.0 colorspace_2.1-0 nlme_3.1-162 raster_3.6-20
[97] conditionz_0.1.0 cli_3.6.1 fansi_1.0.4 dplyr_1.1.2
[101] gtable_0.3.4 R.methodsS3_1.8.2 oai_0.4.0 digest_0.6.31
[105] classInt_0.4-9 crul_1.3 farver_2.1.1 htmltools_0.5.5
[109] R.oo_1.25.0 lifecycle_1.0.3 dismo_1.3-9 httr_1.4.5
Hey, I noticed that issue not too long ago and managed to fix it. It should work if you use the github version of the package rather than the CRAN version. devtools::install_github('PhilipMostert/intSDM)
Hmm, i tried it again with the new version but I'm still having issues with the model output - the output is still NULL after rerunning this morning. However, it did run with the biasNames option this time, while it didn't before. Maybe I am missing something else?
Is there any particular issue that you get when running the models? There was an issue in the last line when printing the summaries, but it shouldn't have returned NULL.
Have you by any chance changed the Save argument instartWorkflow? I see from your first comment that it was trying to save the output of the model. If set to TRUE then the function will try and save the output of the model to the chosen directory instead of returning the object.
Well, that makes a lot of sense! The files saved properly it looks like and I can load them from the saved folders. Looks like it is all good! Thanks for your help.
also just wanted to add I'm excited to try it out with my own data, looks like a really cool package!
| gharchive/issue | 2023-09-19T22:32:16 | 2025-04-01T04:32:57.957477 | {
"authors": [
"PhilipMostert",
"king0708"
],
"repo": "PhilipMostert/intSDM",
"url": "https://github.com/PhilipMostert/intSDM/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
388385160 | Make sure releases are signed
https://www.electron.build/code-signing
once we bring that to ethereum org I can port what we have on ethereum/mist.
It's basically, just set some proper env variables.
| gharchive/issue | 2018-12-06T20:29:25 | 2025-04-01T04:32:57.959498 | {
"authors": [
"PhilippLgh",
"evertonfraga"
],
"repo": "PhilippLgh/mist-shell",
"url": "https://github.com/PhilippLgh/mist-shell/issues/4",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
1645858660 | More data
Swype will properly release their instruction fine tuning dataset real soon:
https://huggingface.co/datasets/swype/instruct-102.4k
And there is also:
https://github.com/nomic-ai/gpt4all
I forgot this one, damn
https://github.com/mobarski/alpaca-libre
Thank you for the reminder. We will collect and format these as soon as possible.
| gharchive/issue | 2023-03-29T13:50:44 | 2025-04-01T04:32:57.977628 | {
"authors": [
"KnutJaegersberg",
"PhoebusSi"
],
"repo": "PhoebusSi/Alpaca-CoT",
"url": "https://github.com/PhoebusSi/Alpaca-CoT/issues/15",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
532933916 | FileValidator won't open in Win10.
Hello,
I wouldl ike to ask for a short support. I downloaded the
"win-photonfilevalidator-2.0 (2)" file. Installed and started.
However the program started and it is showned on the tray, but the window not appear on the screen.
Do you know what I did wrong?
Thank you
Problem solved.
i had a 2nd monitor which was not turned on. I disconnected it from the PC and the window appeared on the first monitor.
| gharchive/issue | 2019-12-04T21:05:14 | 2025-04-01T04:32:57.984715 | {
"authors": [
"kenikol"
],
"repo": "Photonsters/PhotonFileValidator",
"url": "https://github.com/Photonsters/PhotonFileValidator/issues/41",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
666620431 | Index Sid in SubPost table for fast retrieval for sub filter posts
Why doesn't phuks use any indexing on foreign key field like pid in subcomment table for fast queries/retrieval of data. Thanks
We do, but I've forgotten to add a migration for this after we switched from our old .sql dump to migrations for database setup
We do, but I've forgotten to add a migration for this after we switched from our old .sql dump to migrations for database setup
Does it index parentcid in SubPostComment table? Does it require to index parentcid of SubPostComment. Comment retrieval is using pid of SubPostComment. So does it need to index parentcid.
Yes.
Yes.
Can you show the database table schema. I am testing in local development. I am confused what to index and what not to.
Just double checked and the migrations are already creating most of the required indexes. I just pushed a migration with three more indexes that shouldn't impact everyday use a lot but should help anyway.
@migrator.create_model
class SubPostComment(pw.Model):
cid = pw.CharField(max_length=40, primary_key=True)
content = pw.TextField(null=True)
lastedit = pw.DateTimeField(null=True)
parentcid = pw.ForeignKeyField(backref='subpostcomment_set', column_name='parentcid', field='cid', model='self', null=True)
pid =
Just double checked and the migrations are already creating most of the required indexes. I just pushed a migration with three more indexes that shouldn't impact everyday use a lot but should help anyway.
@migrator.create_model
class SubPostComment(pw.Model):
cid = pw.CharField(max_length=40, primary_key=True)
parentcid = pw.ForeignKeyField(backref='subpostcomment_set', column_name='parentcid', field='cid', model='self', null=True)
pid = pw.ForeignKeyField(backref='subpostcomment_set', column_name='pid', field='pid', model=migrator.orm['sub_post'], null=True)
Where is the indexing for pid? Does peewee migration create indexing automatically for foreignkey?
Yes
| gharchive/issue | 2020-07-27T22:20:49 | 2025-04-01T04:32:57.994099 | {
"authors": [
"Polsaker",
"axiomxd"
],
"repo": "Phuks-co/throat",
"url": "https://github.com/Phuks-co/throat/issues/136",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
298183213 | Automatically load Spreadsheet
Implement loading an online version of responses on startup.
Probably will have to export the spreadsheet as csv or similar...
TODO: Direct GDocs loading? GitHub releases? Personal server?
After careful consideration I think the best idea would be to create tags on GitHub and upload curated CSV exports of the GDocs spreadsheet there. The application can then load all available tags from GitHub on startup and if it's newer than the local one it downloads the latest CSV and saves it somewhere to disk.
The CSV should be loaded through some sort of helper class, to avoid code repetition in different Behaviour classes.
I'll see what I can do about inputting from the CSV. I was thinking about doing semi-colon delimiters so the program wouldn't be disrupted by punctuation (Assuming we can keep our text "semi-colon free").
I'm not very familiar with how GitHub and how loading tags from it into the program would work however. Do you have any reference material I can study up on as far as that goes?
@PiMaker
One could take a look at this: https://github.com/octokit/octokit.net
Seems like that would make interfacing with GitHub easier. I will create a sample tag and upload a CSV for testing. Semicolon-Seperated sounds good, that's still called a CSV though.
Awesome, i'll take a look at that in the morning!
And I am indeed @PiMaker
Okay got a good idea for building this, mind assigning me to the issue so I can receive any updates from this thread?
@PiMaker
Alright, I have no clue how to assign you the issue, it doesn't let me from the dropdown I get... Anyway, you should receive updates to this thread anyway, because you are mentioned in it? Once you have a prototype you can create a pull request from a fork of yours, that should allow me to merge your code into the main repo.
I thought about also adding in a general update functionality. Although I feel like that should be a seperate part anyway. Have you focused your prototype idea more on parsing the CSV or downloading updates from GitHub? Because if you want we could split up that task, so I can take on one or the other as well.
Okay, maybe there's a way to assign myself! I'm very unfamiliar with this platform so if I get updates then it sounds good!
And i'm focusing currently on the parsing the CSV to start with, if you could figure out downloading updates from GitHub so we can combine when i'm finished that would be very helpful.
Sounds like a plan. For the format of the CSV, if you could somewhat base that on the spreadsheet already on Google Docs that would be very helpful, as it would minimize effort to create a new version from new entries in there.
Yeah, what I did is I exported the Google Docs to a CSV file after swapping the delimiter to a semi-colon and I've been testing my code with that. I just now got it working so I can store the code as members in structs. Tomorrow (hopefully) i'll perfect that and figure out how your code works so I can start working on her reading in the behaviours
The interesting part is going to be how to integrate my browser process instance detection function to avoid detecting multiple processes in one single browser instance. I think that will have to be a hardcoded edge-case for now, as long as you don't figure out some magical solution. (You'll see once you take a look at my code)
Okay so I've got it down a basic thing to be able to read the information into my struct the main issue i'll have now is integrating my C++ into your C# so that I can finish it up by making the function so MonikAI can read it
I apologize for entering the thread late. I started created a csv parser the first day I saw the repository but the code was messy and I did not get a chance to improve it until today but I saw that it was already being worked on. I've submitted a PR with the changes that I made but my code is in C#. Is there any reason why the parser needs to be done in C++? It seems like using unmanaged code with the core codebase might make things more complicated to change and extend in the future.
@bwroberts48 Is your code posted anywhere? We can work together to convert it to C# in case my code isn't up to par with the rest of the repository (although if that's the case, I don't think you'd want my help!)
I linked my parser to the ApplicationBehavior class and from the testing I did everything seems to be working fine. I will take a look at linking it to the WebBrowserBehavior next if there are no issues to resolve.
On a side note, the auto updater seems to be broken for me. For some reason, whenever I run MonikAI, the auto updater immediately maxes out the 60 hourly GitHub API requests that I can make. I will investigate this issue further at a later time.
@zchaudhry11
Only reason I was writing in C++ was because I don't know C# so your code makes the process a lot easier. I'm a novice to programming so I'll look at your code but I bet it's more elegant than mine
I'm not a professional by any means so feel free to make any corrections if something looks problematic or ask any questions if you have any.
C# is an amazing programming language and you should give it a shot since there's no better way to learn than diving head first into a new codebase!
Merged the pull request from @zchaudhry11 and implemented the remaining few bits myself. Thank you all!
| gharchive/issue | 2018-02-19T07:34:35 | 2025-04-01T04:32:58.004163 | {
"authors": [
"PiMaker",
"bwroberts48",
"zchaudhry11"
],
"repo": "PiMaker/MonikAI",
"url": "https://github.com/PiMaker/MonikAI/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
427291562 | VideoFileRenderer
I used the same implementation of the VideoFileRenderer you made, it worked but I am having the following issues:
I have recorded a 3 seconds video file, its size is about 40 MegaBytes, is that normal, I tried to change the resolution of video captured so that its quality gets reduced, but this did not make any difference.
I could not play the video except with external apps like MXPlayer (could not play it with the normal video player that comes embedded with the android) I wrote about that I found that it is of extension .y4m, I tried to convert it programmatically into mp4 or something playable by the native VideoView or even the video player that comes with the android but could not.
This is not an issue: how to do the same with audio, i.e. what is the VideoFileRenderer for audio
I really appreciate your effort in webrtc, Thanks in advance.
At first, I didn't commit any actual code into this repo, all the source code commit history are the same with the official repo, the real content of this repo is the sdk/android_gradle folder, which is a collection of build scripts. Although I contribute several patches into the official webrtc repo :)
Regarding your questions: if you read VideoFileRenderer's source code, you will find answers of your questions :)
| gharchive/issue | 2019-03-30T15:04:48 | 2025-04-01T04:32:58.009379 | {
"authors": [
"Piasy",
"cppgarbagecollector"
],
"repo": "Piasy/webrtc",
"url": "https://github.com/Piasy/webrtc/issues/7",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1051310930 | ✨ Add raw_content property to UserMessage
This will remove text formatting (ex Italics, Bold, Code Blocks). This will ame it easier to deal with text.
I suppose I can take a look at this, do emoji's need to removed as well?
I don't think so because I think it would be good if you can search for an emoji in a cleaned string
I think it should be called clean_content, because raw_content tells me nothing has changed to the content
I think it should be called clean_content, because raw_content tells me nothing has changed to the content
Sure! It'll make more sense for dpy devs too.
Just putting my grain of salt.
I personally wouldn't put the burden on cleaning the text -- removing the content aforementioned in the comment thread -- on the library but rather in something else, like user code. Okay, maybe if this comes often enough, why not make something external that is standalone.
I would rather have a property or a method that will nullify effects of certain mentions like @everyone or @here, replace different mentions[^1] by their textual counterpart. For example, <#0123456789> becomes #my-channel-has-been-mentioned.
If cleaning the message text will be useful at multiple places in the library, why not. If it is convenient for the end user so they don't have headaches with regex, also why not !
[^1]: Roles, channels, users, whatever is mentionable, you name it!
#237
Should it also convert emojis like :eyes: to unicode?
| gharchive/issue | 2021-11-11T19:54:19 | 2025-04-01T04:32:58.036931 | {
"authors": [
"Lunarmagpie",
"Sly-Little-Fox",
"beastmatser",
"lapin-b"
],
"repo": "Pincer-org/Pincer",
"url": "https://github.com/Pincer-org/Pincer/issues/192",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1944737449 | fix(tidb): fix script integrationtest.sh
Correct start pd cmd params.
Specify using store tikv for testing.
/hold
Currently, running with store tikv will result in the following errors.
ERRO[0663] 8 tests failed
ERRO[0663] run test [executor/admin] err: sql:admin checksum table cache_admin_table_with_index_test;: failed to run query
"admin checksum table cache_admin_table_with_index_test;"
around line 66,
we need(172):
admin checksum table cache_admin_table_with_index_test;
Db_name Table_name Checksum_crc64_xor Total_kvs Total_bytes
executor__admin cache_admin_table_with_index_test 0 2 2
but got(172):
admin checksum table cache_admin_table_with_index_test;
Db_name Table_name Checksum_crc64_xor Total_kvs Total_bytes
executor__admin cache_admin_table_with_index_test 0 0 0
ERRO[0663] run test [executor/executor] err: sql:change pump to node_state ='paused' for node_id 'pump1';: failed to run query
"change pump to node_state ='paused' for node_id 'pump1';"
around line 693,
we need(132):
change pump to node_state ='paused' for node_id 'pump1';
Error 1105 (HY000): URL scheme must be http, https, unix, or unixs:
change
but got(132):
change pump to node_state ='paused' for node_id 'pump1';
Error 1105 (HY000): node pump, id pump1 from etcd 127.0.0.1:2379 not found
ERRO[0663] run test [expression/builtin] err: sql:select * from tb5 where cast(a as unsigned int)=0;: failed to run query
"select * from tb5 where cast(a as unsigned int)=0;"
around line 280,
we need(156):
select * from tb5 where cast(a as unsigned int)=0;
a b
Level Code Message
Warning 1690 constant 1.844674407370955e+20 overflows bigint
select * from tb5 whe
but got(156):
select * from tb5 where cast(a as unsigned int)=0;
a b
Level Code Message
Warning 1690 evaluation failed: constant 184467440737095500000 overflows LongLong
ERRO[0663] run test [expression/charset_and_collation] err: sql:admin recover index t a;: run "admin recover index t a;" at line 359 err Error 1105 (HY000): [components/tidb_query_executors/src/table_scan_executor.rs:425]: Data is corrupted, missing data for NOT NULL column (offset = 0)
ERRO[0663] run test [expression/issues] err: sql:select * from t where field('A', a collate utf8mb4_general_ci, b) > 1;: failed to run query
"select * from t where field('A', a collate utf8mb4_general_ci, b) > 1;"
around line 1445,
we need(79):
select * from t where field('A', a collate utf8mb4_general_ci, b) > 1;
a b
sele
but got(79):
select * from t where field('A', a collate utf8mb4_general_ci, b) > 1;
a b
a A
ERRO[0663] run test [planner/cascades/integration] err: sql:select /*+ HASH_AGG() */ a, count(distinct a) from t;: failed to run query
"select /*+ HASH_AGG() */ a, count(distinct a) from t;"
around line 135,
we need(78):
select /*+ HASH_AGG() */ a, count(distinct a) from t;
a count(distinct a)
1 2
but got(78):
select /*+ HASH_AGG() */ a, count(distinct a) from t;
a count(distinct a)
2 2
ERRO[0663] run test [planner/core/plan_cache] err: sql:select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934;: failed to run query
"select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934;"
around line 209,
we need(263):
select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934;
Error 1690 (22003): DECIMAL value is out of range
but got(263):
select * from IDT_20290 where col2 * 049015787697063065230692384394107598316198958.1850509 >= 659971401668884663953087553591534913868320924.5040396 and col2 = 869042976700631943559871054704914143535627349.9659934;
Error 1690 (22003): value is out of range in ''
ERRO[0663] run test [session/vars] err: sql:SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable';: failed to run query
"SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable';"
around line 3,
we need(98):
SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable';
variable_value
SET G
but got(98):
SELECT variable_value FROM mysql.tidb WHERE variable_name = 'tikv_gc_enable';
variable_value
true
/unhold
| gharchive/pull-request | 2023-10-16T09:07:51 | 2025-04-01T04:32:58.045284 | {
"authors": [
"purelind"
],
"repo": "PingCAP-QE/ci",
"url": "https://github.com/PingCAP-QE/ci/pull/2519",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2014636052 | chore(apps/prod/tekton/setup): bump tekton operator to v0.61.0
With components updated:
hub "v1.9.0"
dashboard "v0.28.0"
pipeline "v0.39.0"
chains "v0.11.0"
Signed-off-by: wuhuizuo wuhuizuo@126.com
I have already done a preliminary review for you, and I hope to help you do a better job.
This pull request updates the Tekton operator and several other components to the newest versions. There are no potential problems found within this pull request.
Suggestion:
It is always a good idea to test the updated components before merging the changes into the main branch.
If it failed I will revert it.
/approve
| gharchive/pull-request | 2023-11-28T14:56:14 | 2025-04-01T04:32:58.048874 | {
"authors": [
"wuhuizuo"
],
"repo": "PingCAP-QE/ee-ops",
"url": "https://github.com/PingCAP-QE/ee-ops/pull/804",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2023159936 | feat(charts): add chart git-cdn
We will deploy it to proxy git clone globally to save CI/CD time and brandwidth.
It's based open source project git_cdn.
Signed-off-by: wuhuizuo wuhuizuo@126.com
/approve
| gharchive/pull-request | 2023-12-04T06:57:30 | 2025-04-01T04:32:58.050730 | {
"authors": [
"wuhuizuo"
],
"repo": "PingCAP-QE/ee-ops",
"url": "https://github.com/PingCAP-QE/ee-ops/pull/819",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2596141522 | Customization of wave function for TileReflection
The sharp, clear-cut edges of the tiles in TileReflection are achieved through the tangent function, which approaches infinite on one side and minus infinite on the other side, resulting in very sudden jumps.
But what if we chose a different function? For example if we chose sine, the image would be bulged in a tile pattern but it wouldn't have clear edges. Other functions would behave in different, interesting ways.
This is cool, definitely has some interesting effects!
I'm mostly unsure about the naming of the Wave Type parameter and its menu options, which are pretty technical and might not make much sense to a user without an understanding of what the effect is doing internally. Ideally there could be some names that describe what behaviour is produced?
@cameronwhite what names would you suggest?
I haven't thought much about names for the new modes, but for example you might label the existing mode as Tile Shape -> Curved Edges or something like that?
The current mode looks like the image is reflected in a mirror made up of tiles that have some curvature along their edges. Ideally we could have some labels for the new modes that give some idea of the general look of the effect
@cameronwhite if we focus on the visual effect and not the wave function as such, we'd have to think about this differently, so I reduced the number of options to two to start with (with more tile types coming in the future). Also, I gave the enum the name of TileType and the values are CleavedEdges (which is what we already have) and CurvedEdges (the new option). English is not my first language, so if you can think of better names, that would be cool.
@cameronwhite I implemented your suggestions 👍
| gharchive/pull-request | 2024-10-18T01:27:33 | 2025-04-01T04:32:58.058870 | {
"authors": [
"Lehonti",
"cameronwhite"
],
"repo": "PintaProject/Pinta",
"url": "https://github.com/PintaProject/Pinta/pull/1051",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
15739453 | Cosmetic bugfixes
Created new branch for these commits as we decided to look again at the changes for Pinta 1.6 (https://github.com/PintaProject/Pinta/pull/61). I needed this in order to be able to continue working on master without mixing up things so much.
Closing this due to lack of time to look into it
| gharchive/pull-request | 2013-06-19T11:21:31 | 2025-04-01T04:32:58.060330 | {
"authors": [
"ericksson"
],
"repo": "PintaProject/Pinta",
"url": "https://github.com/PintaProject/Pinta/pull/63",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
131314427 | Cant compile it
This shit cannot be compiled
So don't use it. Stupid and ignorant issues like this are not tolerated.
| gharchive/issue | 2016-02-04T11:07:53 | 2025-04-01T04:32:58.077823 | {
"authors": [
"finnishhacker",
"ludo237"
],
"repo": "Pirate-Crew/IPTV-Extreme",
"url": "https://github.com/Pirate-Crew/IPTV-Extreme/issues/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2000842386 | Unable to install the component
Hello,
I've just tried to install the pirateweather component but got no entities after installing.
In the log file I have the following error message:
Logger: homeassistant.setup
Source: setup.py:379
Unable to prepare setup for platform pirateweather.weather: Platform not found (cannot import name 'WeatherEntityFeature' from 'homeassistant.components.weather' (/***/homeassistant/lib/python3.10/site-packages/homeassistant/components/weather/init.py)).
Anybody knows a fix for this problem?
Thank you very much.
I'll tag @alexander0042 so he can take a look at this but some helpful information to know is:
Which version of the integration are installing and what version of HA are you running?
Hello,
Home Assistant Version: 2023.7.3
Pirate-Weather: Version 1.3.2
Hi, thanks for opening this issue, and apologies for the issue here. The hiccup is that you're running the pre-release version of Pirate Weather, but your version of HA is tool old.
However, HACS should be able to prevent this, so let me take a look at the manifest to see what went wrong here!
Argh, just realized I've installed HA Core - will this even work in core?
The pre-release version of the integration requires HA version 2023.10 so you'd need to update your version of HA to run that version. The release version 1.2.3 will work with your HA version so have you tried installing that and seeing if it works?
I'm not a HA expert but I'll tag @alexander0042 again as he might know more information on if the integration would work in Core.
I've updated HA to the latest version and it's working - thanks!
Glad you were able to get it working. Since this issue has been solved I'm going to close this issue.
| gharchive/issue | 2023-11-19T13:04:27 | 2025-04-01T04:32:58.083162 | {
"authors": [
"alexander0042",
"cloneofghosts",
"km8742"
],
"repo": "Pirate-Weather/pirate-weather-ha",
"url": "https://github.com/Pirate-Weather/pirate-weather-ha/issues/143",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
311535732 | Avoid panic when blending two pixels with zero alpha.
I ran into a situation where trying to overlay two rgba images led to a panic.
The root cause was a division by zero.
Here's a minimal reproducible example program run against image = "0.18.0":
extern crate image;
use image::{ImageBuffer, Rgba, RgbaImage};
use image::imageops::overlay;
fn main() {
let mut a: RgbaImage = ImageBuffer::new(1, 1);
a.put_pixel(0, 0, Rgba([0, 0, 0, 0]));
let mut b: RgbaImage = ImageBuffer::new(1, 1);
b.put_pixel(0, 0, Rgba([0, 0, 0, 0]));
overlay(&mut a, &b, 0, 0);
}
I don't have a ton of graphics experience, so I'm not sure if there's a "correct" way to blend two completely transparent pixels in terms of retaining the color components.
This patch just uses the blended colors and leaves alpha at zero.
This patch is the simplest thing I could think of to fix, but I'd be happy to change to something more idiomatic (num::CheckedDiv trait and explicit handling of the None condition?)
It seems that the final_alpha is set to 1.0? Should this not be zero?
Doh, I wasn't thinking and was only testing against images with completely opaque pixels. I'll amend the commit.
That said --- is there any reason to do anything with the color values at all when blending two transparent pixels?
If not, then presumably it'd be more efficient to return a transparent white/black pixel rather than doing calculations on the color components.
I think blending is good, because it preserves the information in the image to neighboring pixels. For example, if you are super-sampling textures, white/black pixels can cause trouble.
Thanks for the quick reply!
I did some reading, and I don't think this kind of compositing is well defined when both pixels have zero alpha.
The referenced stack overflow answer and SVG spec, defines compositing in terms of "premultiplied" color components (as defined in the original Porter and Duff paper).
But since alpha is in the domain [0, 1],
let alpha_final = bg_a + fg_a - bg_a * fg_a;
must always be greater than the alpha of either pixel.
Which means that it can only be zero when both pixel's alpha are zero.
But in that case, the "premultiplication" step
let (bg_r_a, bg_g_a, bg_b_a) = (bg_r * bg_a, bg_g * bg_a, bg_b * bg_a);
let (fg_r_a, fg_g_a, fg_b_a) = (fg_r * fg_a, fg_g * fg_a, fg_b * fg_a);
will set all of the color components to zero.
Or am I missing something in my logic there?
Premultiplied color components do not conserve the color information in pixels when making them transparent. This is a different way of encoding images, which is not used here.
The code I quoted wasn't hypothetical, it's from the RGBA blend implementation and has the comment "We premultiply our channels bu their alpha, as this makes it easier to calculate".
Or are you suggesting that we switch to a different blending algorithm in the case when both source alphas are zero?
@lynaghk I messed this up, sorry.
I was not aware of premultiplying colors being used here. Not sure if this is the right algorithm.
@bvssvni No worries. I'm not too familiar w/ the tradeoffs for color blending and graphics stuff like this, so I'm not sure what the appropriate thing to do would be --- all I wanted from my changes was to avoid a panic.
If you do find another way through and want a hand, let me know.
But no worries if it'd be easier to just do the changes yourself too.
Another issue that there is no conversion to sRGB before blending. This algorithm is very opinionated. However, that is a design decision that could be dealt with in a later version.
Can you add a unit test that blends two transparent pixels and see what happens to the alpha channel? At least the resulting alpha channel should be zero.
I think it should simply return instead of setting the alpha to 1.0.
Closing in favor of https://github.com/PistonDevelopers/image/pull/759
Thanks for pointing this out!
Sorry for the delay on unit tests, I was away from the Internet.
Returning makes sense, given that the blend operation mutates. Thanks for the fix and your work on this library!
| gharchive/pull-request | 2018-04-05T09:27:49 | 2025-04-01T04:32:58.095973 | {
"authors": [
"bvssvni",
"lynaghk"
],
"repo": "PistonDevelopers/image",
"url": "https://github.com/PistonDevelopers/image/pull/744",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1081735411 | yarn start error
yarn install success,but yarn start with this error:
Attempted import error: 'CanvasSlicImg' is not exported from '../../pixels-metaverse'.
yarn install success,but yarn start with this error: Attempted import error: 'CanvasSlicImg' is not exported from '../../pixels-metaverse'.
Sorry, since I need to make some upgrades to the feature to participate in the contest, the code has not been released yet, so the update to the react-pixel-metaverse package has not been released yet, so you may not find the corresponding CanvasSlicImg component.
很抱歉,由于我需要对功能进行一些升级才能参与黑客松比赛,对react-pixel-metaverse包的更新尚未发布,因此您可能找不到相应的CanvasSlicImg组件。
You can now access the link , which was the previous version, but can be used for learning. In addition, the project will be released once before Christmas, and then you can see that the warehouse code is working properly. Thank you for your attention and hope we can make progress together.
当前您可以访问链接 https://github.com/xiangzhengfeng/pixels-metavers,这是以前的版本,但可用于学习。此外,该项目 将在圣诞节前发布一次,到时候您可以看到仓库代码正常工作了。感谢您的关注,希望我们能共同进步。
| gharchive/issue | 2021-12-16T03:38:40 | 2025-04-01T04:32:58.132191 | {
"authors": [
"lhzwb2008",
"qianduanxinlv"
],
"repo": "PixelsMetaverse/pixels-metaverse",
"url": "https://github.com/PixelsMetaverse/pixels-metaverse/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1864337680 | 出现这个问题
我有8G的GPU,但是还是出现这样的问题?
G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\whisper\timing.py:58: NumbaDeprecationWarning: The 'nopython' keyword argument was not supplied to the 'numba.jit' decorator. The implicit default value for this argument is currently False, but it will be changed to True in Numba 0.59.0. See https://numba.readthedocs.io/en/stable/reference/deprecation.html#deprecation-of-object-mode-fall-back-behaviour-when-using-jit for details. def backtrace(trace: np.ndarray): Use 8 cpu cores for computing Traceback (most recent call last): File "G:\AI\VALL\VALL-E-X-master\VALL-E-X-master\ui.py", line 84, in <module> whisper_model = whisper.load_model("medium").cpu() File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\whisper\__init__.py", line 154, in load_model return model.to(device) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 1145, in to return self._apply(convert) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 797, in _apply module._apply(fn) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 797, in _apply module._apply(fn) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 797, in _apply module._apply(fn) [Previous line repeated 2 more times] File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 820, in _apply param_applied = fn(param) File "G:\AI\VALL\VALL-E-X-master\myenv\lib\site-packages\torch\nn\modules\module.py", line 1143, in convert return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking) torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 4.00 GiB total capacity; 3.05 GiB already allocated; 0 bytes free; 3.06 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
看起来是在加载whisper(另一个ai模型,用于识别声音中的文字)的时候出现了错误。
目前vallex需要输入一段声音,这段声音中的文字内容,以及要合成的文字内容,输出合成的声音。
如果你没有输入原音频中的文字内容,那么就会调用whisper模型自动识别原音频中的文字内容。这将花费额外的显存。
可以尝试手动输入原音频中的文本内容来避免调用whisper
感谢哥哥答复,我运行 ui时(python launch-ui.py),没有启动成功就出现这个报错。很奇怪
可以尝试修改源代码,删除加载whisper模型的部分
我试试看看~
可以了,打开ui了。是whisper和程序强GPU了。去掉whisper就可以
| gharchive/issue | 2023-08-24T03:28:48 | 2025-04-01T04:32:58.175782 | {
"authors": [
"40740",
"ZYFDroid"
],
"repo": "Plachtaa/VALL-E-X",
"url": "https://github.com/Plachtaa/VALL-E-X/issues/22",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2435373023 | chore(appeals): health-check (eviction settings)
Adds a timeout (5 mins) for health check eviction
Issue ticket number and link
Type of change 🧩
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation update
[x] Other (please explain in the description section above)
Checklist before requesting a review
[ ] I have performed a self-review of my own code
[ ] I have double checked this work does not include any hardcoded secrets or passwords
[ ] I have made corresponding changes to the documentation
[ ] I have provided details on how I have tested my code
[ ] I have referenced the ticket number above
[ ] I have provided a description of the ticket
[ ] I have included unit tests to cover any testable code changes
Unsupported argument
| gharchive/pull-request | 2024-07-29T13:04:57 | 2025-04-01T04:32:58.179974 | {
"authors": [
"fabfab-kc"
],
"repo": "Planning-Inspectorate/appeals-back-office",
"url": "https://github.com/Planning-Inspectorate/appeals-back-office/pull/334",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1529125494 | chore(web/applications): log sessionaccount and AD access token
Describe your changes
Issue ticket number and link
Type of change 🧩
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation update
[ ] Other (please explain in the description section above)
Checklist before requesting a review
[ ] I have performed a self-review of my own code
[ ] I have double checked this work does not include any hardcoded secrets or passwords
[ ] I have made corresponding changes to the documentation
[ ] I have provided details on how I have tested my code
[ ] I have referenced the ticket number above
[ ] I have provided a description of the ticket
[ ] I have included unit tests to cover any testable code changes
:tada: This PR is included in version 1.0.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 1.0.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2023-01-11T14:12:18 | 2025-04-01T04:32:58.186028 | {
"authors": [
"AnthonyJWalsh",
"lutamKC"
],
"repo": "Planning-Inspectorate/back-office",
"url": "https://github.com/Planning-Inspectorate/back-office/pull/668",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1933746743 | [common] instruction definition
Adds
opcode identifiers
operation identifiers
decoded instruction structure definition
LGTM
| gharchive/pull-request | 2023-10-09T20:27:48 | 2025-04-01T04:32:58.215488 | {
"authors": [
"Tako-San",
"mmamayka"
],
"repo": "Plumbus-Group-Inc/Plumbus-VM",
"url": "https://github.com/Plumbus-Group-Inc/Plumbus-VM/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
816390963 | Modification of QS after a success
Allowance of modifications of QS after a roll, but only if the roll was a success (otherwise, you could work-around by just adding 3FP to the roll)
Needed for future implementations anyways, like Athlet.
Low prio, if accepted at all.
Alternatively, allow to edit the Fertigkeitspunkte after a roll (currently, this works for dice values and modificators, but not for FP.
| gharchive/issue | 2021-02-25T12:27:06 | 2025-04-01T04:32:58.218599 | {
"authors": [
"mckee14"
],
"repo": "Plushtoast/dsa5-foundryVTT",
"url": "https://github.com/Plushtoast/dsa5-foundryVTT/issues/139",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1231421223 | Fix error recovery mode for SubmitTx
As part of #233 we implemented a simple approach to error recovery with the use of reconnecting-websocket. However, there's one special endpoint for which simple replaying is not going to work. We need to check if a transaction has been sent before resubmitting it (otherwise UTXOs may be consumed by TX and we'll get an error when trying to send it for the second time).
How to address this problem: maintain request queue as a list of transactions instead of a list of String payloads. Before retrying, use https://ogmios.dev/mini-protocols/local-tx-monitor/ to see if a transaction is in local node mempool.
The reason why we are delaying this: the chance to observe the error is low.
| gharchive/issue | 2022-05-10T16:38:15 | 2025-04-01T04:32:58.220509 | {
"authors": [
"klntsky"
],
"repo": "Plutonomicon/cardano-transaction-lib",
"url": "https://github.com/Plutonomicon/cardano-transaction-lib/issues/398",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1123817023 | [BUG] Unable to reinitialize plugin after getting GDPR consent
Hi, first of all, thanks for the plugin, I've been using it in my iOS app that was released recently, and now I'd like to implement GDPR consent, I just realized that your plugin also supports this. I'm almost done debugging it, but I noticed that the plugin won't reinitialize after getting user consent (in case they choose non-personalized ads), I couldn't find "is_personalized" parameter or something like that in your initialization function (https://github.com/Poing-Studios/godot-admob-ios/blob/master/plugin/admob/src/main/admob.mm#L143), so I'm curious how do you forward consent to the Google Mobile Ads SDK?
I'm using https://github.com/Shin-NiL/Godot-Android-Admob-Plugin + https://github.com/pschw/GodotConsentPlugin for the Android version of my app and I'm able to reinitialize the admob plugin with the "is_personalized" parameter set to false when users don't consent, Shin-NiL's plugin also configures GADRequest (https://github.com/Shin-NiL/Godot-Android-Admob-Plugin/blob/main/admob-plugin/godotadmob/src/main/java/shinnil/godot/plugin/android/godotadmob/GodotAdMob.java#L115) as described in the admob documentation (https://developers.google.com/admob/ump/android/quick-start#forward-consent / https://developers.google.com/admob/ump/ios/quick-start#forward-consent). I just want to make sure your plugin works the same to comply with GDPR, since I'm thinking of using it in android too instead of Shin-NiL's.
Btw, the consent works as expected with "Test Europe User Consent" enabled or using a vpn, but when testing it in my location (Colombia), it still shows a window but only with a continue button and a different text that says something like "ads keep our app free, select allow in the next screen...", then it shows ATT consent as expected (I use another plugin for that), is that intended? because the console prints "Consent Form Available" and then "User consent obtained. Personalization not defined." when I was actually expecting "Consent Form not Available" or "User consent not required. For example, the user is not in the EEA or the UK.".
Hope I made myself understood, and please let me know if these are bugs or missing features, I'd like to contribute but idk much about those languages nor creating plugins, anyways, thanks for reading.
Plugin v2.0.4
Editor plugin v1.2.4
Godot v3.4.2
Hello, thank you for using our plugin!
I couldn't find "is_personalized" parameter or something like that in your initialization function
We don't use is_personalized parameter because when I developed this part of GDPR, in the documentation of AdMob I remember it said that it was deprecated to do it this way, so we only use UMP and to talk about whether it is personalized or not, you need to change it within the AdMob application: http://admob.google.com/
I just want to make sure your plugin works the same to comply with GDPR
Yes it does, but i have not done much tests because i don't live in Europe.
From what I analyzed in the codes, I don't know how "GodotConsentPlugin" will communicate with "Godot-Android-Admob-Plugin" can you tell me how you are using these two Plugins?
For example, if in AdMob you defined by code to always be isPersonalized = true, from what I understand, regardless of what the user selects, it will always be personalized, I believe that what the "Forward consent" is used for the game to always demonstrate personalized ads or not
I saw that in the code it is considering these 3 consent statuses:
enum CONSENT_STATUS {
UNKNOWN,
NON_PERSONALIZED,
PERSONALIZED
}
But I remember that when I tested as i was developing, regardless of whether the user clicked "Consent" or "Do not consent", AdMob can't know what he clicked, so for me there's no way to know if it was granted or not, but I'll do some tests to guarantee
Check the documentation: https://developers.google.com/admob/ump/android/quick-start#present_the_form_if_required
There are only four Status: https://developers.google.com/admob/ump/android/api/reference/com/google/android/ump/ConsentInformation.ConsentStatus
ConsentStatus.UNKNOWN: Unknown consent status.
ConsentStatus.REQUIRED: User consent required but not yet obtained.
ConsentStatus.NOT_REQUIRED: User consent not required. For example, the user is not in the EEA or the UK.
ConsentStatus.OBTAINED: User consent obtained. Personalization not defined.
but when testing it in my location (Colombia), it still shows a window but only with a continue button and a different text that says something like "ads keep our app free, select allow in the next screen..."
It should not happen, maybe you forgot to export again? I will test it, by the way i also live on South America.
then it shows ATT consent as expected (I use another plugin for that), is that intended?
I don't think so
because the console prints "Consent Form Available" and then "User consent obtained. Personalization not defined."
Oh, maybe as you have accepted before "being" in europe, what you selected was saved, for that you would need to reset the consent using this method: https://github.com/Poing-Studios/godot-admob-ios/blob/master/plugin/admob/src/main/admob.mm#L97
Also: I've done some research now and it doesn't seem to be deprecated to declare "npa=1", they added it to the updated "EU Consent" page maybe it's a good idea to add the "isPersonalized" parameter but I really don't know if it's recommended since AdMob says this on end of page: https://developers.google.com/admob/ump/ios/quick-start#forward-consent
If non-personalized ads are requested, the ad request URL currently includes &npa=1. However, note that this is an internal implementation detail of the Google Mobile Ads SDK and is subject to change.
If you want to help me contribute to this part of the GDPR, you can join the Discord link: https://github.com/Poing-Studios/godot-admob-android/discussions/63
This is something I really need help with as the AdMob documentation isn't very good and as I don't live in Europe it just makes it more difficult
Also: Our Android and iOS Plugin code should be identical, so GDPR should work the same for Android and iOS
I don't recommend using multiple Plugins for the "same" purpose because in addition to being bad to maintain the plugins, you will probably have to rewrite code for each operating system, Android and iOS
Maybe this old issue can help you in any way: https://github.com/Poing-Studios/godot-admob-android/issues/57
can you tell me how you are using these two Plugins?
So this is how I use "GodotConsentPlugin" and reinitialize "Godot-Android-Admob-Plugin": https://github.com/bouncymarble/bouncymarble.github.io/blob/main/ConsentPlugin.gd
Btw, I had to add a neutral age screen that asks the user for their age before asking for GDPR consent (Google Play rejected my app the first time because consent shouldn't be shown to children under 13).
I saw that in the code it is considering these 3 consent statuses
Yeah I was wondering why the statuses are different, also the form interface is very different as you can see in the demo video (https://youtu.be/PJ2H8ZK8O_w), but I just realized they actually use "Personalized Ad Consent SDK" which is now deprecated (https://developers.google.com/admob/android/eu-consent/) instead of UMP SDK, so I think I must migrate to your plugin.
so we only use UMP and to talk about whether it is personalized or not, you need to change it within the AdMob application: http://admob.google.com/
Ok, should I set it to non-personalized or keep it personalized (default)?
Oh, maybe as you have accepted before "being" in europe, what you selected was saved
Hmm you might be right, maybe I forgot to reset it. I just need to make sure the user is not in europe (with the signal "consent_info_update_success" or "consent_status_changed") so I don't show them a "data usage" button from my game settings to update their consent.
you can join the Discord link
Thanks for the link.
I don't recommend using multiple Plugins for the "same" purpose because in addition to being bad to maintain the plugins, you will probably have to rewrite code for each operating system, Android and iOS
Yeah I've been trying to merge my Android and iOS projects with your plugin so it's easier for me to release updates.
Maybe this old issue can help you in any way: Poing-Studios/godot-admob-android#57
Well, I've had no issues updating the consent for european users, just added a "data usage" button in my game settings panel with this code:
func _on_TB_UpdateGDPR_button_up(): MobileAds.reset_consent_state() MobileAds.request_user_consent()
Although it still shows the button for non european users, but maybe it's due to the above mentioned, so I need to test it again. Thank you so much for your help.
So this is how I use "GodotConsentPlugin" and reinitialize "Godot-Android-Admob-Plugin": https://github.com/bouncymarble/bouncymarble.github.io/blob/main/ConsentPlugin.gd
I think this will not work, i just tested here on my phone and the buttons "Consent" and "Do not Consent" have the same result: Get the ConsentStatus.OBTAINED which means this: https://developers.google.com/admob/ump/android/api/reference/com/google/android/ump/ConsentInformation.ConsentStatus#public-static-final-int-obtained.
I had to add a neutral age screen that asks the user for their age before asking for GDPR consent (Google Play rejected my app the first time because consent shouldn't be shown to children under 13).
I didn't know that, maybe we can include this in the plugin or create a new plugin to display this part of getting the user's age and thus displaying the form or not
also the form interface is very different as you can see in the demo video (https://youtu.be/PJ2H8ZK8O_w)
Yes, i think this is a deprecated version
Ok, should I set it to non-personalized or keep it personalized (default)?
Unfortunately I don't know the answer to this question :(
But where you can configure this is here:
Although it still shows the button for non european users
Remember to unmark this button:
If you mark this, whenever the user is or not on EEA, it will show the Form
Closing this as there's no need to reinitialize the plugin, it seems the UMP handles user choice and somehow forwards it to the Google Mobile Ads SDK. Although there's this issue when "don't consent" button is pressed (or "manage options" and then "confirm choices" without selecting any consent option), ads won't load due to error code 3 (ERROR_CODE_NO_FILL), but this has nothing to do with the plugin, so the only way to load ads is by resetting and updating consent. More info about what admob requires to load either personalized or non-personalized ads: https://support.google.com/admob/answer/9760862?hl=en&ref_topic=9756841
Also, here are the answers to some questions about UMP and GDPR that @gumaciel and I asked on the Google Mobile Ads SDK Technical Forum: https://groups.google.com/g/google-admob-ads-sdk/c/xfLEdy06pyc
the console prints "Consent Form Available" and then "User consent obtained. Personalization not defined." when I was actually expecting "Consent Form not Available" or "User consent not required. For example, the user is not in the EEA or the UK."
Regarding this issue, it turns out I didn't forget to unmark "Test Europe User Consent", and what is actually triggering this message "Consent Form Available" is the IDFA form that is required on iOS, no matter location. On Android it returns "Consent Form not Available" as expected if not in the EEA.
Btw, I had another issue when trying to start my app with no internet connection, the plugin tries to initialize and returns status 0 which I thought was ok, but @gumaciel tested it too and always gets status 1 (MobileAds.AdMobSettings.INITIALIZATION_STATUS.READY). The thing is when I turn on internet, initialize() keeps returning status 0, it seems like it can only be called once. Anyways, what I had to do is make sure there's internet connection to initialize the plugin. For this I use Timer and HTTPRequest nodes, here's the code in case anyone finds it useful:
func _on_Timer_CheckInternet_timeout():
check_internet()
func check_internet():
if http_request.get_http_client_status() != 0:
return
http_request.request("https://google.com")
func _on_HTTPRequest_request_completed(result, response_code, _headers, _body):
if result == 0 and response_code == 200:
if not MobileAds.get_is_initialized():
MobileAds.ready_to_initialize()
| gharchive/issue | 2022-02-04T05:13:28 | 2025-04-01T04:32:58.257890 | {
"authors": [
"bouncymarble",
"gumaciel"
],
"repo": "Poing-Studios/godot-admob-ios",
"url": "https://github.com/Poing-Studios/godot-admob-ios/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
167625551 | Error when starting program
Error:
A JavaScript error occurred in the main process
Uncaught Exception:
Error: spawn python ENOENT
at exports._errnoException (util.js:949:11)
at Process.ChildProcess._handle.onexit (internal/child_process.js:182:32)
at onErrorNT (internal/child_process.js:348:16)
at _combinedTickCallback (internal/process/next_tick.js:74:11)
at process._tickCallback (internal/process/next_tick.js:98:9)
Operating System:
OSX El Capitan
Same error when starting.
OS : Windows 7
Same error, windows 10 x64, latest commit 5f96174.
screenshot shows fixes applied to this issue in past are present
| gharchive/issue | 2016-07-26T14:44:38 | 2025-04-01T04:32:58.278509 | {
"authors": [
"JVenberg",
"pyr0ball",
"tzull"
],
"repo": "PokemonGoF/PokemonGo-Bot-Desktop",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot-Desktop/issues/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
167725296 | KeyError: 'fort_id'
KeyError: 'fort_id'
After #1039 change i update my dev version now my problem is that .
Other Information
OS: Ubuntu
Git Commit: 9c99cc0de7e5b0c35762cae9b657c495d32e3ade
same here
`
[23:14:59] Starting PokemonGo Bot....
[23:14:59] Scanning area for objects....
[23:14:59] Walking from (51.500575, -0.127703) to [51.500575, -0.127703] 0.00km
[████████████████████████████████████████] 100%
[23:15:01] Attempting to evolve all pokemons ...
[23:15:04] Something rustles nearby!
Traceback (most recent call last):
File "./pokecli.py", line 246, in
main()
File "./pokecli.py", line 237, in main
bot.take_step()
File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/init.py", line 46, in take_step
self.work_on_cell(cell, location)
File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/init.py", line 189, in work_on_cell
if self.catch_pokemon(pokemon) == PokemonCatchWorker.NO_POKEBALLS:
File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/init.py", line 337, in catch_pokemon
return_value = worker.work()
File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/cell_workers/pokemon_catch_worker.py", line 31, in work
response_dict = self.create_encounter_api_call()
File "/home/undg/pokemon/bot/PokemonGo-Bot/pokemongo_bot/cell_workers/pokemon_catch_worker.py", line 400, in create_encounter_api_call
fort_id = self.pokemon['fort_id']
KeyError: 'fort_id'
`
I think who change it on other Issues #1045 #1046 maybe have same problem
Commit just been reverted, please repull. #1065
@aslafy-z Thank you when i change to head 1056 working ;)
@aslafy-z :( after visit pokestore now this problem :(
@Zgrkaralar This is another isue due to your outdated config file. Recreate it from configs/config.json.example
Now problem is this
Please open another issue, and someone close this one please.
Commit has been reverted, please repull. #1065
| gharchive/issue | 2016-07-26T22:16:05 | 2025-04-01T04:32:58.287119 | {
"authors": [
"Zgrkaralar",
"aslafy-z",
"surfaace",
"und3rdg"
],
"repo": "PokemonGoF/PokemonGo-Bot",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/1067",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
170194360 | Error in bot startup
Hello , is experiencing an error when booting the bot, do not know what to do, please help me .
The following image:
Please ask this question in the #help section on slack. This is not an issue with the bot.
| gharchive/issue | 2016-08-09T15:20:18 | 2025-04-01T04:32:58.288637 | {
"authors": [
"RedViper9",
"piai46"
],
"repo": "PokemonGoF/PokemonGo-Bot",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/3302",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
167171249 | Update included yaml import, but missing in requirements.txt
Expected Behavior
Program runs
Actual Behavior
ImportError: No module named yaml
Steps to Reproduce
*Clone repo
*Install
*Run
Other Information
OS: Mac OSX
Git Commit: 90225f7133ede69f10167e5ae816391023de0219
Was introduced in 49d3ef6
duplicate and already fixed in #355
| gharchive/issue | 2016-07-23T06:43:27 | 2025-04-01T04:32:58.290645 | {
"authors": [
"eiselems"
],
"repo": "PokemonGoF/PokemonGo-Bot",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/356",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
167182812 | infinite loop if Out of pokeballs
[#] Something rustles nearby!
[#] A Wild Pidgey appeared! [CP 25]
[x] Out of pokeballs...
[#] A Wild Zubat appeared! [CP 167]
[x] Out of pokeballs...
[#] A Wild Pidgey appeared! [CP 25]
[x] Out of pokeballs...
[#] A Wild Zubat appeared! [CP 167]
[x] Out of pokeballs...
[#] Something rustles nearby!
[#] A Wild Pidgey appeared! [CP 25]
[x] Out of pokeballs...
[#] A Wild Zubat appeared! [CP 167]
[x] Out of pokeballs...
[#] A Wild Pidgey appeared! [CP 25]
[x] Out of pokeballs...
[#] A Wild Zubat appeared! [CP 167]
[x] Out of pokeballs...
[#] Something rustles nearby!
[#] A Wild Pidgey appeared! [CP 25]
[x] Out of pokeballs...
[#] A Wild Zubat appeared! [CP 167]
[x] Out of pokeballs...
[#] A Wild Pidgey appeared! [CP 25]
[x] Out of pokeballs...
[#] A Wild Zubat appeared! [CP 167]
[x] Out of pokeballs...
This should be fixed with the newest dev branch.
| gharchive/issue | 2016-07-23T12:46:06 | 2025-04-01T04:32:58.295026 | {
"authors": [
"Hatsunyan",
"s1x"
],
"repo": "PokemonGoF/PokemonGo-Bot",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot/issues/427",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1510196923 | A FAQ would be awesome
Reviewing the discord chat reveals that many of the same questions are asked and answered over and over. A linkable FAQ would be great
Heres a start
FAQ
Is Pokemon Unity still in development?
Yes
Is the Unity part of this project being worked on?
No
Is there a more up to date version and runnable version of this?
Not that we know
How do I start developing my own Pokemon game?
Thats completely up to you. The maintained Pokemon code in this repository is actually NOT related to Unity at all. This means someone can take the code here and start their own Pokemon game in any way that supports loading C# DLL libraries. Unity is just 1 way :)
What relevance do the folders other than the Pokemon Unity folder have?
The repo contains several projects. Each folder (probably) represents one of those projects. The Pokemon Unity folder is just where the Unity related files are found
TODO: add breakdown of contents of each parent folder
Unity FAQ
Why is my screen pink?
Unity metafiles were broken sometime in the past. Because of this, images in (probably) all scenes are corrupted and show up pink
How do I learn Unity?
There are several great tutorials on YouTube to learn Unity. To name a few
Brackeys
Code Monkey
Jason Weimann
Unity
Sebastian Lague
Game Maker's Toolkit
Why isnt the Unity side of things being worked on?
There is only 1 developer currently active on this project. Their focus is on the non-Unity Pokemon framework
Why is my screen pink?
The answer can be a pretty lengthy one, but it's just a unity issue with the .meta files. The why, the how, and the solution is discussed here:
https://discord.com/channels/285560546244427777/285565647507685379/1046810010132435067
Why isnt the Unity side of things being worked on?
I wouldnt say it's because i'm the only dev, but that we dont have any "Unity-developers", it's way more appealing to write programming code, than to struggle with getting unity to behave nice. Those that are comfortable with unity, are mostly working on their own private projects... It's not easy to manage a unity project in git, as there's lots of problems that occur just on the .meta files, if the team isnt in sync or on the same page, in terms of development style/practice, and communication (overlapping edits in the same scene, across different commits can cause major problems when a conflict emerges).
At the end of the day... the project is geared more towards being an API framework, than a game kit (a playable unity demo/sample used as a foundational resource to start you game from).
How to download?
This is a Github question, not really specific to project. But since a lot of people new to the project are also new to github, i recommend just reading this guide and following the steps. Better than downloading a zip, and the frustrations that come after from not receiving bug fixes and updates.
https://docs.github.com/en/desktop/contributing-and-collaborating-using-github-desktop/adding-and-cloning-repositories/cloning-a-repository-from-github-to-github-desktop
or...
https://docs.github.com/en/repositories/creating-and-managing-repositories/cloning-a-repository
Downloading zips aren't always the answers because some files don't get moved over into zip
There are settings on some projects on repository to limit what will and will not go into a "zip" download. Though I don't believe I'm using those settings...
When you clone a project, it bypasses those settings and gives you a one-for-one copy of the project and all of its contents
Literally, "cloning" instead of downloading a "package"
This was an answer i gave to a user in discord chat. But when you "clone" the project, the project lives on your computer as well as on the git cloud it came from... so if a problem is "fixed" on the server, you dont need to download the entire project over again just for the files that were corrected (also, you dont need to worry about your changes be lost or deleted). If you make any edits and want to share it with the project for others to download, then this is also the same steps needed in order to do that.
I downloaded the Github, like instructed above, but how do i play it? I want to experience a pokemon game inside unity...
This isnt a game kit, nor a pokemon clone. It's a resources for your game to implement pokemon features (like a plug-in or API). The resources are downloaded assets you can add to your existing or new project, and shortcut the need to have to program all the pokemon logic yourself from zero. This is a transferrable drag-and-drop library, and for the time being, there really isnt any flagship game to showcase or demo the content in usage. (Unless you wish to be the first, and link back to us here, for us to tag you... no one contributing back to this repository is a "game developers" :shrug: )
What about the youtube video that i saw, where the game was running in unity? How do i download and play that?
That's found in the release tab, as an .exe download...
https://github.com/PokemonUnity/PokemonUnity/releases
you can grab the project that was created by the original author of this project repository. But it's dated back to 2015, and has a lot of incomplete code that we're attempting to resolve by rewriting and redoing the code source from the ground up. So if you want the latest changes and fixes, there is no "game" to play; and yes... you will download a git repository with a unity project that may contain a lot of errors (preventing you from playing any scene).
| gharchive/issue | 2022-12-24T21:52:57 | 2025-04-01T04:32:58.319506 | {
"authors": [
"Xerner",
"herbertmilhomme"
],
"repo": "PokemonUnity/PokemonUnity",
"url": "https://github.com/PokemonUnity/PokemonUnity/issues/129",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
393752346 | [Sample app Kotlin] separate sample-kotlin module
Creates a separate sample-kotlin module. As a start, ScanActivity and DeviceActivity are implemented, with parts of functionalities outcommented to make for an easier review.
Also all layouts have been copied over from the Java sample and cleaned up a little bit.
I have added Work In Progress to the title as I have seen that there are some TODOs in the code. Feel free to rename when you finish
Have a Happy New Year! :)
In this case I will put it on top of my Open Source priority list
One last thing — I have just released version 1.8.0 which is now using androidx. I have migrated the java sample app already. Could you also update the kotlin one? From now on I do not expect more such changes
| gharchive/pull-request | 2018-12-23T14:03:50 | 2025-04-01T04:32:58.362151 | {
"authors": [
"dariuszseweryn",
"lukaszkalnik"
],
"repo": "Polidea/RxAndroidBle",
"url": "https://github.com/Polidea/RxAndroidBle/pull/520",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1069058593 | Pack load error icon
Including a small little icon after pressing Reload Packs if any of the packs failed to load would make it easy to see if you need to check the console. Additionally, a check mark could indicate no pack errors.
I plan on redoing the majority of that functionality, as I don't like the tabs feature. Thoughts?
| gharchive/issue | 2021-12-02T02:23:37 | 2025-04-01T04:32:58.393470 | {
"authors": [
"Astrashh",
"solonovamax"
],
"repo": "PolyhedralDev/BiomeTool",
"url": "https://github.com/PolyhedralDev/BiomeTool/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
702390706 | [lit-html] Clarify behavior of classMap
The test for classMap say that dynamic classes can not override and remove static classes, but this is only true on the first render. Subsequent renders will remove static classes. This is clearly broken, but what is the right behavior? To allow overriding or not? If we're going to change this, 2.0 is the time.
I think the most important thing is that classMap can remove the classes it previously set. I don't think the behavior when you have a conflict between a static class and a classMap class is important. We should just add docs that you shouldn't do that.
We can't change this behavior w/o potentially breaking people though, so I do want to settle on what we want, even if we recommend against class clashes like this.
| gharchive/issue | 2020-09-16T02:07:11 | 2025-04-01T04:32:58.413077 | {
"authors": [
"justinfagnani",
"sorvell"
],
"repo": "Polymer/lit-html",
"url": "https://github.com/Polymer/lit-html/issues/1278",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
144036224 | Add carbon elements
Let's add carbon elements to the catalog. Should close https://github.com/PolymerElements/carbon-route/issues/57
This looks good locally. I needed to stamp v0.8.1 of carbon-route to get our updated docs (v0.8.0 had invalid markdown).
The demo looks pretty good but I'd like to wait for https://github.com/PolymerElements/carbon-route/issues/53 to land before we push this out though, so that there's a useful URL bar in the demo. After that lands (and we stamp v0.8.2) I'd say that this is good to go.
great! ^ cc @tjsavage
Sounds good - let's wait for it to land then push.
Ok, it's landed and released as v0.8.2, LGTM!
| gharchive/pull-request | 2016-03-28T17:45:46 | 2025-04-01T04:32:58.417596 | {
"authors": [
"blasten",
"rictic",
"tjsavage"
],
"repo": "Polymer/polymer-element-catalog",
"url": "https://github.com/Polymer/polymer-element-catalog/pull/264",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
197519963 | Polymer 2.0 - Dblclick is not bubbling up
http://codepen.io/anon/pen/ObYaqX?editors=1000
If you look at the example you'll notice a blue box with black box inside
The blue box is listening to the dblclick event, however when you click the black box the dblclick event does not trickle up to the blue box... any idea why?
(If you replace dblclick with click you'll see it works that way, which is odd)
Here's a slightly updated example that works x-browser: http://codepen.io/anon/pen/wJBJyd?editors=1000.
Closing since this no longer seems to repro on the example. Please re-open if the problem resurfaces.
| gharchive/issue | 2016-12-26T00:17:03 | 2025-04-01T04:32:58.420488 | {
"authors": [
"afroradiohead",
"sorvell"
],
"repo": "Polymer/polymer",
"url": "https://github.com/Polymer/polymer/issues/4234",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
170665613 | Do notify once per property at most
If you have multiple behaviors defining the same property, all setting notify
to true, Polymer just blindly added multiple effects. This resulted in multiple
DOM events sent upwards.
The fix is to just check if there is already a notify effect registered for a
given property, before we register the effect.
Reference issue #3418.
Fixes #1791.
Fixes PolymerElements/paper-input#66
I hit this bug that this solves with paper-input on-value-change firing twice for every key press...whatever became of this since it has not been merged?
Closing per https://github.com/Polymer/polymer/issues/1791#issuecomment-321436929
| gharchive/pull-request | 2016-08-11T14:53:11 | 2025-04-01T04:32:58.422992 | {
"authors": [
"TimvdLippe",
"dman777",
"kaste"
],
"repo": "Polymer/polymer",
"url": "https://github.com/Polymer/polymer/pull/3856",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
146141329 | 1.14.9 fails on Windows
Older version 1.14.8 works, but new version 1.14.9 returns:
ERROR finding /Work/QLine/WebClients/AppPortal/elements/bower_components/bower_components/bower_components/bower_components/web-animations-js/web-animations-next-lite.min.js
There is a package neon-animation which references web-animations-js.
+1 Getting the same issue
+1 Same issue on Mac
+1 Same issue on ubuntu
Same here. In my case I have a local element in project\component folder (not in bower_components) that links ../bower_components/paper-dialog/paper-dialog.html and it ends up looking for project\component\bower_components\paper-dialog\import\neon-animation\import\neon-animation\import\web-animations-js\web-animations-next-lite.min.js
This is a really bad since gulp-vulcanize will automatically get the latest 1.14.x version of this tool.
Working on it
Unpublished v1.14.9 until I can get to the bottom of the breakage.
| gharchive/issue | 2016-04-05T23:49:02 | 2025-04-01T04:32:58.431745 | {
"authors": [
"ElsewhereGames",
"azakus",
"davidlwatsonjr",
"dudabone",
"govis",
"kwaclaw",
"xwsxethan"
],
"repo": "Polymer/vulcanize",
"url": "https://github.com/Polymer/vulcanize/issues/332",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1975120904 | feat: 🎸 add NFT module
JIRA Link
✅ Closes: DA-913
Changelog / Description
add endpoint to allow management of NFT collections, including issuing, redeeming and fetching NFT details
Checklist -
[ ] New Feature ?
[ ] Updated swagger annotation (if API structure is changed) ?
[ ] Unit Test (if possible) ?
[ ] Updated the Readme.md (if required) ?
:tada: This PR is included in version 4.0.0-alpha.6 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2023-11-02T22:11:07 | 2025-04-01T04:32:58.464405 | {
"authors": [
"polymath-eric"
],
"repo": "PolymeshAssociation/polymesh-rest-api",
"url": "https://github.com/PolymeshAssociation/polymesh-rest-api/pull/217",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2573591900 | docs: ✏️ update readme + export join creator params
Description
Update readme + fix doc generator warning
Breaking Changes
JIRA Link
Checklist
[ ] Updated the Readme.md (if required) ?
/fast-forward
:tada: This PR is included in version 26.0.0-alpha.2 :tada:
The release is available on:
npm package (@alpha dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 26.0.0-beta.1 :tada:
The release is available on:
npm package (@beta dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 26.0.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2024-10-08T15:48:03 | 2025-04-01T04:32:58.472127 | {
"authors": [
"polymath-eric",
"prashantasdeveloper"
],
"repo": "PolymeshAssociation/polymesh-sdk",
"url": "https://github.com/PolymeshAssociation/polymesh-sdk/pull/1329",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
962684342 | Removes YouTube Premium fullscreen pop-up
Rename the tweak to NoYTAnnoyances
https://github.com/PoomSmart/NoYTPremium
| gharchive/pull-request | 2021-08-06T12:16:06 | 2025-04-01T04:32:58.481165 | {
"authors": [
"PoomSmart",
"nicknitewolf"
],
"repo": "PoomSmart/YouAreThere",
"url": "https://github.com/PoomSmart/YouAreThere/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
701286565 | nom start errors - please help
Failed to compile.
./node_modules/material-ui/styles/withStyles.js
Module not found: Can't resolve '@babel/runtime/core-js/map' in '/home/test/downloads/nifty-game/dapp/node_modules/material-ui/styles'
/home/test/downloads/nifty-game/dapp/scripts/start.js:11
throw err;
^
Can you help, please?
can you fixed? that mean material-ui is not in the map in babel..
| gharchive/issue | 2020-09-14T17:23:33 | 2025-04-01T04:32:58.509373 | {
"authors": [
"JoseMariani",
"rocheston"
],
"repo": "PortalNetwork/nifty-game",
"url": "https://github.com/PortalNetwork/nifty-game/issues/73",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2126726853 | fix: respect cache max age from configs
Problem:
max_age passed in configs is not getting respected. Due to this, the sent max age is not applied to cache sometimes.
Example config:
{
"cache": {"mode": "simple", "maxAge": 60}
}
Using the above config, cache should be expired after 60 seconds.
Should it be max_age?
Yes. Fixed the example.
| gharchive/issue | 2024-02-09T09:02:33 | 2025-04-01T04:32:58.511373 | {
"authors": [
"VisargD",
"vrushankportkey"
],
"repo": "Portkey-AI/gateway",
"url": "https://github.com/Portkey-AI/gateway/issues/201",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1105198247 | Min version of Posthog to run?
What's the minimum version of Posthog this plugin runs on? When I tried on Posthog 1.29, I saw this ERROR in the plugin log:
Task "Export historical events" not found for plugin "Migrator 3000 (Beta)" with config id [object Object]
I have confirmed this should work on PostHog 1.29.0.
Can you send me a screenshot of your system status on the instance you've enabled this plugin in?
Sorry, I misspoke, I was on 1.27.0.
I can try to upgrade to 1.29.0 or 1.29.1 as I've been having issues upgrading to 1.30.0 on Heroku.
On 1.27.0 here's my system status:
1.29.1 is working for me! Cheers!
| gharchive/issue | 2022-01-16T21:57:49 | 2025-04-01T04:32:58.515898 | {
"authors": [
"avk",
"yakkomajuri"
],
"repo": "PostHog/posthog-plugin-migrator3000",
"url": "https://github.com/PostHog/posthog-plugin-migrator3000/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1305121279 | Add param for tx value col, let user data override database, & bug fix
Make the value column in tx data tables a parameter. This lets users pick
firm/non-firm or custom names.
Let user tx data override the database table
Add docstring to agg_transmission_constraints table
Add error messages
Expand testing
Codecov Report
Merging #198 (51184ee) into develop (cc7da97) will decrease coverage by 0.65%.
The diff coverage is 54.78%.
@@ Coverage Diff @@
## develop #198 +/- ##
===========================================
- Coverage 71.58% 70.93% -0.65%
===========================================
Files 16 16
Lines 3258 3354 +96
===========================================
+ Hits 2332 2379 +47
- Misses 926 975 +49
Impacted Files
Coverage Δ
...owergenome/run_powergenome_multiple_outputs_cli.py
0.00% <0.00%> (ø)
powergenome/GenX.py
56.32% <8.89%> (-5.80%)
:arrow_down:
powergenome/generators.py
82.62% <69.70%> (-0.22%)
:arrow_down:
powergenome/util.py
69.52% <76.32%> (+0.79%)
:arrow_up:
powergenome/transmission.py
88.54% <80.95%> (+0.38%)
:arrow_up:
powergenome/eia_opendata.py
92.86% <100.00%> (+0.15%)
:arrow_up:
powergenome/load_profiles.py
78.20% <100.00%> (+1.03%)
:arrow_up:
powergenome/nrelatb.py
89.15% <100.00%> (-0.05%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 6ee7432...51184ee. Read the comment docs.
| gharchive/pull-request | 2022-07-14T18:00:20 | 2025-04-01T04:32:58.603341 | {
"authors": [
"codecov-commenter",
"gschivley"
],
"repo": "PowerGenome/PowerGenome",
"url": "https://github.com/PowerGenome/PowerGenome/pull/198",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
811206566 | Update request - Determine which groups a user is a member of in a logon script
Summary of the update request
Link to the post in the old blog:
https://devblogs.microsoft.com/scripting/how-can-i-determine-which-groups-a-user-belongs-to/
Description of what needs to be changed:
Needs to show implicit remoting of AD cmdlets from the DC, then use get-ADPrincipalGroupMembership to get the group memberships. Use whoami.exe .upn, to get the date, then look the user up using the UPN.
PR #45 adds this to the blog, so I am closing this issue.
| gharchive/issue | 2021-02-18T15:30:58 | 2025-04-01T04:32:58.606048 | {
"authors": [
"doctordns"
],
"repo": "PowerShell/Community-Blog",
"url": "https://github.com/PowerShell/Community-Blog/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
715489049 | Error in last two SuppressMessageAttribute examples in readme.MD
In Suppressing rules in README.md, the last two examples omit the CheckID parameter. When used like this, the constructor for SuppressMessageAttribute throws an exception
Eg. when using the example provided under 'Suppress violations in all the functions' in a script:
[Diagnostics.CodeAnalysis.SuppressMessageAttribute('PSAvoidUsingWriteHost', Scope='Function', Target='*')]
Param()
Write-Output 'Hello world'
When executing this script, PowerShell throws this error:
Cannot find an overload for ".ctor" and the argument count: "1".
The example starting on line 302 should read:
[Diagnostics.CodeAnalysis.SuppressMessageAttribute('PSAvoidUsingWriteHost', '', Scope='Function', Target='*')]
Param()
and on line 307
[Diagnostics.CodeAnalysis.SuppressMessageAttribute('PSAvoidUsingWriteHost', '', Scope='Function', Target='start-b*')]
Param()
Thanks @hairyken you are correct, we will make this fix
| gharchive/issue | 2020-10-06T09:10:57 | 2025-04-01T04:32:58.634458 | {
"authors": [
"SydneyhSmith",
"hairyken"
],
"repo": "PowerShell/PSScriptAnalyzer",
"url": "https://github.com/PowerShell/PSScriptAnalyzer/issues/1592",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
610716632 | [Feature Request] Better display of UNC paths
Summary of the new feature/enhancement
Currently, when the CWD is a UNC path, it is displayed like so
PS Microsoft.PowerShell.Core\FileSystem::\\wsl$\Ubuntu\home\stan\projects\personal\yarntest>
Would it be possible to display it as a normal path instead? like so
PS \\wsl$\Ubuntu\home\stan\projects\personal\yarntest>
Or is that something my profile broke?
Proposed technical implementation details (optional)
Not sure how complex this would be to make behave as you expect, but you can always use the New-PSDrive cmdlet to add a simpler address to the path, e.g.:
New-PSDrive -Provider FileSystem -Root \\wsl$\Ubuntu\ -Name WSL
Set-Location WSL:
Well my request was mostly to improve the default since it does look a bit "wrong".
Something you can add to your profile:
function prompt {
if ($null -eq $PWD.Drive -and $PWD.Provider.Name -eq 'FileSystem') {
return "PS $($PWD.ProviderPath)> "
}
return "PS $($PWD.Path)> "
}
Maybe worth adding as a default.
No worries, @thlac is working on a fix already.
Bad bot
bad bot
bad bot
| gharchive/issue | 2020-05-01T11:58:20 | 2025-04-01T04:32:58.644798 | {
"authors": [
"SeeminglyScience",
"Stanzilla",
"vexx32"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/12544",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
170208284 | Remove dead code in StopJob.cs (line 164)
Jason found this through code inspection. This code will never run because of the previous if statement and in fact the code is duplicated in the HandleStopJobCompleted() where it is correct. So we can simply remove the code at line 164.
var parentJob = job as ContainerParentJob;
if (parentJob != null && parentJob.ExecutionError.Count > 0)
{
foreach (
var e in
parentJob.ExecutionError.Where(
e => e.FullyQualifiedErrorId == "ContainerParentJobStopError"))
{
WriteError(e);
}
}
If we have a "code cleanup" label at some point, this would fall under it.
Hi @powershell, I just submit a PR (#5091) for this issue.
Hmm, I seem to be missing something with this issue and the associated PR. HandleStopJobCompleted appears to be only used for Job2 instances and not Job instances. While the code is duplicated, removing it for Job instances doesn't appear to be correct.
| gharchive/issue | 2016-08-09T16:13:26 | 2025-04-01T04:32:58.646950 | {
"authors": [
"PaulHigin",
"dantraMSFT",
"joeyaiello",
"travisty-"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/1696",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1293148872 | "Failed to create CoreCLR, HRESULT: 0x80070008" when using latest
Prerequisites
[X] Write a descriptive title.
[X] Make sure you are able to repro it on the latest released version
[X] Search the existing issues.
[X] Refer to the FAQ.
[X] Refer to Differences between Windows PowerShell 5.1 and PowerShell.
Steps to reproduce
In my build script I have a step that spawns a docker container based on the powershell:latest image that fails with the error in the title.
Here is the command executed on the build server
docker run --rm -e AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY -e COMMIT_SHA1=$CIRCLE_SHA1 --network project_default mcr.microsoft.com/powershell pwsh -command '
Install-Module -Name AWS.Tools.Installer -Force
Install-AWSToolsModule AWS.Tools.S3 -Force -CleanUp
Invoke-WebRequest -Uri http://entrypoint-api/docs/v1.0/schema.json -OutFile /tmp/schema.json
Write-S3Object -File /tmp/schema.json -BucketName "some_bucket" -Key "some_key/$env:COMMIT_SHA1.json" -AccessKey $env:AWS_ACCESS_KEY_ID -SecretKey $env:AWS_SECRET_ACCESS_KEY
'
The command has worked without any issue until few days ago. Then we started getting the error specified above: "Failed to create CoreCLR, HRESULT: 0x80070008"
Unable to find image 'mcr.microsoft.com/powershell:latest' locally
latest: Pulling from powershell
018f9d1d: Pulling fs layer
Digest: sha256:844fb73db69ed0bbe82d159d4ba50aea06ef4d42bda766c52f2928683f5c8d09
Status: Downloaded newer image for mcr.microsoft.com/powershell:latest
Failed to create CoreCLR, HRESULT: 0x80070008
Exited with code exit status 137
CircleCI received exit code 137
After some testing, I noticed that the issue is only happening when using latest which is pointing at ubuntu-22.04.
If I use 7.2.0-debian-10, 7.2-debian-10, 7.2-alpine-3.14, alpine-3.14, ubuntu-22.04 it works without any problem.
In issue #13166 it was suggested to set COMPlus_EnableDiagnostics to 0. This didn't help.
Expected behavior
Being able to run a container with latest image of powershell.
Actual behavior
Container can't start
Error details
Unable to find image 'mcr.microsoft.com/powershell:latest' locally
latest: Pulling from powershell
018f9d1d: Pulling fs layer
Digest: sha256:844fb73db69ed0bbe82d159d4ba50aea06ef4d42bda766c52f2928683f5c8d09
Status: Downloaded newer image for mcr.microsoft.com/powershell:latest
Failed to create CoreCLR, HRESULT: 0x80070008
Exited with code exit status 137
CircleCI received exit code 137
Environment data
Don't have it.
Visuals
No response
The tag ubuntu-22.04 doesn't work for me either. (no surprise given that latest points to the same version)
Status: Downloaded newer image for mcr.microsoft.com/powershell:ubuntu-22.04
mcr.microsoft.com/powershell:ubuntu-22.04
Failed to create CoreCLR, HRESULT: 0x80070008
Process exited with code 137
What version of powershell are you using? I'll look into replicating this my side and resolving soon!
@anamnavi the faulty command is launched in a mcr.microsoft.com/powershell container.
Basically it's a step of a job in a CircleCI build. The job uses mcr.microsoft.com/powershell as executor and itself spawns another container based off the same image.
In our tests we didn't need to change the outer container but just the one we were using to execute those commands.
I'm not the OP, but I'm hitting the same issue (with a simpler script, but the same container). My host is a relatively ancient Debian box running docker 19.03. The host does have ~3GB free RAM and running that command on the host just reports a max long value, as does running it inside the container.
Running the full powershell command on a Windows 10 host runs without issues, so it does seem to be something host-specific. Using a different powershell container also works on my ancient host.
@anamnavi Here is what I got in the host container
root@d1e24fe67e4c:~/project# docker run --rm mcr.microsoft.com/powershell
Failed to create CoreCLR, HRESULT: 0x80070008
root@d1e24fe67e4c:~/project# docker-compose up -d
[...]
root@d1e24fe67e4c:~/project# /bin/sh -c 'grep hierarchical_memory_limit /sys/fs/cgroup/memory/memory.stat'
hierarchical_memory_limit 4320133120
root@d1e24fe67e4c:~/project# docker-compose down
[...]
root@d1e24fe67e4c:~/project# /bin/sh -c 'grep hierarchical_memory_limit /sys/fs/cgroup/memory/memory.stat'
root@d1e24fe67e4c:~/project# hierarchical_memory_limit 4320133120
@anamnavi
Here it is
log.txt
The log shows a failure loading /opt/microsoft/powershell/7-lts/libcoreclr.so which is the CoreCLR itself.
@Kralizek Digging into the error HRESULT: 0x80070008 points an issue with a syscall being blocked due to SECCOMP in an container environment. For example, https://github.com/dotnet/runtime/issues/1634 shows that this error happened when .NET called sched_setaffinity to set thread's CPU affinity mask in a Snap confined environment that doesn't have the process-control interface.
You will want to read this article: https://www.linuxserver.io/blog/anatomy-of-an-issue, which points out a similar problem due to using old versions of Docker engine. Their resolution may be helpful to you: https://docs.linuxserver.io/faq#jammy.
@daxian-dbw thanks for the reply. Do you have an idea why older images and images based off different distributions like alpine don't suffer of the same problem?
I'm not sure, but article I referenced to has the following that may be an explanation:
The problem is that if a new syscall gets introduced but Docker doesn't have it included in the profile then your containers can't use it and if, as in this case, the syscall replaces an older one, new distros using it will break previously functional containers.
Does either of the resolutions suggested at https://docs.linuxserver.io/faq#jammy work for you?
@daxian-dbw I'll try when I get back to work. Currently on vacation
deploy:
resources:
limits:
pids: 1
accidently had the same pids twice in my docker-compose.yml, causing this error.
| gharchive/issue | 2022-07-04T13:05:44 | 2025-04-01T04:32:58.662894 | {
"authors": [
"FelixSFD",
"Kralizek",
"Pentadome",
"TravisEz13",
"anamnavi",
"daxian-dbw",
"uecasm"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/17624",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
340355311 | PSCore - Where is the PowerShell Core GA version(s)?
On the releases page, I'm looking for the PowerShell Core 6.0 GA version. But, I can't find it.
So, which of the non-Preview are GA releases.
Or, should I assume that all non-Preview are actually GA's?
I think this should be address better in the Release Notes. My organization only support GA release in their products.
Preview releases are marked as preview, everything else is GA. But I agree, it is not very easy to find out as one has to scroll down quite a bit.
The latest GA release is 6.0.2 here
Maybe it would be better to display the latest GA always at the very top or at least at 2nd place?
We have links on main page https://github.com/PowerShell/PowerShell#get-powershell
@iSazonov
I see now. Of course... "Stable" = "GA". Understood!
:)
@iSazonov
Forgot to mention, my main point is more about the "Release Notes" not mention when the product release reached GA status. I think it would be a "nice" to include in the notes.
:)
I believe this could be added in release check list.
@iSazonov,
That would be Great!!
:)
@SteveL-MSFT Is the request about "GA in Release Notes" addressed? Can we close the issue?
I believe this can be closed
| gharchive/issue | 2018-07-11T18:27:40 | 2025-04-01T04:32:58.668922 | {
"authors": [
"MaximoTrinidad",
"SteveL-MSFT",
"bergmeister",
"iSazonov"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/7271",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
378875524 | Start-Transcript sporadically loses exe output
About half the time there is not output result in the $logFile even though on the console there was plently of output. The other half of the time there is output in the $logFile. It seems very random on if the output works or not. Actually I would say a little less than half the time the helm ouput does NOT appear in the $logFile.
I'm using PowerShell core 6.1 on Windows 10.
Start-Transcript -Force -IncludeInvocationHeader -Path $logFile
try {
helm upgrade --install --wait --force --debug -f ./helm/zookeeper/values.yaml platform-zookeeper ./helm/zookeeper
} finally {
Stop-Transcript
}
This also happen with kubectl commands as well.
This appears to be a similar issue.
http://web.archive.org/web/20130608164153/http://connect.microsoft.com/PowerShell/feedback/details/315875/unable-to-capture-all-session-output-into-a-transcript
Do you get a random output in transcript file run by run?
| gharchive/issue | 2018-11-08T19:09:38 | 2025-04-01T04:32:58.671775 | {
"authors": [
"AceHack",
"iSazonov"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/8215",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
218263174 | Incorrect Stringbuilder usage
StringBuilder doesn't have a constructor that takes a char. In this case, the char is only used to inadvertently define capacity when the intent is to populate the StringBuilder with a newline.
Addresses https://github.com/PowerShell/PowerShell/issues/3457
Looked through the code for StringBuilder and only found this one occurrence of incorrect usage
I'm closing this PR as it is a duplicate of #3458
| gharchive/pull-request | 2017-03-30T17:01:00 | 2025-04-01T04:32:58.673755 | {
"authors": [
"SteveL-MSFT",
"daxian-dbw"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/pull/3461",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
214486198 | Register-PSRepository in non-interactive mode?
Team,
I'm using Register-PSRepository for automation on internal build system. On PS 5.1. the cmdlet prompts and I get the following message(from the build logs):
Exception calling "ShouldContinue" with "2" argument(s): "Windows PowerShell is in NonInteractive mode. Read and Prompt functionality is not available."
The cmdlet doesn't have -Force/-Confirm parameters and it doesn't respect the $ConfirmPreference variable. Is there a workaround for this?
Thanks,
Nedko
@bmanikm That command doesn't seem to download nuget.exe. In my environment, I have been running that command, but I don't get the error about non-interactive mode until I get to the publish-module step, which fails because it's trying to prompt about downloading nuget.exe.
@godefroi Correct, Install-PackageProvider only installs NuGet provider, and it doesn't download nuget.exe. NuGet.exe is required only during Publish-Module or Publish-Script operations.
Please take a look at
https://github.com/PowerShell/PowerShellGet/blob/development/tools/build.psm1#L72 to understand the logic for bootstrapping NuGet.exe in automation scripts.
This is extremely unsatisfying. @bmanikm's suggestion results in downloading NuGet in every iteration/build/deployment!
We end up needing something like this in our deployment scripts?
if(!(Get-PackageProvider NuGet)) {
Install-PackageProvider Nuget -ForceBootstrap -Force
}
It's a bit frustrating that I need to write so many lines for that, and then I have the same problem when installing modules.
In automation scripts, please use the below simple command to bootstrap the NuGet provider. This bootstrap the NuGet provider only if it is not available on the local machine.
Get-PackageProvider NuGet -ForceBootstrap
In non-interactive mode this step requires additional privileges, at least account under which is running this commandlet has to have SeAssignPrimaryTokenPrivileges and SeServiceLogonRight. I've just extended this script by adding SeAssignPrimaryTokenPrivilege options. Currently Register-PSRepository works fine for me from our Chef recipes.
This should be reopened @bmanikm - I have similar use case with Docker (installing pswh on docker image behind proxies) and it would be very welcomed
Here is my Recipe how to upgrade PowerShellGet under the docker
Write-Host
Write-Host "Updating PowerShellGet ..." -ForegroundColor Green
# Original way to install PowerShellGet does not work under docker or without user confirms
#Install-Module PowerShellGet -Repository PSGallery -Force -Confirm -AllowClobber -Scope CurrentUser
# Here is the alternative install script
$nugetPackageProvider = $null
$nugetPackageProvider = Get-PackageProvider NuGet -ErrorAction SilentlyContinue
if (!$nugetPackageProvider) {
$nugetPackageProvider = Install-PackageProvider Nuget -ForceBootstrap -Force
Write-Host "Installed NuGet package provider version $($nugetPackageProvider.Version)"
}
Write-Host "Installing PowerShellGet module..."
$userModulesPath = "$([Environment]::GetFolderPath("MyDocuments"))/WindowsPowerShell/Modules"
New-Item -ItemType Directory -Path $userModulesPath -Force | Out-Null
Save-Module PowerShellGet -Path $userModulesPath
Remove-Module PowerShellGet, PackageManagement -Force
Import-Module PowerShellGet
$psGetModule = Get-Module PowerShellGet
$installedVersion = $psGetModule.Version
I was having a similar issue when trying to run a PowerShell script in my release pipeline that needed access to a module in our private Azure DevOps Artifacts feed, so I was attempting to do:
Register-PSRepository -Name MyRepoName -SourceLocation 'https://pkgs.dev.azure.com/[MyOrganization]/_packaging/[MyFeed]/nuget/v2' -InstallationPolicy Trusted
Install-Package MyPrivatePowerShellModule -InstallUpdate -AllowClobber
However, when the script runs on the MS hosted agent, it ends up in an infinite wait loop outputting:
"C:\windows\System32\WindowsPowerShell\v1.0\powershell.exe" -NoLogo -NoProfile -NonInteractive -ExecutionPolicy Unrestricted -Command ". 'D:\a_temp\527651b0-116e-46aa-a599-141b4dfc9899.ps1'"
Registering PSRepository
[Minimal] [CredentialProvider]DeviceFlow: https://pkgs.dev.azure.com/[MyOrganization]/_packaging/[MyFeed]/nuget/v2
[CredentialProvider]To sign in, use a web browser to open the page https://microsoft.com/devicelogin and enter the code EB64QSSD6 to authenticate.##[section]Starting: PowerShell Script
I came across this feedback post though and was able to get it working with the following code (ensuring $PersonalAccessToken contains a PAT with permissions to read from the provider):
$pat = ConvertTo-SecureString '$PersonalAccessToken' -AsPlainText -Force
$credential = New-Object System.Management.Automation.PSCredential 'Username@DoesNotMatter.com', $pat
Install-PackageProvider NuGet -Scope CurrentUser -Force > $null
if ($null -eq (Get-PSRepository -Name 'MyRepoName' -ErrorAction SilentlyContinue))
{ Register-PSRepository -Name 'MyRepoName' -SourceLocation 'https://pkgs.dev.azure.com/[MyOrganization]/_packaging/[MyFeed]/nuget/v2' -InstallationPolicy Trusted -Credential $credential }
Install-Module -Name 'MyPrivatePowerShellModule' -Repository 'MyRepoName' -Force -AllowClobber -Credential $credential
Import-Module -Name 'MyPrivatePowerShellModule' -Force
The trick was to first do Install-PackageProvider NuGet -Scope CurrentUser -Force, and then also to use Personal Access Token credentials. I hope this helps anybody else who stumbles across this issue.
| gharchive/issue | 2017-03-15T18:35:03 | 2025-04-01T04:32:58.692777 | {
"authors": [
"Jaykul",
"Vit0hA",
"bmanikm",
"deadlydog",
"dmitriyse",
"dmpe",
"f0rt",
"godefroi"
],
"repo": "PowerShell/PowerShellGet",
"url": "https://github.com/PowerShell/PowerShellGet/issues/101",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
171257845 | update-markdownhelp should update the help file name in each of the cmdlets MD and also update the GUID in the module landing page
Steps to reproduce
Create MDs from a MAML file that doesn't have a standard help file name.
Run an update on MDs with the module installed
Expected behavior
Cmdlet MDs should have the external help file name to be updated with original xml help file name from the module.
Also, the landing page should have the GUID updated if not already there.
Actual behavior
The external help file name in the Cmdlets MD remains to be the same as the MAML file that was used to convert to MDs. This poses a problem when we cab the file and release the updated content changes will not take into effect.
v0.6.1
@sankethka could you split this into two issues? These are separate asks.
resolved in 0e7749df
| gharchive/issue | 2016-08-15T20:45:44 | 2025-04-01T04:32:58.708165 | {
"authors": [
"jongeller",
"sankethka"
],
"repo": "PowerShell/platyPS",
"url": "https://github.com/PowerShell/platyPS/issues/193",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
163964292 | BLOCKER: PSSA Errors Need to Be Fixed
Errors from PS script analyzer are causing the tests to fail like this.
This resource module cannot be released until these errors are fixed or suppressed and the tests are passing.
#12
| gharchive/issue | 2016-07-05T22:55:25 | 2025-04-01T04:32:58.712032 | {
"authors": [
"Indhukrishna",
"kwirkykat"
],
"repo": "PowerShell/xAzurePack",
"url": "https://github.com/PowerShell/xAzurePack/issues/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1943743232 | Superhero Penance I image not showing
Hi, I found that some superhero images are not showing, like Penance I , Atom and Goliath. Please assign it to me under hacktoberfest 2023 label. I can work on it.
Found that the Superhero API has some Heros' images missing. Server is responding with a 404.error message. I don't know how to work around this situation now. Please assign it to someone who can work on it.
| gharchive/issue | 2023-10-15T06:22:39 | 2025-04-01T04:32:58.722554 | {
"authors": [
"Shybymh"
],
"repo": "Prajwal0225/My-Favorite-SuperHero",
"url": "https://github.com/Prajwal0225/My-Favorite-SuperHero/issues/187",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
912731849 | Adding the small screen fonts view and changing vote rank
Well detailed description of the change :
I worked on the changing the position of the icons on small screens and also I changed the vote rank so if it is 0 it does not say - 0
Context of the change :
- Why is this change required ?
Does it solve a problem ? (please link the issue)
Type of change :
[ ] Bug fix
[ ] New feature
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
Preview (Screenshots) :
If it is possible, please link screenshots of your changes preview !
Checklist:
[ ] I have read the CONTRIBUTING document.
[ ] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] All new and existing tests passed.
Reviewers
@PraveenMalethia
@PraveenMalethia I dont know if I changed the vote rank to not say -0 the correct way. There might be a simpler way but this works.
| gharchive/pull-request | 2021-06-06T08:59:52 | 2025-04-01T04:32:58.731904 | {
"authors": [
"srikarkandula"
],
"repo": "PraveenMalethia/mumble-vue",
"url": "https://github.com/PraveenMalethia/mumble-vue/pull/42",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1311597055 | Updated prefect_dbt.cloud.jobs import
Updated prefect_dbt.cloud.jobs import to reflect accurate name for prefect_dbt.cloud.jobs import()
Summary
Relevant Issue(s)
Checklist
[ ] Summarized PR's changes in CHANGELOG.md
Thanks! The tests are fixed in https://github.com/PrefectHQ/prefect-dbt/pull/22 so that needs to go in first.
| gharchive/pull-request | 2022-07-20T17:49:14 | 2025-04-01T04:32:58.782903 | {
"authors": [
"ahuang11",
"belasobral93"
],
"repo": "PrefectHQ/prefect-dbt",
"url": "https://github.com/PrefectHQ/prefect-dbt/pull/31",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2287833274 | Remove orphan tabs following deletion of old controllers
Questions
Answers
Description?
While upgrading to PrestaShop v9, remove tabs who lost their parent relation with the previous row deletion. this cleans the table and fixes an issue reported while opening the Stock page.
Type?
bug fix
BC breaks?
Nope
Deprecations?
Nope
Fixed ticket?
Fixes https://github.com/PrestaShop/PrestaShop/issues/35879
Sponsor company
PrestaShopCorp
How to test?
Upgrade from PS 8 to 9, then open the Stock page. It should load without error.
Hi @Quetzacoalt91
Thank you for your PR, I tested it on :
8.1.5 to 9.0.0, that works without any exception
but when I try to upgrade from 8.0.5 to 9.0.0, that what I have :
recording(202).webm
Waiting for your correction
Seen with @AureRita: Because the stack trace relates to a module it is interesting to see if the error can be reproduced on the dev branch as well.
| gharchive/pull-request | 2024-05-09T14:38:07 | 2025-04-01T04:32:58.973322 | {
"authors": [
"AureRita",
"Quetzacoalt91"
],
"repo": "PrestaShop/autoupgrade",
"url": "https://github.com/PrestaShop/autoupgrade/pull/698",
"license": "AFL-3.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1749438023 | Deploy kanbanbot v1
The first kanbanbot feature are done. We have to deploy the main branch using the automatic deploiement based en github flow and docker.
Kanban bot deployed
| gharchive/issue | 2023-06-09T08:46:47 | 2025-04-01T04:32:58.978042 | {
"authors": [
"WahbiPS",
"lartist"
],
"repo": "PrestaShop/kanbanbot",
"url": "https://github.com/PrestaShop/kanbanbot/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
691413822 | show staff affiliates on affiliates page #192
see checklist on #192 for a list of changes on this PR.
Codecov Report
Merging #196 into develop will increase coverage by 0.00%.
The diff coverage is 100.00%.
@@ Coverage Diff @@
## develop #196 +/- ##
========================================
Coverage 98.08% 98.09%
========================================
Files 82 82
Lines 2767 2778 +11
========================================
+ Hits 2714 2725 +11
Misses 53 53
| gharchive/pull-request | 2020-09-02T20:54:29 | 2025-04-01T04:32:58.992543 | {
"authors": [
"codecov-commenter",
"thatbudakguy"
],
"repo": "Princeton-CDH/cdh-web",
"url": "https://github.com/Princeton-CDH/cdh-web/pull/196",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1398308609 | Slamp checks
Merge checking passes into master
The main functionality is not yet ready. Please checkpick the repl related updates and only PR those.
| gharchive/pull-request | 2022-10-05T20:00:26 | 2025-04-01T04:32:59.001781 | {
"authors": [
"vgene",
"yebinchon"
],
"repo": "PrincetonUniversity/cpf",
"url": "https://github.com/PrincetonUniversity/cpf/pull/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
374646700 | NEON configuration language support
NEON is a configuration language similar to YAML, with the notable addition of so-called "NEON entities". NEON is used mainly by the Nette PHP framework.
See #2140.
| gharchive/pull-request | 2018-10-27T13:45:05 | 2025-04-01T04:32:59.002756 | {
"authors": [
"RunDevelopment",
"jahudka"
],
"repo": "PrismJS/prism",
"url": "https://github.com/PrismJS/prism/pull/1600",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2090922710 | Add background to main-page and improve its CSS
I visited the website and think the css of front page could be improved and background could be added.
Please assign it to me under JWOC 24
@Priyanshi662
please assign me under JWOC 24.
@Priyanshi662
thank you so much.
@Priyanshi662
| gharchive/issue | 2024-01-19T16:52:14 | 2025-04-01T04:32:59.021620 | {
"authors": [
"KhushiAgarwal22"
],
"repo": "Priyanshi662/FunFusion",
"url": "https://github.com/Priyanshi662/FunFusion/issues/57",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1997810628 | Input value goes back to initial values after reopening the program
Code: - I_x =? {48} in^4
This is the original value the I_x was set, when the program is executed and new values are replaced the calculations are done for the updated values, however when I close and reopen the program the I_x value goes back to 48 instead of last value. It would be useful if it's replaced by the last user input value.
Hi!
Did you try to save the file with the new values? You need to save it to persist the new values for the next time.
Yeah! it works, thank you :)
| gharchive/issue | 2023-11-16T21:23:56 | 2025-04-01T04:32:59.051602 | {
"authors": [
"Proektsoftbg",
"strsthapa"
],
"repo": "Proektsoftbg/Calcpad",
"url": "https://github.com/Proektsoftbg/Calcpad/issues/243",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1582996934 | Update Wiki for Stripe
Need to update Wiki for Stripe
Take out Apple Pay for Checkout
https://github.com/ProgramEquity/amplify/blob/main/server/routes/api/checkout.js
From @masmei
https://stripe.com/docs/api/payment_intents/object#payment_intent_object-setup_future_usage
^^ this is on saving customer data
We can store data from payments object in our DB
@manishapriya94 could you please elaborate what we have to do for this issue?
From what I gather, we just have to remove references to Apple Pay from the wiki page and rename it?
| gharchive/issue | 2023-02-13T20:13:22 | 2025-04-01T04:32:59.055739 | {
"authors": [
"iofall",
"manishapriya94"
],
"repo": "ProgramEquity/amplify",
"url": "https://github.com/ProgramEquity/amplify/issues/492",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1411832362 | CVE-2015-5237
Grype scan reported CVE-2015-5237.
Please investigate and update .grype.yaml.
A false-positive scan result from grype.
mc team confirmed that mc is using the latest version of protobuf which is versioned v1.28.1. This new versioning number made grype think that an older/vulnerable protobuf is used in mc.
Closing as not an issue and updating grype ignore list.
| gharchive/issue | 2022-10-17T16:00:32 | 2025-04-01T04:32:59.074407 | {
"authors": [
"mocsharp"
],
"repo": "Project-MONAI/monai-deploy-workflow-manager",
"url": "https://github.com/Project-MONAI/monai-deploy-workflow-manager/issues/515",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2697742374 | Return information about specified user in report Portal: api spec change
API deprecation: https://reportportal.io/docs/api/service-api/versions/5.11/get-user-using-get/
API new: https://reportportal.io/docs/api/service-api/get-user
Connected issue:
https://github.com/ProjectAlita/projectalita.github.io/issues/512
| gharchive/pull-request | 2024-11-27T09:02:36 | 2025-04-01T04:32:59.090796 | {
"authors": [
"plageat"
],
"repo": "ProjectAlita/application-tools",
"url": "https://github.com/ProjectAlita/application-tools/pull/84",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1521449795 | Dan/2023/01/remove unused functions
Addresses the enumerated unused functions in https://github.com/spearbit-audits/review-seaport-1.2/issues/97.
Note that @Saw-mon-and-Natalie left a todo two days ago to add others.
Update about additional enumerations in the above Spearbit Issue that haven't yet been added to this PR:
LowLevelHelpers.sol#L25
LowLevelHelpers.sol#L112
ZoneInteraction.sol#L227
Update about additional enumerations in the above Spearbit Issue that haven't yet been added to this PR:
LowLevelHelpers.sol#L25
LowLevelHelpers.sol#L112
ZoneInteraction.sol#L227
Those were all removed in other PRs, I believe.
True. These were in the changes happening during the early stages of the audit: https://github.com/ProjectOpenSea/seaport/commit/4de24dfd64b56a966d348405cedf3c4ac7e4450e
Closing as we decided to keep all the helpers in PointerLibraries even if they're not all currently being used.
| gharchive/pull-request | 2023-01-05T21:27:22 | 2025-04-01T04:32:59.116222 | {
"authors": [
"0age",
"DJViau",
"JustDravee"
],
"repo": "ProjectOpenSea/seaport",
"url": "https://github.com/ProjectOpenSea/seaport/pull/838",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1903687006 | Fix typo in README.md
right at the end Godot as written as "GoDot", which is incorrect
There's also one in the requirements.txt file at "tokenzie"(should be "tokenize") and "confiugration"
https://github.com/ProjectUnifree/unifree/blob/8a5456fb10b24954c3e4b308154cc8ef985a728a/configs/godot.yaml#L3 also calls it "GoDot".
| gharchive/pull-request | 2023-09-19T20:34:19 | 2025-04-01T04:32:59.136406 | {
"authors": [
"NotHyper-474",
"aragubas",
"gsteinLTU"
],
"repo": "ProjectUnifree/unifree",
"url": "https://github.com/ProjectUnifree/unifree/pull/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1952194600 | Make Watchy_GSR compatible?
First of all, I love your watchface!
I just wish it was compatible with Watchy_GSR.
Do you know about Watchy_GSR? (https://github.com/GuruSR/Watchy_GSR)
I've seen many people in the watchy discord interested in your watchface, and a few that were interested in it being converted to Watchy GSR compatibility. If I could code myself, I would try to convert it, but I figured you would be the one that knows your code best! Just thought maybe you didn't know about Watchy_GSR and thought I would let you know there is some interest in it's conversion.
Feel free to close this "issue" that isn't really an issue, just wanted to communicate this idea to you.
This would be awesome
| gharchive/issue | 2023-10-19T12:55:35 | 2025-04-01T04:32:59.138608 | {
"authors": [
"Izaic",
"grnqrtr"
],
"repo": "Prokuon/watchy-starfield",
"url": "https://github.com/Prokuon/watchy-starfield/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
438877827 | Scrolling down the mod list shows invisible items
The list seems to be empty after 2nd page but the items are still clickable. Have about 50 mods installed now. After searching with pattern scrolling not works anymore.
it seems to only show about 20 mods for me, how about you Knito?
@Hellzbellz123 Same for me. Thanks for seconding me. I really thought that only happens to me.
yeah it seems to be a bug, if you happen to remember mod names you can filter the list to show ones past the 20 limiut
Exactly.
| gharchive/issue | 2019-04-30T16:54:21 | 2025-04-01T04:32:59.149232 | {
"authors": [
"Hellzbellz123",
"Knito58"
],
"repo": "Prospector/ModMenu",
"url": "https://github.com/Prospector/ModMenu/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
443660254 | RawFileReader update?
I’m not sure which version of Thermo’s RawFileReader is currently in use by ProteoWizard (I assume 4.0.26), but they do have 2 newer versions (4.0.89 and 5.0.6) available on their sharepoint site, with release notes mentioning unspecified “bug fixes in the libraries” and a change to a single set of binaries (compared to the 3 previous sets for each of Windows, MacOS, and Linux).
Update is merged, I don't see any need for this to remain open.
| gharchive/issue | 2019-05-14T00:42:39 | 2025-04-01T04:32:59.157808 | {
"authors": [
"FarmGeek4Life"
],
"repo": "ProteoWizard/pwiz",
"url": "https://github.com/ProteoWizard/pwiz/issues/541",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
238397463 | Don't allow a user to use localhost as a connection address
Tons of issues with people using localhost and then being unable to connect to the SQL socket, require using an IP which avoids that issue.
this is only force localhost to 127.0.0.1 in assing config or block this parameter? (question for.. know?)
Technically if it is set by a plugin I don't think there is much you can do there, though making a huge warning stating you must use the external ip instead of localhost on the docs should make a difference.
This is for panel DB connections, not plugins.
For now I've put a warning on the environment setup for database connections, and the command will also test connections, so I think I'm going to close this as 8722571 addresses it for now.
| gharchive/issue | 2017-06-25T18:34:23 | 2025-04-01T04:32:59.354539 | {
"authors": [
"DaneEveritt",
"Doc94",
"Polarcraft"
],
"repo": "Pterodactyl/Panel",
"url": "https://github.com/Pterodactyl/Panel/issues/518",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2322752816 | Accessing quarto metadata programmatically via QuartoNotebookWorker
Now that we have the expand extension pipeline, I was thinking that downstream packages might want to expand differently depending on, for example, the output format that quarto renders to. I think that's different from the loading hooks we currently have. So I wonder how we could make the full quarto metadata accessible, like you can access it from knitr with knitr::pandoc_to(), as one example.
Stored in https://github.com/PumasAI/QuartoNotebookRunner.jl/blob/85f65102f63c2c4eac98c5cb51eed14fc2948535/src/QuartoNotebookWorker/src/NotebookState.jl#L8, so we can make an official API for that I guess.
Yes, although there's currently no open way to access anything QNW-related, is there? The extension methods hook in at package load time, you couldn't easily reach that from the notebook itself. And I think we were holding off on adding any API into the notebook workspace due to it not being available in a REPL.
you couldn't easily reach that from the notebook itself. And I think we were holding off on adding any API into the notebook workspace due to it not being available in a REPL.
Would only be accessed in 3rd-parth package extensions, not directly in a notebook, e.g.
module PackageNameQuartoNotebookWorkerExt
import PackageName
import QuartoNotebookWorker
function QuartoNotebookWorker.expand(::PackageName.Foo)
opts = QuartoNotebookWorker.options() # hypothetical public API, used at runtime.
# ...
end
end
Right, ok, so it should stay that way then. One could not develop such functionality in scripts is the only drawback I can see.
One could not develop such functionality directly within notebooks is the only drawback I can see.
Nope, you can't. Revise ideally should just work with this fine and functionality can just be developed directly in the extension file. If that doesn't work then we should make it work.
Yes for "serious" development that's true, it just adds overhead if you want to quickly try out things. I would have liked an "escape hatch" of some kind. Actually, maybe that's already possible by requiring the right UUID manually. Just thinking out loud here, I think the extension is in general the right place to do this kind of stuff.
Actually, maybe that's already possible by requiring the right UUID manually.
Given the way the LOAD_PATH is done you should just be able to always do using QuartoNotebookWorker in a cell as an escape hatch if needed. (Worked for me just now).
| gharchive/issue | 2024-05-29T08:53:26 | 2025-04-01T04:32:59.379974 | {
"authors": [
"MichaelHatherly",
"jkrumbiegel"
],
"repo": "PumasAI/QuartoNotebookRunner.jl",
"url": "https://github.com/PumasAI/QuartoNotebookRunner.jl/issues/139",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
430995219 | No attempts for a long time
I'm having this issue inly for exact user, cause I see that for some other usernames this bruter works well
I was facing the same issue, then I found out that my other server is bruting the same account. so just terminated other one (guess multiple login attempts cause errors)
They use the same proxies, so it’s not a good idea to use multiple servers to attack your account
| gharchive/issue | 2019-04-09T14:20:08 | 2025-04-01T04:32:59.387093 | {
"authors": [
"Pure-L0G1C",
"ali3shah",
"sshkrv"
],
"repo": "Pure-L0G1C/Instagram",
"url": "https://github.com/Pure-L0G1C/Instagram/issues/88",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1636843715 | update redirects for opengov changes
Add redirects to the CN mkdocs file for the opengov related changes
Based on these PRS:
https://github.com/PureStake/moonbeam-mkdocs/pull/105/files
https://github.com/PureStake/moonbeam-mkdocs/pull/106/files
Approved both OpenGov related PRs
| gharchive/pull-request | 2023-03-23T04:47:30 | 2025-04-01T04:32:59.388820 | {
"authors": [
"albertov19",
"eshaben"
],
"repo": "PureStake/moonbeam-mkdocs",
"url": "https://github.com/PureStake/moonbeam-mkdocs/pull/112",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1007171196 | Android build error when including in pubspec.yaml
This plugin causes a build error on Android when simply being included in pubspec.yaml, without any other changes being made to my project:
Execution failed for task ':app:checkDebugDuplicateClasses'.
> A failure occurred while executing com.android.build.gradle.internal.tasks.CheckDuplicatesRunnable
> Duplicate class org.threeten.bp.Clock found in modules jetified-threetenbp-1.4-no-tzdb.2-no-tzdb (org.threeten:threetenbp:1.4.2) and jetified-threetenbp-1.4.2 (org.threeten:threetenbp:1.4.2)
> ... (hundreds of similar errors)
My minSdkVersion is 24, above the 21 required in the README file.
Hi,
it's a problem with your gradle config, try with the suggestions here or here or here and let me know if it works.
Simone
| gharchive/issue | 2021-09-25T19:03:04 | 2025-04-01T04:32:59.397158 | {
"authors": [
"palkerecsenyi",
"sstasi95"
],
"repo": "PurpleSoftSrl/sumup_flutter_plugin",
"url": "https://github.com/PurpleSoftSrl/sumup_flutter_plugin/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1425432426 | 🛑 Gitea is down
In 27c1e0e, Gitea (https://git.pxlbuzzard.com) was down:
HTTP code: 503
Response time: 103 ms
Resolved: Gitea is back up in 620150f.
| gharchive/issue | 2022-10-27T10:50:14 | 2025-04-01T04:32:59.411755 | {
"authors": [
"PxlBuzzard"
],
"repo": "PxlBuzzard/upptime",
"url": "https://github.com/PxlBuzzard/upptime/issues/685",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
707561817 | Syntax Issue
File "filemovergui.py", line 45
".wmv",".mov",".qt",".rm",]"CodePerfectPlus",
^
SyntaxError: invalid syntax
When I run python3 filemovergui.py
Thanks, @Raavtube
For pointing out Bug. It's now fixed.
| gharchive/issue | 2020-09-23T17:50:07 | 2025-04-01T04:32:59.413465 | {
"authors": [
"Raavtube",
"codePerfectPlus"
],
"repo": "Py-Contributors/FileMoverGUI",
"url": "https://github.com/Py-Contributors/FileMoverGUI/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
712570224 | Find the Missing Number
Before jumping into Pr first comment for assign.
Problem statement : Find the Missing Number
Probelm Link : https://edabit.com/challenge/7YaJhC4terApw5DFa
Choose one Cateorgy - Very Easy, Easy, Medium, Hard, Very Hard, Expert
question category : Easy
Before submitting a PR please Check some details.
[ ] fileName formet should be in camelcase
example- returnTheNextNumber.py
[ ] Choose label for Problem , example- "good first issue", "easy"
[ ] In file there should be problem statement and link to problem
/* Problem-Task : Create a function that takes an array of numbers between 1 and 10 (excluding one number) and returns the missing number.
* Problem Link : https://edabit.com/challenge/7YaJhC4terApw5DFa
*/
int main() {
int bla;
return 0;
}
This repository is vaild for HacktoberFest2020
Read before making PR - How to contribute on Github
Good Luck,
Can you assign me this issue
Can you assign this to me?
@pranav230 can u assign this to me
Please assign me this.
@pranav230 can you assign this to me ....
i am working on this issues ,I will create PR soon
can you please assign this to me
@pranav230 I Added FindMissingNumber.cpp ..
please check my code and review my PR
PR of assign Person will review.. sorry
| gharchive/issue | 2020-10-01T07:22:40 | 2025-04-01T04:32:59.419362 | {
"authors": [
"NidhiNivedita6",
"VaderX",
"aayushi-droid",
"pranav230",
"yasharth291",
"zahinekbal"
],
"repo": "Py-Droid/Edabit-Solutions",
"url": "https://github.com/Py-Droid/Edabit-Solutions/issues/112",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1347583624 | Make options changeable after codec or (de)muxer has been opened
Overview
Currently, as best I can tell, it is only possible to set options once: when a codec or (de)muxer is opened. There are cases where one might want to change some option afterward. My use case is changing -crf in the libx264 encoder during encoding based on user preference and changing network conditions.
Existing FFmpeg API
The av_opt_set*() in libavutil/opt.h. av_opt_get*() are probably also prudent to implement.
Doxy link: https://ffmpeg.org/doxygen/trunk/group__avoptions.html
Expected PyAV API
The av_opt functions are very generic, so the easiest route would probably be to guarantee that all relevant classes have self.ptr exposed that is guaranteed to be passable to the av_opt functions. This could be hidden from the user so that they only have to pass a CodecContext, OutputContainer or InputContainer.
Example based on aiortc's h264.py:
from av.avutil import av_opt_set_int, av_opt_get_int
class H264Encoder(Encoder):
# [...]
def set_crf(self, crf: int)
av_opt_set_int(self.codec, "crf", crf) # search_flags could be optional, default to 0
def get_crf(self) -> int:
return av_opt_get_int(self.codec, "crf")
Errors returned by av_opt_get_int() could be raised as Exceptions, which simplifies the get functions like above.
It would also be useful to expose AV_OPT_SEARCH_* and optionally allow setting them on the set and get calls.
Here's a little proof-of-concept that works well enough for my purposes. In context.pyx:
cdef class CodecContext(object):
# Default search_flags to AV_OPT_SEARCH_CHILDREN
def opt_set_int(self, name: str, value: int, search_flags: int = 1):
ret = lib.av_opt_set_int(self.ptr, name, value, search_flags)
if ret != 0:
raise RuntimeError(f'av_opt_set_int() = {ret}')
def opt_get_int(self, name: str, search_flags: int = 1):
cdef int64_t val
ret = lib.av_opt_get_int(self.ptr, name, search_flags, &val)
if ret != 0:
raise RuntimeError(f'av_opt_get_int() = {ret}')
return val
And of course in avutil.pxd right after av_opt_set_int():
cdef int av_opt_get_int(
void *obj,
char *name,
int search_flags,
int64_t *out_val
)
It'd be better if all classes inherited some base class from which void* ptr can be gotten, perhaps via an abstract property or something.
| gharchive/issue | 2022-08-23T08:55:20 | 2025-04-01T04:32:59.424271 | {
"authors": [
"Tjoppen"
],
"repo": "PyAV-Org/PyAV",
"url": "https://github.com/PyAV-Org/PyAV/issues/1021",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2268572841 | feat: support pyclass on tuple enums (closer but not there yet)
This is an extension of this PR, which hasn't been updated in two weeks:
https://github.com/PyO3/pyo3/pull/4072
Aiming to close this issue:
https://github.com/PyO3/pyo3/issues/3748
@newcomertv in that PR added support for tuple enums but only with ._0 ._1 access to elements.
This PR attempts to do the following:
[x] respond to minor comments from #4072
[x] support [0] [1] access on tuple elements with __getitem__()
[ ] with single-element tuples, automatically forward to that element
I've got the Rust-side macros working but haven't yet managed to get them wired up on the Python side... Main complication is that the __getitem__() implementation has to return a Box<dyn Any> (as the type won't be known) and I haven't figured out how to deal with that on the other side.
Will give it another go soon but putting this up in case anyone has suggestions...
Thanks for continuing the work here! These double underscore methods are a bit special in how their code generation works.
I think we can reuse
// `slot` is `__GETITEM__`/`__LEN__` from `pyo3-macros-backend/src/pymethods`
slot.generate_type_slot(variant_cls_type, &spec, &name, ctx)
to generate these special definitions and thread them back up into PyClassImplsBuilder::new slots argument. Then in impl_complex_enum_tuple_variant_cls we can just generate the corresponding functions
fn __getitem__(slf: PyRef<Self>, idx: usize) -> PyResult<PyObject> {...}
fn __len__(slf: PyRef<Self>) -> usize {...}
There should be no need to use Box<dyn Any>, because they need to be convertible to a PyObject anyway. Maybe this helps you a bit while continuing.
Just for future reference: We might also want to generate __match_args__ to allow tuple pattern matching on the Python side.
Closingin favour of #4072 which has picked this up again!
Thanks @Icxolu
| gharchive/pull-request | 2024-04-29T09:41:56 | 2025-04-01T04:32:59.489305 | {
"authors": [
"Icxolu",
"carderne"
],
"repo": "PyO3/pyo3",
"url": "https://github.com/PyO3/pyo3/pull/4135",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
642307411 | Do not use PyObject_RichCompareBool to detect invalid comparison
Fixes #985
I think this is the right thing to do in the general case, but I suspect that there's a reasonable speed-up associated with using PyObject_RichCompareBool, since most objects actually are equal when the object identity is the same.
I have not looked at this issue in any detail, and I haven't tried to do any benchmarks to see if this change even really had a major speed impact, but I'm wondering how much work it would be to either special case certain types where it is known that pointer comparison is valid, like int, str, list, tuple, datetime.datetime, etc, or even something more elaborate using traits (or an extra field in PyAny indicating whether pointer comparison is "safe" for a given ).
I would be curious to see how this change affected the speed of comparison of, say, two identical tuples or lists, since for container types this optimization is already made in __eq__:
>>> a = [float('nan')]
>>> b = [float('nan')]
>>> a == b
False
>>> a == a
True
>>> b == b
True
Another possible performance regression would be two lists that are not the same object but consistent of a large number of elements all of which are the same object, like so:
>>> OBJECT = "bbbbb"
>>> a = [OBJECT] * 10000
>>> b = [OBJECT] * 10000
>>> a == b
| gharchive/pull-request | 2020-06-20T04:37:01 | 2025-04-01T04:32:59.492884 | {
"authors": [
"kngwyu",
"pganssle"
],
"repo": "PyO3/pyo3",
"url": "https://github.com/PyO3/pyo3/pull/986",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
927549489 | Add note to ExclusionContainer.add_raster(...) that anonymous (lambda) functions are not permitted.
Change proposed in this Pull Request
Doc-string change for ExclusionContainer.add_raster(...) and changes to tests:
Anonymous functions are not permitted for the codes=... keyword argument.
Description
Non-anonymous functions can be used. As an example I've changed the related test (which was for the meantime inactive) from a lambda function to an equivalent formulation using functools.partial and np.greater_equal.
Motivation and Context
Anonymous functions lead to complications with python's multiprocessing, c.f. https://github.com/PyPSA/pypsa-eur/issues/249 .
How Has This Been Tested?
n/a
Type of change
[x] Bug fix (non-breaking change which fixes an issue)
[n/a ] New feature (non-breaking change which adds functionality)
[n/a] Breaking change (fix or feature that would cause existing functionality to change)
Checklist
[x] I tested my contribution locally and it seems to work fine.
[ ] I locally ran pytest inside the repository and no unexpected problems came up.
[x] I have adjusted the docstrings in the code appropriately.
[n/a] I have documented the effects of my code changes in the documentation doc/.
[n/a] I have added newly introduced dependencies to environment.yaml file.
[x] I have added a note to release notes doc/release_notes.rst.
[ ] I have used pre-commit run --all to lint/format/check my contribution
great!
| gharchive/pull-request | 2021-06-22T19:06:54 | 2025-04-01T04:32:59.498506 | {
"authors": [
"FabianHofmann",
"euronion"
],
"repo": "PyPSA/atlite",
"url": "https://github.com/PyPSA/atlite/pull/171",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
755745092 | intermediate_source/rpc_param_server_tutorial.rst 번역
문서 URL
번역을 하실 / 요청하실 URL을 남겨주세요. (예. https://tutorials.pytorch.kr/beginner/saving_loading_models.html)
URL: https://tutorials.pytorch.kr/intermediate/rpc_param_server_tutorial.html
(대략적인) 예상 완료 일정
예상하시는 완료 일정이 있으시다면 알려주세요. (예. 1달 내, 12월 중 등)
(반드시 지키셔야 하는 일정이 아닙니다 - 일정이 너무 늦어질 경우 다른 번역자를 위해 남겨주세요.)
2주일 내
관련 이슈
현재 번역 요청 / 진행 내역을 보기 위해 각 버전의 메인 이슈를 참조합니다.
(특별한 일이 없다면 수정하지 않으셔도 됩니다.)
#90 (v1.6)
진행되지 않은지 오래되어 이슈를 닫습니다.
제가 잘못 닫거나, 진행 중이신 경우 이슈를 다시 열어주시고 덧글로 알려주세요.
| gharchive/issue | 2020-12-03T01:49:15 | 2025-04-01T04:32:59.557121 | {
"authors": [
"9bow",
"hongsam123"
],
"repo": "PyTorchKorea/tutorials-kr",
"url": "https://github.com/PyTorchKorea/tutorials-kr/issues/195",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
921249232 | Added nn.Module support for FasterRCNN backbone
What does this PR do?
Fixes #660
Before submitting
[x] Was this discussed/approved via a Github issue? (no need for typos and docs improvements)
[x] Did you read the contributor guideline, Pull Request section?
[x] Did you make sure your PR does only one thing, instead of bundling different changes together?
[x] Did you make sure to update the documentation with your changes?
[x] Did you write any new necessary tests? [not needed for typos/docs]
[ ] Did you verify new and existing tests pass locally with your changes?
[ ] If you made a notable change (that affects users), did you update the CHANGELOG?
PR review
[x] Is this pull request ready for review? (if not, please submit in draft mode)
Anyone in the community is free to review the PR once the tests have passed.
If we didn't discuss your PR in Github issues there's a high chance it will not be merged.
Did you have fun?
Sure! 🙃
@Borda, I made the changes. Let me know what you think!
PS: Tests had failed due to the issue below, which I see has been fixed now...
This change is causing tests to fail in master. Please check!
3bf65da#diff-81b21ad681f00e8e80279aaf4d5e2c3a304b5a587b79850cff8d4472147f7f0aL92
Hey @abhayraw1 seems there is one failing test left, do you mind checking it out! thanks for your work so far!!
Not sure what is wrong here:
boxes = target["boxes"]
degenerate_boxes = boxes[:, 2:] <= boxes[:, :2]
if degenerate_boxes.any():
# print the first degenerate box
bb_idx = torch.where(degenerate_boxes.any(dim=1))[0][0]
degen_bb: List[float] = boxes[bb_idx].tolist()
> raise ValueError("All bounding boxes should have positive height and width."
" Found invalid box {} for target at index {}."
.format(degen_bb, target_idx))
E ValueError: All bounding boxes should have positive height and width. Found invalid box [137.5, 556.25, 743.75, 556.25] for target at index 0.
as I see the [137.5, 556.25] <= [743.75, 556.25] shall be true, but it fails...
| gharchive/pull-request | 2021-06-15T10:29:01 | 2025-04-01T04:32:59.563859 | {
"authors": [
"Borda",
"SeanNaren",
"abhayraw1"
],
"repo": "PyTorchLightning/lightning-bolts",
"url": "https://github.com/PyTorchLightning/lightning-bolts/pull/661",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1054080139 | [RFC] Introduce Strategy in favor of TrainingTypePlugin
Proposed refactor
Part of #9932
Related PR: #10548
Motivation
As we are introducing StrategyPlugin in favor of TrainingTypePlugin for this release.
@ananthsub raised some really good points for discussion regarding StrategyPlugin. Reference
Proposed solution
Three main points to be discussed here
Drop Plugin from the name of the class
It should be under pytorch_lightning/strategy directory
Over time, these classes should be renamed as DDPStrategy and so on.
Would like your thoughts? @PyTorchLightning/core-contributors
cc @borda @justusschock @awaelchli @akihironitta @kaushikb11 @ananthsub
Yes, definitely. We agreed on this I think. This was already set in motion by the introduction of the strategy argument in the Trainer.
Over time, these classes should be renamed as DDPStrategy and so on.
Has to be done for 1.6. We are targeting stable API for these components at 1.6.
Should we name all classes without the "Plugin"/"Strategy" suffix? Here is how this would look like:
DataParallel, (or just DP)
DDP2,
DDP,
DDPSpawn,
DDPFullySharded,
DeepSpeed,
Horovod,
IPU,
SingleDevice,
SingleTPU,
TPUSpawn,
TrainingType,
Parallel,
DDPSharded,
DDPSpawnSharded,
On naming, Loggers, Profilers, Accelerators, and ClusterEnvironments are suffixed with the component name (e.g. TensorboardLogger, SimpleProfiler, GPUAccelerator, SlurmEnvironment).
If we drop the suffix for strategy:
Would we also make this change for other components of Lightning so the naming scheme is unified across components (leading to one less thing to think about for users) ?
Do you think there's a risk with naming collisions? For instance, the IPU accelerator and IPU strategy could both be IPU classes. As a result, users would likely need to change the import for them to be:
from pytorch_lightning.accelerators import IPU as IPUAccelerator
from pytorch_lightning.strategy import IPU as IPUStrategy
Proposed plan:
Rename classes XPlugin to XStrategy
Move files from pl/plugins/training_type to pl/strategy or pl/strategies
Reroute pl/plugins/training_type to pl/strategy with a deprecation message
Step 3) is necessary because of certain usages of these plugins, even when they are not being subclassed. One example is:
trainer = Trainer(plugins=DDPPlugin(find_unused_parameters=False)) in which case we don't want to break user's code.
More follow-ups will include:
Rename TTP registry
Rename tests, comments, docstrings that use the TTP terms
@awaelchli the step 1 will be breaking change for trainer = Trainer(plugins=DDPPlugin(find_unused_parameters=False)) use case already, as the DDPPlugin won't exist anymore?
To keep backward compatibility we will have to copy the folder pl/plugins/training_type to pl/strategy, and only rename Xplugin to Xstrategy in the new folder pl/strategy and kept pl/plugins/training_type unchanged? In the mean time, reroute pl/plugins/training_type to pl/strategy with a deprecation message.
Do we want to do this way?
We also have to remember to rename the test files and test names that include the word plugin.
We also have to remember to rename the test files and test names that include the word plugin.
I could take care of it.
Closing this issue, as the Strategy transition has been completed! Thank you to everyone involved ❤️
| gharchive/issue | 2021-11-15T20:40:02 | 2025-04-01T04:32:59.575210 | {
"authors": [
"ananthsub",
"awaelchli",
"four4fish",
"kaushikb11"
],
"repo": "PyTorchLightning/pytorch-lightning",
"url": "https://github.com/PyTorchLightning/pytorch-lightning/issues/10549",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
709789011 | ref: deprecated results obj, added support for simpler comms (1/n)
Here we move to deprecate results objects in favor of simpler syntax. Results obj support will remain until 0.10.0 but not 1.0.0.
(ie: 0.10.0 is 1.0.0 but backwards compatible for anyone who really needs it).
The new pattern is to decouple logging from hooks. That means that step, step_end, epoch_end are independent from logging.
example
def training_step(...):
loss = ...
return loss
# equivalent
def training_step(...):
return {'loss': loss}
# to log
def training_step(...):
self.log('anything', x, on_step=True, on_epoch=True)
Passing around step results.
In the case the user still needs/wants to do something with the output of each batch, the other hooks are still there.
def training_step(...):
return {'loss': loss, 'random_thing': [1, 'a', Tensor(), ...]}
def training_epoch_end(self, training_step_outputs):
for d in training_step_outputs:
random_thing = d['random_thing']
Hooks and .log are decoupled...
user can log from anywhere
def training_epoch_end(...):
some_new_val = ...
self.log('my_new_val', some_new_val)
cc @ananthsub @awaelchli @justusschock
Hello @williamFalcon! Thanks for opening this PR.
In the file pytorch_lightning/trainer/logging.py:
Line 70:1: W293 blank line contains whitespace
Do see the Hitchhiker's guide to code style
cc @ananthsub @awaelchli @justusschock
Hi, after updating to version 1.0.4, i think below approach seems to be not working as desired
def training_step(...):
return {'loss': loss, 'random_thing': [1, 'a', Tensor(), ...]}
def training_epoch_end(self, training_step_outputs):
for d in training_step_outputs:
random_thing = d['random_thing']
training_step_outputs is always empty. On further debugging i find that training_step is never getting called, instead this function training_step_and_backward is called with run_training_batch
ok, will fix asap. thanks for reporting!
mind replicating it on a boring model?
@edenafek (p0)
@nazim1021! Thank you for reporting! I created an issue to track this, would be great if you can try to repro using boring model and paste a colab link in the new issue](https://github.com/PyTorchLightning/pytorch-lightning/issues/4452)
@williamFalcon @edenlightning sure, of course. Thanks for prompt reply
i ll add the colab link to new issue
| gharchive/pull-request | 2020-09-27T18:18:23 | 2025-04-01T04:32:59.584537 | {
"authors": [
"edenlightning",
"nazim1021",
"pep8speaks",
"williamFalcon"
],
"repo": "PyTorchLightning/pytorch-lightning",
"url": "https://github.com/PyTorchLightning/pytorch-lightning/pull/3681",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
711657802 | Support OmegaConf hparams on LightningModule
What does this PR do?
Fixes bug in this stacktrace: https://gist.github.com/ananthsub/07f99b2acacb7ef46bf426f0323cc615
The LightningModule in question does this in the init:
class MyModule(LightningModule):
def __init__(self, cfg=None, visualize_cfg=None, benchmark_cfg=None, **kwargs):
super().__init__()
self._cfg = cfg
self._visualize_cfg = visualize_cfg
self._benchmark_cfg = benchmark_cfg
self.save_hyperparameters()
...
where all the cfgs are OmegaConf containers. It doesn't look like the framepointer logic for save_hyperparameters handles this well. The LightningModule hparams convert OmegaConf hparams to AttributeDict from this conversion: https://github.com/PyTorchLightning/pytorch-lightning/blob/master/pytorch_lightning/core/lightning.py#L1640-L1641
This causes checks in the tensorboard logger to never trigger, because the type(hparams) is type(params) is <class 'pytorch_lightning.utilities.parsing.AttributeDict'>. This results in YAML dumping to file issues, as seen in the stacktrace
Before submitting
[ ] Was this discussed/approved via a Github issue? (no need for typos and docs improvements)
[x] Did you read the contributor guideline, Pull Request section?
[x] Did you make sure your PR does only one thing, instead of bundling different changes together? Otherwise, we ask you to create a separate PR for every change.
[ ] Did you make sure to update the documentation with your changes?
[ ] Did you write any new necessary tests?
[ ] Did you verify new and existing tests pass locally with your changes?
[ ] If you made a notable change (that affects users), did you update the CHANGELOG?
PR review
Anyone in the community is free to review the PR once the tests have passed.
If we didn't discuss your PR in Github issues there's a high chance it will not be merged.
Did you have fun?
Make sure you had fun coding 🙃
Abandoning this PR. The workaround was to
Merge the OmegaConfs into a single object
set self.hparams = <merged omegaconf>
Drop the call to save_hyperparameters()
Ideally hparams could be set in the init instead
| gharchive/pull-request | 2020-09-30T06:02:36 | 2025-04-01T04:32:59.591388 | {
"authors": [
"ananthsub"
],
"repo": "PyTorchLightning/pytorch-lightning",
"url": "https://github.com/PyTorchLightning/pytorch-lightning/pull/3736",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
850171484 | [WIP] Move NaN/Inf detection to a separate utilities file
What does this PR do?
Fixes #6815
Before submitting
[ ] Was this discussed/approved via a GitHub issue? (not for typos and docs)
[ ] Did you read the contributor guideline, Pull Request section?
[ ] Did you make sure your PR does only one thing, instead of bundling different changes together?
[ ] Did you make sure to update the documentation with your changes? (if necessary)
[ ] Did you write any new necessary tests? (not for typos and docs)
[ ] Did you verify new and existing tests pass locally with your changes?
[ ] Did you update the CHANGELOG? (not for typos, docs, test updates, or internal minor changes/refactorings)
PR review
Anyone in the community is free to review the PR once the tests have passed.
Before you start reviewing make sure you have read Review guidelines. In short, see the following bullet-list:
[ ] Is this pull request ready for review? (if not, please submit in draft mode)
[ ] Check that all items from Before submitting are resolved
[ ] Make sure the title is self-explanatory and the description concisely explains the PR
[ ] Add labels and milestones (and optionally projects) to the PR so it can be classified
Did you have fun?
Make sure you had fun coding 🙃
Hello @ananthsub! Thanks for opening this PR.
In the file pytorch_lightning/trainer/training_tricks.py:
Line 37:121: E501 line too long (121 > 120 characters)
Do see the Hitchhiker's guide to code style
| gharchive/pull-request | 2021-04-05T07:54:04 | 2025-04-01T04:32:59.599505 | {
"authors": [
"ananthsub",
"pep8speaks"
],
"repo": "PyTorchLightning/pytorch-lightning",
"url": "https://github.com/PyTorchLightning/pytorch-lightning/pull/6834",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
926246836 | SWT2 with db4 error
Hi! Not an issue in itself I guess (might be), but I am posting here because so far I cannot find a solution to this, even though I searched StackOverflow and the PYWT forum. I am trying to apply swt2 with a db4 decomposition to an (516, 862) 2D array. The data has been padded to meet the 2**level criteria and the max level has been calculated to be 3 using swt_max_level. My need is to extract the coefficients at each level (up to max_level), however I when the level > 1 I keep receiving the start_level must be less than 1 error, even though I specifically declare it to be 0. Am I missing something?
has been padded to meet the 2**level criteria
This should work. If I want to do a 3-level decomposition of data that is shape (516, 862) then it needs to be padded to a multiple of 2**3 on each axis. So, the padded shape should be (516, 864) in this case.
The following works for me.
import numpy as np
import pywt
x = np.random.randn(516, 862)
x_pad = np.pad(x, pad_width=((0, 0), (0, 2)), mode='reflect')
y = pywt.swt(x_pad, wavelet='db4', level=3)
If you are not padding as above, it is expected that the max level will be one since 862 is divisible by 21, but not 22
If you are still seeing a problem, can you paste the script similar to the example above that reproduces the issue?
Ohhhh, nooo, this was so obvious. I padded the signal but to consider the 2*n not the exponent. So much in my face. Code's fine and thanks so much! Sometimes all it takes is a clear mind to point it out to you after spending to much time on it.
Great, glad it is working now.
| gharchive/issue | 2021-06-21T13:58:07 | 2025-04-01T04:32:59.603104 | {
"authors": [
"cavrinceanu",
"grlee77"
],
"repo": "PyWavelets/pywt",
"url": "https://github.com/PyWavelets/pywt/issues/592",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2086371886 | Run screenshotbot.io
Test screenshotbot.io integration
@PycKamil Arnold from Screenshotbot here (saw your accidental PR on swift-snapshot-testing). Feel free to ping me if you're stuck with anything :)
| gharchive/pull-request | 2024-01-17T14:57:07 | 2025-04-01T04:32:59.604889 | {
"authors": [
"PycKamil",
"tdrhq"
],
"repo": "PycKamil/swift-snapshot-testing",
"url": "https://github.com/PycKamil/swift-snapshot-testing/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
714317879 | Update README.md
Correct typo
Thanks for that
| gharchive/pull-request | 2020-10-04T13:47:30 | 2025-04-01T04:32:59.613865 | {
"authors": [
"KJonline",
"perjury"
],
"repo": "Pyhive/HA-Hive-Custom-Component",
"url": "https://github.com/Pyhive/HA-Hive-Custom-Component/pull/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2015503793 | Gerneated Tests are empty when run on a fresh project.
When i delete the pythagora_tests folder first and then generate the tests, it always generates me an almost empty unit-test file with only the lines
` ` ` ` javascript
in it.
After i leave the existing folder and just delete the unit-test-file, it will run fine again.
Typo in the title. Mus say: "Generated"
Update: Now i also get it, when the pythagora_tests folder already exists, while expanding tests with
npm run clean && npx pythagora --expand-unit-tests --force --path matchesValue.test.ts
The result is then:
Content of my handwritten tests
//Expanded tests using Pythagora:
```javascript
Seems a bit non-deterministic here.
| gharchive/issue | 2023-11-28T23:36:26 | 2025-04-01T04:32:59.635188 | {
"authors": [
"bogeeee"
],
"repo": "Pythagora-io/pythagora",
"url": "https://github.com/Pythagora-io/pythagora/issues/67",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.