added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:39:10.768681
2023-12-13T09:09:18
2039266988
{ "authors": [ "Mogost", "robin-blanchard" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7221", "repo": "jazzband/django-constance", "url": "https://github.com/jazzband/django-constance/issues/535" }
gharchive/issue
Unmodified value isn't set to new default when default is changed Describe the problem I didn't find in the docs what to expect when the default value is changed for some key while the value was not modified before change. I did find two distinct behaviours, based on whether or not the value has been modified or read in the past. Steps to reproduce First case: Key has not been modified or read Let's add a new entry to the constance config: CONSTANCE_CONFIG = { 'SOME_PARAMETER': ("old default", 'Test key', str), } Now, modify the configuration with a new default value: CONSTANCE_CONFIG = { 'SOME_PARAMETER': ("new default", 'Test key', str), } Conclusion, SOME_PARAMETER has been set to the new default value, as it was not modified before. Checking what is going on on the database side seems relevant in this issue: # SELECT * FROM constance_constance; id | key | value ----+-----+------- (0 rows) Second case: Read key before modifying default value Start with a similar configuration: CONSTANCE_CONFIG = { 'SECOND_PARAMETER': ("old default", 'Test key', str), } Now, read it from the console: # python manage.py constance get SECOND_PARAMETER old default Let's check the database now: # SELECT * FROM constance_constance; id | key | value ----+------------------+------------------------------ 7 | SECOND_PARAMETER | gAJYCwAAAG9sZCBkZWZhdWx0cQAu (1 row) Reading the key made it write to the database Let's try to change the default value with new default: Conclusion: SECOND_PARAMETER kept the old default value while it was never modified. Third case: Modify and reset value Again, let's start with a similar configuration: CONSTANCE_CONFIG = { 'THIRD_PARAMETER': ("old default", 'Test key', str), } From the Admin page, or the console, set it to any other value On the database side, the value is inscribed to it at this point And use 'Reset to default' Now, as in the second case, change the configuration to modify the default value. CONSTANCE_CONFIG = { 'THIRD_PARAMETER': ("new default", 'Test key', str), } Conclusion, resetting it to the default value doesn't make it act as if it has never been modified. Conclusion I do like the idea of unmodified values "following" the changes of default values, as in the first case. But I guess the current django-constance version doesn't feature it. Is there any reason for the lack of behaviour consistency as it is right now ? Would it be possible to have unmodified variables auto setting themselves to new default values ? It could also be configured with some CONSTANCE_XXX setting System configuration Django version: 3.2.19 Python version: 3.9.12 Django-Constance version: 3.1.0 This behavior comes from this line https://github.com/jazzband/django-constance/blob/master/constance/base.py#L22 So I did a bit of digging. What is happening when the constance isn't set inside the database ? Getting a new constance resulted in: Receiving the default value Having it written to the database. This comes from this line: https://github.com/jazzband/django-constance/blob/master/constance/base.py#L22 How to handle the cases in the issue Since the behaviour I was looking for is "Whenever a constance is set to the default value, I would like it to change when I change the corresponding default value", here's what I did: from constance.backends.database import DatabaseBackend from constance import settings, signals, config from django.db import ( OperationalError, ProgrammingError ) class CustomDatabaseBackend(DatabaseBackend): def set(self, key, value): if key not in settings.CONFIG: raise AttributeError(key) default_value = settings.CONFIG[key][0] if value != default_value: return super().set(key, value) queryset = self._model._default_manager.all() # Set _for_write attribute as get_or_create method does # https://github.com/django/django/blob/2.2.11/django/db/models/query.py#L536 queryset._for_write = True try: constance = queryset.get(key=key) except (OperationalError, ProgrammingError): # database is not created, noop return except self._model.DoesNotExist: # Key doesn't exist, then do nothing return old_value = constance.value # Constance should be modified to default value -> delete it constance.delete(using=queryset.db) if self._cache: self._cache.set(key, value) signals.config_updated.send( sender=config, key=key, old_value=old_value, new_value=value ) @robin-blanchard you are 100% correct. I am unsure, but this behavior might be added directly in django-constance. I'd consider a pull request with such a change. Perhaps it contains some additional constraints, but I don't see them at first glance.
2025-04-01T06:39:10.774790
2024-08-23T15:03:01
2483340691
{ "authors": [ "Mogost", "sebastianmanger" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7222", "repo": "jazzband/django-constance", "url": "https://github.com/jazzband/django-constance/issues/570" }
gharchive/issue
Missing migration in 4.0.0 Describe the problem When using django-constance with django>5, the command to check for missing migrations (makemigrations --check --dry-run) tells me that there is a missing migration. I guess this is because of the editable field, which is set here: https://github.com/jazzband/django-constance/blob/master/constance/migrations/0001_initial.py#L16, but it is not stated on the model (https://github.com/jazzband/django-constance/blob/master/constance/models.py#L7). The created migration looks like this: # Generated by Django 5.1 on 2024-08-23 15:01 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('constance', '0003_drop_pickle'), ] operations = [ migrations.AlterField( model_name='constance', name='value', field=models.TextField(blank=True, null=True), ), ] In case you can reproduce this, I am happy to help. Steps to reproduce In the example-project, install Django > 5 and run python manage.py makemigrations --check --dry-run Run tox as configured here: https://github.com/sebastianmanger/django-constance/commit/2357a61e64f698f18189508d0317806092f5c0f0. Example output: https://github.com/sebastianmanger/django-constance/actions/runs/10527827693/job/29171898640#step:5:542 System configuration Django version: >5 Python version: any Django-Constance version: latest (e428387) That was quick - thanks a lot! Maybe also include this check in tox.ini? django-admin makemigrations --check --dry-run Good idea! Could you suggest a PR?
2025-04-01T06:39:10.777498
2024-02-06T10:11:35
2120408373
{ "authors": [ "BlueMagma2" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7223", "repo": "jazzband/django-redshift-backend", "url": "https://github.com/jazzband/django-redshift-backend/pull/118" }
gharchive/pull-request
Fix: fix constraint creation using the wrong table and column name Subject: Fix a bug regarding foreign key constraint creation Feature or Bugfix Bugfix Detail This was causing issue if a foreign key was pointing to another field than the remote table id eg. Table1(models.Model): remote_column = models.IntegerField(unique=True) Table2(models.Model): local_column = models.ForeignKey(to=Table1, to_field='remote_column') During the migrations, you would get an error saying Table2 doesn't have a column named remote_column which is absurd since that colum is not supposed to be in that table. @shimizukawa I see you merged the last PR, though I admit it was some time ago. Can you have a look ? PS: I apologise for pinging you if you are the wrong person to address this Ha cool, I'm happy to see this was merged, thanks you
2025-04-01T06:39:10.808883
2015-11-06T23:14:51
115609844
{ "authors": [ "diasdavid", "jbenet" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7224", "repo": "jbenet/node-bsdash", "url": "https://github.com/jbenet/node-bsdash/issues/2" }
gharchive/issue
something broke jbenet @ lorien : ~/git/ipfs/specs * overviews % bsdash *** IPFS Bitswap Dash *** Active Requests 0: Provide Workers 0: Task Workers 0: Rebroadcast Worker: Provider Connector: events: 0/1 prints: 1 (25ms delay) /Users/jbenet/git/node-bsdash/state.js:23 var parts = e.event.split(".") ^ TypeError: Cannot read property 'split' of undefined at updateState (/Users/jbenet/git/node-bsdash/state.js:23:22) at Object.State.s.update (/Users/jbenet/git/node-bsdash/state.js:14:5) at DestroyableTransform._transform (/Users/jbenet/git/node-bsdash/index.js:26:13) at DestroyableTransform.Transform._read (/Users/jbenet/git/node-bsdash/node_modules/through2/node_modules/readable-stream/lib/_stream_transform.js:172:10) at DestroyableTransform.Transform._write (/Users/jbenet/git/node-bsdash/node_modules/through2/node_modules/readable-stream/lib/_stream_transform.js:160:12) at doWrite (/Users/jbenet/git/node-bsdash/node_modules/through2/node_modules/readable-stream/lib/_stream_writable.js:323:12) at writeOrBuffer (/Users/jbenet/git/node-bsdash/node_modules/through2/node_modules/readable-stream/lib/_stream_writable.js:309:5) at DestroyableTransform.Writable.write (/Users/jbenet/git/node-bsdash/node_modules/through2/node_modules/readable-stream/lib/_stream_writable.js:236:11) at write (/Users/jbenet/git/node-bsdash/node_modules/ndjson/node_modules/through2/node_modules/readable-stream/lib/_stream_readable.js:623:24) at flow (/Users/jbenet/git/node-bsdash/node_modules/ndjson/node_modules/through2/node_modules/readable-stream/lib/_stream_readable.js:632:7) @diasdavid @whyrusleeping any ideas? Seems that logs are different now, .event is not guaranteed to always exist: example 1 - Didn't break { event: 'log API client connected', session: 'b2b765ac-48b1-459c-8e1f-9fcd83868a3f', system: 'core/server', time: '2015-11-07T18:28:26.632599724Z' } example 1 - Error thrown { id: 5649, level: 5, message: 'bitswap net handleNewStream from <peer.ID bqE6Uf>', module: 'bitswap_network', time: '2015-11-07T18:28:30.288928034Z' } :////// need tests even on the logs
2025-04-01T06:39:10.821505
2024-04-25T13:05:21
2263539195
{ "authors": [ "ArthurConmy", "dtch1997", "jbloomAus" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7225", "repo": "jbloomAus/SAELens", "url": "https://github.com/jbloomAus/SAELens/pull/104" }
gharchive/pull-request
feat: gated sparse autoencoders Description Implements gated sparse autoencoders, which were recently found to solve the shrinkage problem and achieve equal reconstruction loss with 2x average feature sparsity Fixes #103 Type of change Please delete options that are not relevant. [x] New feature (non-breaking change which adds functionality) Checklist: [x] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] My changes generate no new warnings [x] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes [x] I have not rewritten tests relating to key interfaces which would affect backward compatibility You have tested formatting, typing and unit tests (acceptance tests not currently in use) [x] I have run make check-ci to check format and linting. (you can run make format to format code if needed.) Performance Check. If you have implemented a training change, please indicate precisely how performance changes with respect to the following metrics: [ ] L0 [ ] CE Loss [ ] MSE Loss [ ] Feature Dashboard Interpretability Please links to wandb dashboards with a control and test group. Wandb run ongoing: https://wandb.ai/dtch1997/gpt2?nw=nwuserdtch1997 Cluster is being annoying. Here's a colab to run the training. https://colab.research.google.com/drive/1q_X46e0eO9Q1_t283USX1QG41Q5fgDAZ?usp=sharing Comment from Neel: Awesome! I asked Sen (the lead author) for thoughts: The sparsity penalty isn't comparable between the two setups (or even the relationship between L1 and L0 norms IIRC). So the best way to compare them is to plot the Pareto curves. Very excited about this! Report from initial training runs: https://wandb.ai/sae-experiments/sae-experiments/reports/Gated-SAE-analysis--Vmlldzo3ODE5OTAw?accessToken=wcmc1fc13wlz3dze7zl26cmkblfdt6984fi20n79fgaa4x79l4moygyclx31mx3q Final WandB report: https://wandb.ai/sae-experiments/sae-experiments/reports/Gated-SAE-analysis--Vmlldzo3ODE5OTAw?accessToken=wcmc1fc13wlz3dze7zl26cmkblfdt6984fi20n79fgaa4x79l4moygyclx31mx3q Pareto plot: TODO: Clean up the cluster files Merge latest changes from main Thanks for the work so far! I read through the Gated SAE loss func implementation and it seemed right. If setting off runs is cheap, it would be great to get more Gated and Baseline runs in the L0 range 10-100 -- that's the most important region and there are not too many points there Thanks for this! I just merged a seperate PR for this but will look into your transcoder PR shortly!
2025-04-01T06:39:10.826152
2021-11-25T08:09:43
1063271976
{ "authors": [ "ZeSzymi", "denisenko93", "jbogard" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7226", "repo": "jbogard/MediatR", "url": "https://github.com/jbogard/MediatR/issues/675" }
gharchive/issue
Async Notifications dbContext disposed Hi I have problem with DbContext in async Notifications. This is my code for running async notifications: public class ParallelNoWaitPublishMediator : Mediator { private Func<IEnumerable<Func<INotification, CancellationToken, Task>>, INotification, CancellationToken, Task> _publish; public ParallelNoWaitPublishMediator(ServiceFactory serviceFactory) : base(serviceFactory) { _publish = ParallelNoWait; } protected override Task PublishCore(IEnumerable<Func<INotification, CancellationToken, Task>> allHandlers, INotification notification, CancellationToken cancellationToken) { return _publish(allHandlers, notification, cancellationToken); } private Task ParallelNoWait(IEnumerable<Func<INotification, CancellationToken, Task>> handlers, INotification notification, CancellationToken cancellationToken) { foreach (var handler in handlers) { Task.Run(() => handler(notification, cancellationToken), cancellationToken); } return Task.CompletedTask; } } And everytime i use DbContext in notification it gets disposed. The solution is to use service provider which creates new instance of dbContext when i call it, but I wanted something more generic. Is there any way to provide new scope for the handler, run in Task.Run() ? This is because the DbContext is registered with DI as a Scoped Service by default and asp creates and closes the scope on every request. There are several ways to fix this behavior: Register all services used in notification handlers as transient or singletone Override ServiceFactory, create custom scope and close it after complete all handlers Use IHostedService for executing you async notifications Use external libraries like Quartz.NET or Hangfire for executing you async notifications For option 2 there, here's some docs from MS on the subject: https://docs.microsoft.com/en-us/aspnet/core/fundamentals/host/hosted-services?view=aspnetcore-6.0&tabs=visual-studio#consuming-a-scoped-service-in-a-background-task
2025-04-01T06:39:10.845088
2016-04-07T18:09:36
146702445
{ "authors": [ "aalmada" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7228", "repo": "jbox-web/redmine_git_hosting", "url": "https://github.com/jbox-web/redmine_git_hosting/issues/589" }
gharchive/issue
Gemfile.lock issue when installing on Bitnami stack I'm following all the steps from 'get started' but I get stuck on the 'bundle install' in step 2. I get a "you are trying to install in deployment mode after changing your Gemfile". I tried the suggested at the bottom of set 2 but still get the same message. Environment: Redmine version 3.2.1.stable Ruby version 2.1.8-p440 (2015-12-16) [x86_64-linux] Rails version <IP_ADDRESS> Environment production Database adapter Mysql2 SCM: Subversion 1.8.13 Cvs 1.12.13 Git 2.6.1 Filesystem Redmine plugins: redmine_stats 0.0.3 Found an alternative in a note at https://wiki.bitnami.com/Applications/BitNami_Redmine#How_to_install_a_plugin_on_Redmine.3f Instructions are confusing though. I'll try to get help at the Bitnami forum...
2025-04-01T06:39:10.877360
2022-10-11T00:06:01
1403830733
{ "authors": [ "wraeth-eth" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7230", "repo": "jbx-protocol/juice-interface", "url": "https://github.com/jbx-protocol/juice-interface/pull/2188" }
gharchive/pull-request
use useArray in useRewards What does this PR do and why? Provide a description of what this PR does. Link to any relevant GitHub issues, Notion tasks or Discord discussions. Screenshots or screen recordings If applicable, provide screenshots or screen recordings to demonstrate the changes. Acceptance checklist [ ] I have evaluated the Approval Guidelines for this PR. [ ] I have tested this PR in all supported browsers. [ ] I have tested this PR in dark mode and light mode (if applicable). Current dependencies on/for this PR: main PR #2156 PR #2169 PR #2170 PR #2171 PR #2172 PR #2181 PR #2182 PR #2183 Graphite Merge Job Current status: โณ Queued to merge This pull request is currently queued to be merged as part of a stack. This comment was auto-generated by Graphite. Job Reference: BcMq9niLYsok9Kx1gVMU
2025-04-01T06:39:10.907634
2022-08-08T13:57:17
1331915360
{ "authors": [ "fabinsch" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7231", "repo": "jcarpent/proxqp-unofficial", "url": "https://github.com/jcarpent/proxqp-unofficial/pull/6" }
gharchive/pull-request
Add Julia example example of how to use proxqp in Julia using Pycall (https://github.com/JuliaPy/PyCall.jl). for now if i do cmake .. -DCMAKE_PREFIX_PATH=/home/fschramm/workspace/julia-1.7.3 it finds the julia executable (not installed in the path where it searches usually right now on my machine), it adds the unittest but when I run make test I get an error saying Start 112: example-jl-overview-simple.jl Process not started /home/fschramm/workspace/proxqp/examples/julia/overview-simple.jl [permission denied] 112/112 Test #112: example-jl-overview-simple.jl .......................................................................................................................................................................***Not Run 0.00 sec 99% tests passed, 1 tests failed out of 112 Total Test time (real) = 26.15 sec The following tests FAILED: 112 - example-jl-overview-simple.jl (BAD_COMMAND) Errors while running CTest I think the problem is this permission denied we see here. on my machine the command that should be executed by the test works without problems. So I'm still trying to figure out how to have the permission also when I run it here in ctest. already running on my machine, waiting for https://github.com/jrl-umi3218/jrl-cmakemodules/pull/538.
2025-04-01T06:39:10.931101
2016-03-17T10:24:30
141539798
{ "authors": [ "AlekzNet", "aliabbasjp", "dgcrouse" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7232", "repo": "jcjohnson/torch-rnn", "url": "https://github.com/jcjohnson/torch-rnn/issues/38" }
gharchive/issue
Stateful RNN mod in sampling Does just removing the lines: self:resetStates() in the sample function in langmodel.lua will make it a statefull RNN, thus preserving long term context? I removed these lines, and the loss immediately dropped by 15%. Thanks for the suggestion! But it can be very data-dependent. Interesting, every checkpoint saving resets the state. I wonder if this is the reason why the training loss jumps a bit up after every save: Epoch 1.24 / 50, i = 20 / 4250, loss = 1.025447 Checkpoint = cv/checkpoint_20.t7 Epoch 1.25 / 50, i = 21 / 4250, loss = 1.042428 . . . Epoch 1.35 / 50, i = 30 / 4250, loss = 1.013393 Checkpoint = cv/checkpoint_30.t7 Epoch 1.36 / 50, i = 31 / 4250, loss = 1.036500 . . . Epoch 1.47 / 50, i = 40 / 4250, loss = 1.003105 Checkpoint = cv/checkpoint_40.t7 Epoch 1.48 / 50, i = 41 / 4250, loss = 1.020587 Over a year since last comment, closing stale issue.
2025-04-01T06:39:10.954536
2017-01-05T22:05:01
199071699
{ "authors": [ "jcnelson" ], "license": "isc", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7233", "repo": "jcnelson/fskit", "url": "https://github.com/jcnelson/fskit/issues/17" }
gharchive/issue
User-supplied routes for chown() and chmod() Certain filesystems like Syndicate need to be able to handle chown and chmod themselves. Add routes for this, as well as tests. Added.
2025-04-01T06:39:11.017284
2024-03-08T23:42:33
2176929254
{ "authors": [ "Dynamotivation", "tupaschoal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7234", "repo": "jdm-contrib/jdm", "url": "https://github.com/jdm-contrib/jdm/pull/2003" }
gharchive/pull-request
Added Instructions for MathCad I've added the instructions for MathCad / PTC accounts. Please be aware there is another section about a PTC product listing "ptc.com" under its domains. I would appreciate some clarification how this duplication is handled. If one or the other entry takes priority upon visiting ptc.com I suggest we combine "Vuforia Chalk" (the other product) and MathCad into a "PTC Accounts" section instead and combine their URL list, since PTC aquired Vuforia. The Vuforia section has spelling mistakes and an to my knowledge outdated support article anyways. In that case give me a ping and I'll fix it :) That said, Vuforia developer accounts (developer.vuforia.com) are still separate from PTC accounts and would require their own section. It should all be taken care of now! At least until they change their account infrastructure again :) Vuforia Chalk account were changed to easy, since a "Delete Account" button now exists for the vast majority of end users and business accounts. Only organisation admins should need to use the "hard" fallback I have provided. Thanks for your contribution!
2025-04-01T06:39:11.023552
2014-06-27T14:45:52
36670419
{ "authors": [ "bmcbride", "germanbisurgi", "gnom7", "schmunk42" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7235", "repo": "jdorn/json-editor", "url": "https://github.com/jdorn/json-editor/issues/158" }
gharchive/issue
Required fields I'm having some issues figuring out how to make a field required. I've tried the following, which adds "required": ["comment"] to the object after defining the properties, but that doesn't seem to work. Any guidance would be most appreciated. Thanks! Hello, @jdorn I want to spend some time thinking about the best way to handle the asterisks before implementing that. Has it been implemented? @gnom7 https://github.com/jdorn/json-editor#deprecation-notice https://github.com/json-editor/json-editor/issues/44 I am not too familiar with the code, I'd recommend creating a new issue at https://github.com/json-editor/json-editor/issues The documentation says you can use custom styles but does not say you can use custom classes. When you using barebones JSON editor it uses almost no classes that can override your styles. I looked in the source code looking for a feature that does what you want to do but I did not find anything. A solution could be, to extend the theme class and create a new theme for JSON Editor. You can look in the original repository wiki: https://github.com/jdorn/json-editor/wiki#theme-srcthemejs @germanbisurgi, thank you for the suggestion, but it seems that I can't customize form elements this way, only DOM ones (like tables etc.) and it might happen that I'd like to pull one table to the right and another to the left in my form depending on the element which is rendered, or even something more sophisticated, so it would be nice if we had feature to add classes to json form elements, thanks for reply
2025-04-01T06:39:11.044220
2019-02-15T10:54:38
410712737
{ "authors": [ "SunryTeang", "abhishek-singh-a4", "jeancsanchez", "rohankandwal" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7236", "repo": "jeancsanchez/JcPlayer", "url": "https://github.com/jeancsanchez/JcPlayer/issues/79" }
gharchive/issue
Killing notification on App swipe Duplicate of https://github.com/jeancsanchez/JcPlayer/issues/22 But the problem there is, onDestroy doesn't get called when we clear the app from recent task. This is causing crashes when we click on the notification again. onDestroy tricks work well on exiting the app, but not swiping and killing it from tasks. That's a problem that I can't solve yet. I didn't find a solution for removing the notifications when the app is popped from the tasks list. @jeancsanchez I have actually managed to resolve it on your Sample code, but it is not working on my code when I add library as gradle module. Once I am able to fix it, will let you know. that problem error on emulator if real device is work fine @rohankandwal @SunryTeang I have same problem in multiple real devices. ERROR: Failed to resolve: com.github.jeancsanchez:JcPlayer:2.6.16 sir how can i solve it
2025-04-01T06:39:11.200120
2023-02-19T09:39:42
1590623783
{ "authors": [ "jeffhammond" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7237", "repo": "jeffhammond/mukautuva", "url": "https://github.com/jeffhammond/mukautuva/issues/12" }
gharchive/issue
MPI_TYPE_GET_CONTENTS with nested user types is broken From https://github.com/jeffhammond/mukautuva/issues/11. If we do e.g. vector of vectors, MPI_TYPE_GET_CONTENTS does not work. https://github.com/jeffhammond/mukautuva/commit/4539c1c4711fec8cae10977b8527f2931a75c450 fixes this
2025-04-01T06:39:11.798052
2023-03-15T13:52:35
1625574266
{ "authors": [ "ayberkcanturk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7238", "repo": "jellydn/next-swagger-doc", "url": "https://github.com/jellydn/next-swagger-doc/issues/646" }
gharchive/issue
`Argument of type 'Promise<typeof import("...@types/swagger-ui-react/index")>' is not assignable to parameter of type 'DynamicOptions<{ spec: any; }> | Loader<{ spec: any; }>'. Describe the bug I am trying to show an open API document and followed same example on the github. I am receiving this error. Do you have any idea how to fix this? Thank you. Argument of type 'Promise<typeof import("c:/workspace/***/***/node_modules/@types/swagger-ui-react/index")>' is not assignable to parameter of type 'DynamicOptions<{ spec: any; }> | Loader<{ spec: any; }>'. Type 'Promise<typeof import("c:/workspace/***/***/node_modules/@types/swagger-ui-react/index")>' is not assignable to type 'LoaderComponent<{ spec: any; }>'. Type 'typeof import("c:/workspace/***/***/node_modules/@types/swagger-ui-react/index")' is not assignable to type 'ComponentType<{ spec: any; }> | ComponentModule<{ spec: any; }>'. Type 'typeof import("c:/workspace/***/***/node_modules/@types/swagger-ui-react/index")' is not assignable to type 'ComponentModule<{ spec: any; }>'. Types of property 'default' are incompatible. Type 'typeof SwaggerUI' is not assignable to type 'ComponentType<{ spec: any; }>'. Type 'typeof SwaggerUI' is not assignable to type 'ComponentClass<{ spec: any; }, any>'. Construct signature return types 'SwaggerUI' and 'Component<{ spec: any; }, any, any>' are incompatible. The types of 'props' are incompatible between these types. Type 'Readonly<SwaggerUIProps>' is not assignable to type 'Readonly<{ spec: any; }>'. Property 'spec' is optional in type 'Readonly<SwaggerUIProps>' but required in type 'Readonly<{ spec: any; }>'.ts(2345) Reproduction Followed example on github README.md System Info "next": "13.0.7", "next-swagger-doc": "^0.3.6", "swagger-ui-react": "^4.18.1", System: OS: Windows 10 10.0.22621 CPU: (16) x64 AMD Ryzen 9 5900HX with Radeon Graphics Memory: 10.91 GB / 31.35 GB Binaries: Node: 16.13.0 - C:\Program Files\nodejs\node.EXE Yarn: 1.22.11 - ~\AppData\Roaming\npm\yarn.CMD npm: 8.1.0 - C:\Program Files\nodejs\npm.CMD Browsers: Edge: Spartan (44.22621.1344.0), Chromium (110.0.1587.69) Internet Explorer: 11.0.22621.1 Used Package Manager yarn Validations [X] Follow our Code of Conduct [X] Read the Contributing Guide. [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. [X] Check that this is a concrete bug. For Q&A, please open a GitHub Discussion instead. [X] The provided reproduction is a minimal reproducible of the bug. I found the issue. I have used: const SwaggerUI = dynamic<SwaggerUIProps>(import("swagger-ui-react"), { ssr: false, }); instead of const SwaggerUI = dynamic<{spec: any}>(import("swagger-ui-react"), { ssr: false, }); Thank you
2025-04-01T06:39:11.866220
2024-01-23T19:13:24
2096806187
{ "authors": [ "jemmorey" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7239", "repo": "jemmorey/upptime", "url": "https://github.com/jemmorey/upptime/issues/142" }
gharchive/issue
โš ๏ธ SpeakUpConference.com has degraded performance In 1493e04, SpeakUpConference.com (https://www.speakupconference.com) experienced degraded performance: HTTP code: 200 Response time: 9878 ms Resolved: SpeakUpConference.com performance has improved in 84e96c2 after 11 minutes.
2025-04-01T06:39:11.879474
2022-12-16T14:07:55
1500288641
{ "authors": [ "NotMyFault", "lemeurherve" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7240", "repo": "jenkins-infra/repository-permissions-updater", "url": "https://github.com/jenkins-infra/repository-permissions-updater/pull/3032" }
gharchive/pull-request
Provide a GH permissions template See https://github.com/jenkins-infra/helpdesk/pull/3299 I removed the permissions level dropdown, given teams typically have admin permissions over the repository. Looks fine, you might want to add a issue template config and change the number in the filename.
2025-04-01T06:39:11.884181
2018-03-08T22:25:55
303655080
{ "authors": [ "dwnusbaum", "oleg-nenashev" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7241", "repo": "jenkins-infra/repository-permissions-updater", "url": "https://github.com/jenkins-infra/repository-permissions-updater/pull/624" }
gharchive/pull-request
Grant @dwnusbaum permission to release cloudbees-folder Description Grant @dwnusbaum permission to release https://github.com/jenkinsci/cloudbees-folder-plugin. CC @oleg-nenashev @jglick @recena for confirmation from existing maintainers. Submitter checklist for changing permissions Always [x] Add link to plugin/component Git repository in description above When adding new uploaders (this includes newly created permissions files) [x] Make sure to @mention an existing maintainer to confirm the permissions request, if applicable [x] Use the Jenkins community (LDAP) account name in the YAML file, not the GitHub account name [x] All newly added users have logged in to Artifactory at least once @dwnusbaum are you also fine with being a default assignee in JIRA? Currently there is no default assignee, and the tickets appear to be untriaged: https://issues.jenkins-ci.org/browse/JENKINS-49988?jql=project %3D JENKINS AND status in (Open%2C "In Progress"%2C Reopened) AND component %3D cloudbees-folder-plugin @oleg-nenashev Sure you can make me the default.
2025-04-01T06:39:11.900686
2018-01-19T11:14:35
289941720
{ "authors": [ "jstrachan", "rawlingsj" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7242", "repo": "jenkins-x/cloud-environments", "url": "https://github.com/jenkins-x/cloud-environments/issues/20" }
gharchive/issue
make the use of the domain easier to override? the exposecontroller settings and domain are encoded in a secret for thunder which makes it hard to use locally or to change the domain etc. I wonder if we can make it easier to change those? As a work around I hacked the secrets.yaml.dec locally then added these new targets in the Makefile... local-install: helm install jenkins-x/$(CHART) --name $(RELEASE) -f ./myvalues.yaml -f ./secrets.yaml.dec --version $(CHART_VERSION) $(ARGS) local-upgrade: helm upgrade $(RELEASE) jenkins-x/$(CHART) -f myvalues.yaml -f secrets.yaml.dec --version $(CHART_VERSION) $(ARGS) wonder if there's a cleaner way? Not an issue anymore
2025-04-01T06:39:11.965409
2016-07-20T07:52:10
166518875
{ "authors": [ "benpatterson", "martenson", "ssbarnea" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7243", "repo": "jenkinsci/ghprb-plugin", "url": "https://github.com/jenkinsci/ghprb-plugin/issues/386" }
gharchive/issue
publish change list on all releases Please publish change list as part of each release. Currently the plugin points to the Jenkins wiki which states that after version x the change-lists moved to a Github page. When you click this 2nd link you reach to a page that again is missing the current changes. Please fix this so when you click the URL from the plugin manager you would go to a page that DOES contain the change-lists, allowing you to make an informed decision. Jenkins plugins do have the bad habit of breaking Jenkins when updated, please don't make this maintenance process even harder. duplicate of https://github.com/jenkinsci/ghprb-plugin/issues/360 Hi - I volunteered to help out with this repo. I'll group this with 360 as @martenson pointed out and look to address this problem. thanks @benpatterson !
2025-04-01T06:39:12.042025
2017-09-15T16:39:30
258098757
{ "authors": [ "jimklimov" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7244", "repo": "jenkinsci/lockable-resources-plugin", "url": "https://github.com/jenkinsci/lockable-resources-plugin/issues/73" }
gharchive/issue
Add documentation example on how to pass build arguments into the Groovy script Context: the Lockable Resources plugin includes an option to specify the needed resource by one of: exact resource name, matching a label, or evaluating a Groovy script for each configured resource that would return true if this resource is okay for this job. Then one of the "okay" resources (assuming "Number of resources to request" setting is 1) is reserved while the build runs. My use-case, now solved by code in PR #72, is that we have a test farm, where usually any of the systems can be used for a test (so label-matching is in place), except when developers want to run the job against some particular system - and protect it from being used by any other job during this time. The build job (or actually a Multiphase job calling several others) has build arguments, including specification of the environment to use for the test (a "CONTROLLER" in examples below). For the test, I used a build of the plugin with PR above integrated (made by mvn package), and in the job I selected "Meta Data / This build requires lockable resources", further selected "Groovy Expression" and entered the script below. The "Resources" and "Label" fields were left empty, and the "Number of resources to request" was set to 1 (disregard the warning that Given amount 1 is greater than amount of resources: 0.), and "Reserved resources variable name" is LOCKED_CONTROLLER. Note the println lines below end up in jenkins.log, and can be pretty noisy, so comment them away when your job definition works well :) // We need an available testbed resource marked with label "rc-validation-farm" // If the user specified a particular testbed name in CONTROLLER var, require that one instead (regardless of labels) /* println "Inspecting the resource to lock for requested CONTROLLER='" + CONTROLLER + "' (looking at resourceName='" + resourceName + "' resourceDescription='" + resourceDescription + "' resourceLabels='" + resourceLabels + "')" */ if ( CONTROLLER.startsWith("LOCK_LABEL:") ) { def LOCK_LABEL = (CONTROLLER =~ /^LOCK_LABEL:(.*?)$/ )[0][1]; /* println "Looking for LOCK_LABEL='" + LOCK_LABEL + "' among '" + resourceLabels + "' for '" + resourceName + "' (" + resourceDescription + ")" */ if (resourceLabels.contains( LOCK_LABEL )) { // println "ACCEPTED '" + resourceName + "'" return true; } } else { // println "Looking for 'rc:" + CONTROLLER + "' in the name '" + resourceName + "'" if (resourceName == ("rc:"+CONTROLLER) ) { // println "ACCEPTED '" + resourceName + "'" return true; } } // println "Resource '" + resourceName + "' is not suitable for this job" return false; // Tested resource is not appropriate for this build The numerous corresponding lockable resource definitions in Manage Jenkins define a name (like rc:controller1) and labels (like rc-validation-farm rc-model-3), and sometimes comments about nuances of the controller model. The job build parameter CONTROLLER is a predefined Global Choice Parameter, which includes the names like controller1 that our devs can pick, and a default value of LOCK_LABEL:rc-validation-farm. The script above was made generic enough to ease copy-pasting, so it reacts to the build arguments starting with LOCK_LABEL: and picks the rest of the string as the label to look for in resources; otherwise it would look for the requested resource name (with the prefix rc:). As you could guess, there is a second side to the medal: the default CONTROLLER build argument is not usable as a definitive host name, so that build argument has to be replaced with the actually picked value :) For this we use the EnvInject plugin, so in the job I further selected "Prepare an environment for the run" and entered an "Evaluated Groovy script" with: def map = [:] /* println "Requested build arg CONTROLLER=='" + CONTROLLER + "'; the locked resource == '" + LOCKED_CONTROLLER + "'"; */ if ("LOCK_LABEL:rc-validation-farm".equals(CONTROLLER)) { def CTLNAME = ( (LOCKED_CONTROLLER =~ /^rc:(.*?)$/)[0][1] ) println "Extracted CONTROLLER:='" + CTLNAME + "'" map << [ CONTROLLER : CTLNAME ] } return map (the script actually evaluates optional overrides for a few other variables as well, hence this structure and not a simpler one). Note that println here ends up in the build job's log. Finally note, that for sub-jobs called from this one, you should also specify "Predefined parameters" including CONTROLLER=${CONTROLLER} so the one mapped above is in place for those jobs that would access it, and not the default token. The "Build on same node" and "Current build parameters" are useful for such cases as well :) UPDATE: The pull request has been merged for a while now. Over the past couple of years I've referred a number of IRC discussions to this issue. I believe it has thus become a good basis for a documentation chapter about using this trick, so I rephrased the issue name to have it only closed after adding published documentation rather than keeping it a needle of know-how lost in the haystack.
2025-04-01T06:39:12.058346
2015-12-03T11:38:30
120151475
{ "authors": [ "Kentzo", "TheMeier", "jhoblitt" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7245", "repo": "jenkinsci/puppet-jenkins", "url": "https://github.com/jenkinsci/puppet-jenkins/issues/443" }
gharchive/issue
no ordering in config_hash jenkins::sysconfig does not support any kind of ordering. if you need to change JENKINS_ARGS this might lead to a broken sysconfig: example given this config_hash in yaml: jenkins::config_hash: HTTP_PORT: value: -1 HTTPS_PORT: value: 443 HTTPS_CERT: value: '/some/path' HTTPS_KEY: value: '/some/patjh' JENKINS_ARGS: value: '--webroot=/var/cache/$NAME/war --httpPort=$HTTP_PORT --ajp13Port=$AJP_PORT --httpsPort=$HTTPS_PORT --httpsCertificate=$HTTPS_CERT --httpsPrivateKey=$HTTPS_KEY' you might end up with this sysconfig: # port for HTTP connector (default 8080; disable with -1) HTTP_PORT="-1" # port for AJP connector (disabled by default) AJP_PORT=-1 # servlet context, important if you want to use apache proxying PREFIX=/$NAME # arguments to pass to jenkins. # --javahome=$JAVA_HOME # --httpPort=$HTTP_PORT (default 8080; disable with -1) # --httpsPort=$HTTP_PORT # --ajp13Port=$AJP_PORT # --argumentsRealm.passwd.$ADMIN_USER=[password] # --argumentsRealm.roles.$ADMIN_USER=admin # --webroot=~/.jenkins/war # --prefix=$PREFIX JENKINS_ARGS="--webroot=/var/cache/$NAME/war --httpPort=$HTTP_PORT --ajp13Port=$AJP_PORT --httpsPort=$HTTPS_PORT --httpsCertificate=$HTTPS_CERT --httpsPrivateKey=$HTTPS_KEY" HTTPS_PORT="443" HTTPS_CERT="/some/path" HTTPS_KEY="/some/patjh" since the variables are set in the wrong order variable expansion does not work... Hash are inherently unordered. I'm not sure what are you suggesting? Clearly i didn't think this through ;) Hash is a bad idea for defaults as this must be ordered, otherwise it makes config_hash quite useless. @Kentzo There's no interpolation under systemd. The order of the environmentfile directives is meaningless. That's not true when env vars depend on one another. E.g. in /etc/default/jenkins JENKINS_ARGS="--webroot=/var/cache/$NAME/war --httpPort=$HTTP_PORT --httpListenAddress=<IP_ADDRESS>" which makes order meaningful. See the comment in https://github.com/jenkinsci/puppet-jenkins/blob/master/templates/jenkins-slave-defaults.erb#L3-L4 and https://www.freedesktop.org/software/systemd/man/systemd.exec.html#Environment EnvironmentFile declarations are parsed into a char ** and set as environ(7). There is no interpolation. I'm not sure how it works, but this is the state of /etc/default/jenkins after installing jenkins with this module on Ubuntu 16.04 Do you mean it's buggy? Debian hasn't been switched over to systemd unit files, yet.
2025-04-01T06:39:12.062097
2017-12-07T16:57:17
280204016
{ "authors": [ "nhathy", "victor-accarini" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7246", "repo": "jenkinsci/slack-plugin", "url": "https://github.com/jenkinsci/slack-plugin/issues/360" }
gharchive/issue
Feature: Custom messages for success and failure Hi, I would like to add a different custom message when my build fails or succeeds. As far as I know I can set a custom message that will be show in all cases. I made something that works the way I want in victor-accarini/slack-plugin, but I would like a feedback from you guys since I'm not a Java dev. Thanks! PS: @gurumaia Are you still maintaining this plugin? Possibly fixed in #169
2025-04-01T06:39:12.066596
2015-02-03T11:07:58
56363673
{ "authors": [ "Brantone", "ctapobep", "jenkinsadmin", "ndeloof" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7247", "repo": "jenkinsci/template-project-plugin", "url": "https://github.com/jenkinsci/template-project-plugin/pull/13" }
gharchive/pull-request
Fix for using with CloudBees Folder plugin https://issues.jenkins-ci.org/browse/JENKINS-24396 Thank you for a pull request! Please check this document for how the Jenkins project handles pull requests Is anyone maintaining this plugin any more?? I don't think so. Might make sense to take over the ownership and maintain the plugin on your own. And don't forget about thorough testing :) Looks like @ndeloof was the last to merge anything ... any insight? Heh, that was awesome and fast, thank you kindly. Are there still plans for maintenance and up-keep, or need a 2nd set of hands to help out? Hit me up on e-mail, can discuss further there. In mean-time ... thanks again!! :) I can't see active maintenance on this plugin, so feel free to request ownership on jenkins-dev ML Cheers.
2025-04-01T06:39:12.068591
2020-06-14T10:30:14
638325193
{ "authors": [ "oleg-nenashev", "timja" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7248", "repo": "jenkinsci/theme-manager-plugin", "url": "https://github.com/jenkinsci/theme-manager-plugin/issues/6" }
gharchive/issue
Convert to the Incubated project before the release? Hi @timja , I would suggest to make it an incubated project so that we have more freedom w.r.t next steps (e.g. merging it into the Jenkins core). I would suggest the following stapes: Add "Incubated project" to the description and docs, similar to https://plugins.jenkins.io/pipeline-as-yaml/ Mark all APIs with @Restricted(Beta.class) Sure
2025-04-01T06:39:12.156524
2021-11-04T08:04:26
1044453531
{ "authors": [ "jeremymailen", "oradkovsky" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7249", "repo": "jeremymailen/kotlinter-gradle", "url": "https://github.com/jeremymailen/kotlinter-gradle/issues/225" }
gharchive/issue
How to configure indentation for the case like this Hello! I'm using 3.3.0 version with Android Studio 2020.3.1 patch 3. Kotlin linter complains about this code: Unexpected indentation (16) (should be 12) What is proper way to configure indentation in such example? KLint supports a few different options for configuring indents, kotlinter-gradle will support all the values you might place in .editorconfig https://github.com/pinterest/ktlint#editorconfig It's hard to tell from your screenshot what the context of those string literals is? Is it in line with the android style guide? Should it be a triple quote """ literal? Thanks for great lib. Will keep on investigating. Even https://ktlint-demo.herokuapp.com/ behaves the same with the same code snippet.
2025-04-01T06:39:12.158128
2013-08-09T17:34:45
17875727
{ "authors": [ "TheSisb", "gjuchault" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7250", "repo": "jeresig/i18n-node-2", "url": "https://github.com/jeresig/i18n-node-2/issues/24" }
gharchive/issue
Is there any way to check if the browser locales are in the application locales list? I'm using express so I'm getting the user's locales with req.headers['accept-language'] Now I have a string of locales: zh,zh-CN;q=0.8,zh-TW;q=0.6,en;q=0.4,en-US;q=0.2 My accepted locales are I18n.expressBind(app, { locales: ['en', 'zh-CN'] }); ``` Is there any way I can find the client's most preferred locale that is also on my list? Closed (the comment fixes the answer)
2025-04-01T06:39:12.168432
2022-10-02T01:45:01
1393610349
{ "authors": [ "enobufs", "jerry-tao" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7251", "repo": "jerry-tao/sctp", "url": "https://github.com/jerry-tao/sctp/pull/1" }
gharchive/pull-request
Reset outgoing stream on inbound is reset Added state to Stream to reset outbound only when the state is open Relates to #187 Hi @jerry-tao, what do you think about this fix. (this branch is based on your fork jerry-tao/sctp:master). Who to react to the incoming OutgoingResetRequest should be Stream layer (not at the Association layer). Also, if the stream already sent its OutgoingResetRequest, it shouldn't send another one, meaning, we need to introduce a state in the Stream class (s.state), and send OutgoingResetRequest if the state is "open", I believe. Which prevented the vnet test case from failing. As RFC 8831 implies, teardown sequence of a stream starts with outgoing (closing), then incoming (closed). This branch also addresses it by not to cause error (s.readErr) because it will eventually be reset by the remote. I am hoping to merge this into your branch, then we can finalize it in your pull-request. @enobufs Sorry for the delay, just come back form vacation.
2025-04-01T06:39:12.171475
2017-02-21T08:36:04
209074710
{ "authors": [ "robertsipka" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7252", "repo": "jerryscript-project/jerryscript", "url": "https://github.com/jerryscript-project/jerryscript/pull/1601" }
gharchive/pull-request
Remove shadowed declarations, undefined identifiers, and specify argument types where it is required. Thanks, I've updated this patch based on your suggestions.
2025-04-01T06:39:12.178024
2021-01-01T18:07:28
777327753
{ "authors": [ "akosthekiss", "lygstate" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7253", "repo": "jerryscript-project/jerryscript", "url": "https://github.com/jerryscript-project/jerryscript/pull/4394" }
gharchive/pull-request
Master fork PLEASE REMOVE THIS TEMPLATE BEFORE SUBMITTING Before submitting a PR, please, make sure that: Changes are in a separate branch, not in master. The branch contains only one commit on top of master (if not, squash them into one commit). The commit message contains fixes #XXXX or closes #XXXX to auto-close the issue(s) that the PR fixes (if any). Tests for the changes have been added (for bug fixes / features). Documentation has been added / updated (if applicable). All new and existing tests passed locally (if not, fix them first and amend the commit). IMPORTANT: Please review the CONTRIBUTING.md file for detailed contributing guidelines. PLEASE REMOVE THIS TEMPLATE BEFORE SUBMITTING @lygstate Please, explain what you are doing. You cannot open a PR by mistake. You cannot open PRs by mistake TWICE. @lygstate Please, explain what you are doing. You cannot open a PR by mistake. You cannot open PRs by mistake TWICE. @lygstate Please, explain what you are doing. You cannot open a PR by mistake. You cannot open PRs by mistake TWICE. It's a mistake and also by github, I am trying to PR to my own jerryscript repo to trigger the CI, but it's default to jerryscript not to my own and this is the result @lygstate Please, explain what you are doing. You cannot open a PR by mistake. You cannot open PRs by mistake TWICE. It's a mistake and also by github, I am trying to PR to my own jerryscript repo to trigger the CI, but it's default to jerryscript not to my own and this is the result @lygstate Now that #4395 has landed, it should be easier to trigger the CI. Any branch you push to your fork will trigger GH Actions on your fork, even without a PR. @lygstate Now that #4395 has landed, it should be easier to trigger the CI. Any branch you push to your fork will trigger GH Actions on your fork, even without a PR. @akosthekiss thanks. @akosthekiss thanks.
2025-04-01T06:39:12.190102
2016-09-27T21:05:24
179613632
{ "authors": [ "boxdot", "coveralls", "jeschkies" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7254", "repo": "jeschkies/renderer", "url": "https://github.com/jeschkies/renderer/pull/60" }
gharchive/pull-request
Triangle subdivide Also contains several fixes in the kdtree. Therefore lets just merge it onto master and continue from there with radiosity and subdivision. Coverage decreased (-1.8%) to 46.033% when pulling 336b657c9688c05d9955cf9858efea78b0012551 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage increased (+1.5%) to 49.365% when pulling 3c692fc1febc132889896250b508f498c82fbc50 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.8%) to 46.033% when pulling 21452b4ed6144a3a83c14069051445b5dc3fb054 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.05%) to 46.779% when pulling f8ed7d1cda90710966a28282d8b55675233e7bb7 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.8%) to 46.033% when pulling 71915703b578fd40ccbb0a29587aac297e58c4ff on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. KDTree::build kannst du ja mit einer kleinen Box aufrufen https://coveralls.io/builds/8119862/source?filename=lib%2Fkdtree.cpp#L101 So coverst du box.surface_area() == 0. Coverage decreased (-1.7%) to 46.082% when pulling df22071f2f522f1af3433c326cc8590b0a5b6ff6 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.1%) to 46.708% when pulling e0ff81e96e487819891ff520c4c524cd7b741839 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-0.7%) to 47.126% when pulling 01d71371778cdff9edf123721bb478ffa13aaa09 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Tests done. Coverage decreased weil wir radiohit erweitert haben und da nichts testen. Awesome work!
2025-04-01T06:39:12.194048
2024-08-26T11:29:41
2486632383
{ "authors": [ "WH-Yoshi", "jesper-raemaekers" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7255", "repo": "jesper-raemaekers/python-polarion", "url": "https://github.com/jesper-raemaekers/python-polarion/pull/181" }
gharchive/pull-request
Project group and some optimization for faster retrieval of workitems. I do this PR to add my personal use case (project group) as it can be useful. It's only here as a workitem retriever for now but can be largely modified. It's implemented so it's not more complicated to use than a project. I believe that my code is clean, but any review is very welcomed as I'm not an expert. Looks good to me, never used it but seems like a nice addition. Also thanks for the cleanup. I'll merge it if you could resolve the merge conflict please
2025-04-01T06:39:12.205623
2015-08-14T21:01:25
101102747
{ "authors": [ "jesserayadkins" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7256", "repo": "jesserayadkins/lily", "url": "https://github.com/jesserayadkins/lily/issues/126" }
gharchive/issue
Allow a notation that's just "This belongs to a certain class". Partial types, I suppose. So the original idea here was to take generics, and impose a constraint on them. It wouldn't be a complete constraint, only something that would say "Anything of this particular kind of a class". It was something like A > function specification for generics. There are two problems with this approach: How Lily distinguish between someone that wants a partial type constraint, and someone who goofed. I'd like to avoid a symbol soup here, so A > ~function doesn't set well. How does one create a list of partially typed values, since the constraint is against a generic? Uhh...You kinda don't. Because you can't refer to generics outside of a generic function. So, on to the new idea. Let's say you're trying to make a list of those partially a function things. It's gonna look like this: var v: list[~function] = [list::select, integer::to_s, string::split] So now you've got a list of things that are a function. That's what the tilde is for: It says "these things are all just a function. I don't know any more about them." Since there's an annotation for ~function supplied, Lily will create a list of kinda-functions. Otherwise, it will retain the same behavior. The reason for this, is that I don't like that sort of behavior (unifying in a way that reduces away all subtype information) without explicit intent. Lily understands variance, and will also allow functions to specify that they want a partial function. Also, with that in mind, Lily should have no problem determining that any particular kind of a function with arguments or a return type of any sort is a function. Lily having a better understanding of variance was also a motivation for this change. Initially, only functions will be able to be specified as being partial (no subtype information). However, in the future, I'll extend this if it turns out there's a use for it on other types. What I'm trying to solve can easily be solved by having a callable type which would act as a parent type to what function is. There would be no need for any of the above to be done, as Lily understands when doing variance to follow the parent's generic count. So just adding a callable type and maybe refining a couple areas is all that's necessary. I'll get around to it when I add the doc thing specified by #127. No need for a whole bug and a new syntax. Besides, I can only see adding what this asks for as being a problem later (Valid questions like "Why not partial lists, partial hashes, partial higher kinds begin showing up). So there's no need for a special bug for this, I think.
2025-04-01T06:39:12.211502
2016-09-07T13:54:24
175508657
{ "authors": [ "Komal2905", "jessesquires" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7257", "repo": "jessesquires/JSQMessagesViewController", "url": "https://github.com/jessesquires/JSQMessagesViewController/issues/1820" }
gharchive/issue
How to set Image and text in same Bubble cell New issue checklist [x] I have read all of the README, documentation, and FAQ. [x] I have reviewed the contributing guidelines. Confirmation: ____ [x] I have searched existing issues and this is not a duplicate. General information Library version(s): iOS version(s): Devices/Simulators affected: Reproducible in the demo project? (Yes/No): Related issues: Bug report Expected behavior Actual behavior Steps to reproduce Crash log? Screenshots? Videos? Sample project? Thanks! :smile: Please provide all the information in the issue template so we can help! And don't forget to read our contribution guidelines. :blush:
2025-04-01T06:39:12.220662
2023-08-27T16:53:05
1868560569
{ "authors": [ "Delapouite", "SimenB", "vire" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7259", "repo": "jestjs/jest", "url": "https://github.com/jestjs/jest/issues/14454" }
gharchive/issue
[Bug]: node18 memory consumption growth up to25% Version 29.6.4 Steps to reproduce This happens on a NX powered private repo - on demand I can produce a reproduction repo if needed Expected behavior with node16 our test suite takes about 2GB memory in run in band mode Actual behavior with node18 our test suite takes about 2.6GB memory and has hard time to finish on CircleCI Additional context During memory leak hunt I can see that string allocation loads the same modules several times leading to significant growth, shouldn't this in --runInBand mode never happen with the given cache/resolver? Environment System: OS: macOS 13.3.1 CPU: (8) arm64 Apple M1 Pro Binaries: Node: 18.17.0 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/node Yarn: 3.6.2 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/yarn npm: 9.6.7 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/npm Update: I've tried --no-cache and this had no effect on loaded modules Probably related to https://github.com/jestjs/jest/issues/11956 Yeah, dupe of that one
2025-04-01T06:39:12.238095
2018-07-15T06:36:23
341296543
{ "authors": [ "JaneJeon", "SimenB", "aaronabramov", "alycda", "chrismwendt", "dandv", "denissabramovs", "fider", "medikoo", "pkuczynski", "svyandun", "thernstig", "vtgn" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7260", "repo": "jestjs/jest", "url": "https://github.com/jestjs/jest/issues/6695" }
gharchive/issue
failing beforeAll() causes even passing tests in the scope to fail ๐Ÿ› Bug Report I'm looking for the best practice for aborting a describe block if the test setup fails. For example, when testing an API, if authenticating fails, it's pointless to run any other tests. As suggested in the issue template, I searched StackOverflow first, where I found this incomplete answer, advising to place the initialization code in a beforeAll block. Hence, raising the issue here because I haven't seen anything documenting the practice of returning failure from beforeAll If a beforeAll block fails, tests in that describe block are still run, and they fail, even if otherwise they would pass! To Reproduce describe('test that a 3rd party API remains consistent', () => { beforeAll(() => expect('login').toBe('successful')); // this will fail test('API function 1', () => expect(1).toBe(1)); // each... test('API function 2', () => expect(2).toBe(2)); // ...of these... test('API function 3', () => expect(3).toBe(3)); // ...will fail too }); Expected behavior Jest should report that the beforeAll() failed, and bail the describe scope without executing further tests in it. If it's somehow intended behavior that all tests should still be executed, and still marked as failed (which I find odd), this should be documented under beforeAll and in the Setup and Teardown guide. Other tests in the file, outside of the failing block, should still be executed. Link to repl or repo (highly encouraged) https://repl.it/@DanDascalescu/beforeAll-failure-should-bail-the-test Run npx envinfo --preset jest System: OS: Linux 4.15 Ubuntu 16.04.4 LTS (Xenial Xerus) CPU: x64 Intel(R) Core(TM) i7-7500U CPU @ 2.70GHz Binaries: Node: 10.5.0 - /usr/local/bin/node Yarn: 1.7.0 - /usr/bin/yarn npm: 6.2.0 - /usr/local/bin/npm npmPackages: jest: ^23.4.1 => 23.4.1 See also #6527 @aaronabramov have we changed the behaviour here for circus? Bailing after a failing lifecycle hook makes sense to me yes! this was a bug in jasmine (or a feature). We changed this behavior in jest-circus. @dandv jest-circus isn't shipped with jest by default yet, but you can test it by yarn add jest-circus and adding "testRunner": "jest-circus/runner" to your jest config I would like to add to this a request for named hooks. The test is marked as a failure and not run, but named hooks (especially when the hooks reside outside of the test file) would aid in debugging. @alycda feel free to open up a separate issue for that ๐Ÿ™‚ Closing this as the behavior is fixed in jest-circus Closing this as the behavior is fixed in jest-circus It doesn't seem to be fixed in latest Jest (v23.6.0). So it's fixed for what users exactly? Facebook internal? See the comment above: https://github.com/facebook/jest/issues/6695#issuecomment-405326998 Install circus (which will become the default in Jest at some point) and you should be good. Also see #7198 Install circus (which will become the default in Jest at some point) Thanks I'll follow. Still why then it's not set as default? Are there any other issues we should be aware of? Is jest-circus considered stable? You can follow #6295 for state of making it the default. jest-circus seems to not fix this. All 3 tests still run and fail. Here's the command I ran: $ yarn run jest --testRunner=jest-circus/runner --bail Am I missing a configuration value somewhere? This is still as broken as ~1yr ago, even with circus... Please consider reopening. To my eyes this just looks like an issue with the reporting. beforeAll is only run once and none of the other tests actually execute, but they all fail with the same reason, and they all print the same error. We should probably avoid printing the individual tests if beforeAll fails. Right now the implementation just checks if we've had an error and if yes, print it and fail the test I've seen a similar issue with reporting. It's hard to disentangle from proprietary code, but what happens is that a test.todo('description', callback); reports the MongoDB error thrown by the previously executed test (a bulk write failure) instead of the correct error that only a description should be specified for .todo tests. I tried throwing an error directly in that failing test but that doesn't reproduce the behavior. Happy to see if the .todo issue reproduces once this is fixed. @SimenB the issue seems to be only partially fixed in circus. since nested hooks are still being executed. To my eyes this just looks like an issue with the reporting. beforeAll is only run once and none of the other tests actually execute, but they all fail with the same reason, and they all print the same error. We should probably avoid printing the individual tests if beforeAll fails. Right now the implementation just checks if we've had an error and if yes, print it and fail the test without actually executing it @SimenB Should we create a separate issue for this? Or was the intention to re-open this to continue the work in this one? Sure, a new issue about the state today and what we want the end state to be makes sense to me Is it a duplicate of #577? @pkuczynski I think so, yes. Still not working as expected... Not via throwing an error or calling done.fail(err) Definitely not stale, still see this issue in latest jest. Undigging topic - imo. test suite should interrupt tests execution with appropriate message when one of hooks will fail or timeout (beforeEach/All, afterEach/All). Unstale As a present for this issue's 5th birthday, I would like to ask a maintainer to reopen #2713, because throwing from beforeAll should really stop executing any further tests. @SimenB? It seems this framework is totally abandonned... RIP!
2025-04-01T06:39:12.247011
2024-08-20T13:53:07
2475719495
{ "authors": [ "Daoortor", "nbirillo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7261", "repo": "jetbrains-academy/kotlin-onboarding-collections", "url": "https://github.com/jetbrains-academy/kotlin-onboarding-collections/pull/70" }
gharchive/pull-request
compile frontend in oldSchool, update task info In #69 there was uncompiled frontend, also task-info.yaml files didn't reflect the new changes. This is fixed @Daoortor could you please fix Qodana?
2025-04-01T06:39:12.262339
2017-09-11T10:15:59
256651489
{ "authors": [ "munnerz" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7262", "repo": "jetstack-experimental/cert-manager", "url": "https://github.com/jetstack-experimental/cert-manager/issues/83" }
gharchive/issue
Supporting auto-generation of CA issuer keypair #79 adds a basic CA issuer that reads a signing keypair from a Secret in the Kubernetes API server in order to issue certificates. For convenience, it may be desirable to support an 'automatically generate a signing keypair' mode. ref #85 #84 Instead of adding a 'generate CA' field to the Issuer, I think we should look at adding a selfsigned certificate issuer and adding an IsCA field to the Certificate resource type (as described in #85 #84) Closing in favour of #84
2025-04-01T06:39:12.265136
2019-05-13T09:11:49
443287523
{ "authors": [ "jonasdkhansen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7263", "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/issues/1669" }
gharchive/issue
Cert manager stuck at attempting to acquire leader lease Describe the bug: Have been running with cert manager for a year now, but suddenly, it is being stuck at "attempting to acquire leader lease", with no error messages. It is working fine on a similar cluster, with the same settings. Environment details:: Kubernetes version (e.g. v1.10.2): 1.11.8-gke.6 Cloud-provider/provisioner (e.g. GKE, kops AWS, etc): GKE cert-manager version (e.g. v0.4.0): v0.6.2 Install method (e.g. helm or static manifests): helm /kind bug After reinstall of helm chart, and some patience, then it worked.
2025-04-01T06:39:12.295897
2018-03-16T16:43:48
305999446
{ "authors": [ "Zetanova", "argusua", "cryptcoin-junkey", "grebois", "infa-ddeore", "mlushpenko", "munnerz", "simonkey007", "urbaniak", "whereisaaron" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7264", "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/issues/398" }
gharchive/issue
Issue certificate using dns01 via route53 stuck on SelfCheck status Is this a BUG REPORT or FEATURE REQUEST?: Uncomment only one, leave it on its own line: /kind bug /kind feature What happened: Trying to configure dns01 route53 provider and it works using staging letsencrypt ClusterIssuer. When changing ClusterIssuer to live, it stuck on status Normal PrepareCertificate 15m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 15m cert-manager-controller Presenting dns-01 challenge for domain auth-service-trunk.gel.net Normal PresentChallenge 15m cert-manager-controller Presenting dns-01 challenge for domain auth.test.gel.tech Normal SelfCheck 14m cert-manager-controller Performing self-check for domain auth-service-trunk.gel.net Normal SelfCheck 14m cert-manager-controller Performing self-check for domain auth.test.geo.tech I've checked route 53 and I see there _acme-challenge. TXT records for both domains. Same live ClusterIssuer works as expected with http01 provider. In log no errors. How to understand what wrong? Can I enable debug mode in some way? What you expected to happen: Successfully issued certificate How to reproduce it (as minimally and precisely as possible): Anything else we need to know?: Environment: Kubernetes version (use kubectl version): Server Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.8", GitCommit:"2f73858c9e6ede659d6828fe5a1862a48034a0fd", GitTreeState:"clean", BuildDate:"2018-02-09T21:23:25Z", GoVersion:"go1.8.3", Compiler:"gc", Platform:"linux/amd64"} Cloud provider or hardware configuration**: AWS Install tools: Installed via helm from https://github.com/kubernetes/charts/tree/master/stable/cert-manager Others: Hi @argusua is your cluster DNS able to resolve external domain names? cert-manager with be trying to fetch the _acme-challenge' using the DNS provided by the cluster to the cert-manager` pod. If that DNS does not allow Internet DNS look-ups, that will fail. If you have this problem, then from k8s 1.9 you can specify custom DNS servers for each Pod, so you could add an Internet DNS server like Google <IP_ADDRESS> to the cert-manager pod. https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pods-dns-config I've noticed the similar issue. Everything was working 4 days ago but when I applied a new certificate resource I got the following error. Error preparing issuer for certificate: error waiting for key to be available for domain cert-manager pod logs I0316 14:00:52.147870 1 server.go:68] Listening on http://<IP_ADDRESS>:9402 I0316 14:00:52.153768 1 leaderelection.go:174] attempting to acquire leader lease... I0316 14:00:52.230488 1 leaderelection.go:184] successfully acquired lease kube-system/cert-manager-controller I0316 14:03:07.518985 1 controller.go:138] clusterissuers controller: syncing item 'letsencrypt-production' I0316 14:03:10.273963 1 helpers.go:122] Setting lastTransitionTime for ClusterIssuer "letsencrypt-production" condition "Ready" to 2018-03-16 14:03:10.273947464 +0000 UTC m=+138.299805182 I0316 14:03:10.279701 1 controller.go:152] clusterissuers controller: Finished processing work item "letsencrypt-production" I0316 14:03:10.279766 1 controller.go:138] clusterissuers controller: syncing item 'letsencrypt-production' I0316 14:03:10.991498 1 controller.go:152] clusterissuers controller: Finished processing work item "letsencrypt-production" I0316 14:04:19.614244 1 controller.go:187] certificates controller: syncing item 'test/test-certificate' I0316 14:04:19.615050 1 sync.go:107] Error checking existing TLS certificate: secret "test-certificate" not found I0316 14:04:19.615205 1 sync.go:209] Preparing certificate with issuer I0316 14:04:19.616105 1 prepare.go:230] Compare "" with "https://acme-v01.api.letsencrypt.org/acme/reg/31265273" I0316 14:08:14.093642 1 helpers.go:165] Setting lastTransitionTime for Certificate "test-certificate" condition "Ready" to 2018-03-16 14:08:14.093626634 +0000 UTC m=+442.119484352 I0316 14:08:14.093733 1 sync.go:213] Error preparing issuer for certificate: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded E0316 14:08:14.110479 1 sync.go:190] [test/test-certificate] Error getting certificate 'test-certificate': secret "test-certificate" not found Certificate apiVersion: certmanager.k8s.io/v1alpha1 kind: Certificate metadata: name: test-certificate namespace: test spec: secretName: test-certificate issuerRef: name: letsencrypt-production kind: ClusterIssuer dnsNames: - test.mydomain.com acme: config: - dns01: provider: route53 domains: - test.mydomain.com Cluster Issuer apiVersion: certmanager.k8s.io/v1alpha1 kind: ClusterIssuer metadata: name: letsencrypt-production spec: acme: server: https://acme-v01.api.letsencrypt.org/directory email<EMAIL_ADDRESS> privateKeySecretRef: name: letsencrypt-production dns01: providers: - name: route53 route53: accessKeyID: AKIAXXXXXXXXXXXXX secretAccessKeySecretRef: name: route53-config key: secret-access-key region: us-east-1 Certificate detailed description Name: test-certificate Namespace: test Labels: <none> Annotations: <none> API Version: certmanager.k8s.io/v1alpha1 Kind: Certificate Metadata: Cluster Name: Creation Timestamp: 2018-03-16T14:08:31Z Generation: 0 Resource Version: 33688876 Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/test/certificates/test-certificate UID: 9rf2b603-98uj-11e8-8521-1239c7046d88 Spec: Acme: Config: Dns 01: Provider: route53 Domains: test.mydomain.com Common Name: Dns Names: test.mydomain.com Issuer Ref: Kind: ClusterIssuer Name: letsencrypt-production Secret Name: test-certificate Status: Acme: Authorizations: <nil> Conditions: Last Transition Time: 2018-03-16T14:19:33Z Message: Error checking ACME domain validation: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded Reason: ErrCheckAuthorization Status: False Type: Ready Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning ErrorPrepareCertificate 26m (x29 over 5h) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded Warning ErrorCheckCertificate 10m (x30 over 5h) cert-manager-controller Error checking existing TLS certificate: secret "test-certificate" not found Normal PrepareCertificate 10m (x30 over 5h) cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 10m (x30 over 5h) cert-manager-controller Presenting dns-01 challenge for domain test.mydomain.com Normal SelfCheck 9m (x30 over 5h) cert-manager-controller Performing self-check for domain test.mydomain.com TXT record was created in Route53 hosted zone and it was publicly accessible cert-manager: v0.2.3 kubernetes: v1.7.9 cloud provider: AWS Hi @whereisaaron Thanks for fast response. Looks like yes, dns resolution works inside cert-manager container: Also, today status changed for certificate to: Status: Acme: Authorizations: <nil> Conditions: Last Transition Time: 2018-03-16T15:12:40Z Message: Error checking ACME domain validation: [error waiting for authorization for domain "auth.test.gel.tech": context canceled, error waiting for authorization for domain "auth-service-trunk.gel.net": context canceled] Reason: ErrCheckAuthorization Status: False Type: Ready Same here: $ kubectl get clusterissuer -oyaml apiVersion: v1 items: - apiVersion: certmanager.k8s.io/v1alpha1 kind: ClusterIssuer metadata: clusterName: "" creationTimestamp: 2018-03-18T11:00:27Z generation: 0 name: letsencrypt-prod namespace: "" resourceVersion: "3499055" selfLink: /apis/certmanager.k8s.io/v1alpha1/letsencrypt-prod uid: 90d32342-2a9b-11e8-ae12-4ccc6ad2499b spec: acme: dns01: providers: - name: aws route53: accessKeyID: AK232AHSKGRLPPABQ hostedZoneID: "" region: eu-central-1 secretAccessKeySecretRef: key: secret-access-key name: route53-credentials-secret email<EMAIL_ADDRESS> privateKeySecretRef: key: "" name: ingress-key-pair server: https://acme-staging.api.letsencrypt.org/directory status: acme: uri: https://acme-staging.api.letsencrypt.org/acme/reg/232323 conditions: - lastTransitionTime: 2018-03-18T11:00:28Z message: The ACME account was registered with the ACME server reason: ACMEAccountRegistered status: "True" type: Ready kind: List metadata: resourceVersion: "" selfLink: "" kubectl describe certificate Name: rig2-tls Namespace: default Labels: <none> Annotations: <none> API Version: certmanager.k8s.io/v1alpha1 Kind: Certificate Metadata: Cluster Name: Creation Timestamp: 2018-03-18T11:02:58Z Generation: 0 Owner References: API Version: extensions/v1beta1 Block Owner Deletion: true Controller: true Kind: Ingress Name: rig2-zcash-miner UID: 95c1ed1c-2a2a-11e8-ae12-4ccc6ad2499b Resource Version: 3499961 Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/default/certificates/rig2-tls UID: eb1acec6-2a9b-11e8-ae12-4ccc6ad2499b Spec: Acme: Config: Dns 01: Provider: aws Domains: rigs.umine.farm Common Name: Dns Names: rigs.umine.farm Issuer Ref: Kind: ClusterIssuer Name: letsencrypt-prod Secret Name: rig2-tls Status: Acme: Authorizations: <nil> Conditions: Last Transition Time: 2018-03-18T11:06:38Z Message: Error checking ACME domain validation: error waiting for key to be available for domain "rigs.umine.farm": context deadline exceeded Reason: ErrCheckAuthorization Status: False Type: Ready Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal SelfCheck 7m (x5 over 21m) cert-manager-controller Performing self-check for domain rigs.umine.farm Warning ErrorCheckCertificate 4m (x6 over 22m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-tls" not found Normal PrepareCertificate 4m (x6 over 22m) cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 4m (x6 over 22m) cert-manager-controller Presenting dns-01 challenge for domain rigs.umine.farm Warning ErrorPrepareCertificate 39s (x6 over 18m) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "rigs.umine.farm": context deadline exceeded and route53 record is created and working well My ingress: ingress: enabled: true annotations: #{} kubernetes.io/ingress.class: nginx nginx.ingress.kubernetes.io/rewrite-target: '/' kubernetes.io/tls-acme: "true" certmanager.k8s.io/acme-challenge-type: 'dns01' certmanager.k8s.io/acme-dns01-provider: 'aws' # nginx.ingress.kubernetes.io/ssl-redirect: "false" path: /rig2 hosts: - rigs.umine.farm tls: #[] - secretName: rig2-tls hosts: - rigs.umine.farm @whereisaaron thanks for the tip regarding dns. I can confirm where I exec into cert manager and try pining that address it fails. External addresses are ok, but I configured my cluster woith the same name as dns address, so I guess that's why the query doesn't go to upstream server which is <IP_ADDRESS> Not fully fixed. Got it working for rigs.umine.farm and got fake certificate from staging acme server. Changed server to prod and was hoping to get a new certificate (proper one) but that didn't work. Changed DNS to another name and started getting errors during self-check again. Warning ErrorCheckCertificate 17m (x2 over 20m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-cert" not found Normal PrepareCertificate 17m (x2 over 20m) cert-manager-controller Preparing certificate with issuer Warning ErrorPrepareCertificate 17m cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "monitor.umine.farm": context deadline exceeded Normal PresentChallenge 17m (x2 over 20m) cert-manager-controller Presenting dns-01 challenge for domain monitor.umine.farm Normal SelfCheck 16m (x2 over 19m) cert-manager-controller Performing self-check for domain monitor.umine.farm Warning ErrorCheckCertificate 1m (x5 over 15m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-cert" not found Normal PrepareCertificate 1m (x5 over 15m) cert-manager-controller Preparing certificate with issuer Warning ErrorPrepareCertificate 1m (x4 over 11m) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "monitor.umine.farm": context deadline exceeded Normal PresentChallenge 1m (x5 over 15m) cert-manager-controller Presenting dns-01 challenge for domain monitor.umine.farm Normal SelfCheck 1m (x5 over 14m) cert-manager-controller Performing self-check for domain monitor.umine.farm 'context deadline exceeded' means the HTTP request timed out. With the range of stuff you have reported my guess would be you've having some connectivity issues with Internet requests from your cluster. It looks like your cert-manager is fine, but it has, perhaps intermittent, trouble making DNS requests and making HTTP requests to the Internet to check the challenges are in place. Looks like DNS and HTTP request are correct Request performed from the cert-manager container: dig acme-v01.api.letsencrypt.org ; <<>> DiG 9.11.2-P1 <<>> acme-v01.api.letsencrypt.org ;; global options: +cmd ;; Got answer: ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 48485 ;; flags: qr rd ra; QUERY: 1, ANSWER: 3, AUTHORITY: 0, ADDITIONAL: 1 ;; OPT PSEUDOSECTION: ; EDNS: version: 0, flags:; udp: 4096 ;; QUESTION SECTION: ;acme-v01.api.letsencrypt.org. IN A ;; ANSWER SECTION: acme-v01.api.letsencrypt.org. 60 IN CNAME api.letsencrypt.org-ng.edgekey.net. api.letsencrypt.org-ng.edgekey.net. 60 IN CNAME e14990.dscx.akamaiedge.net. e14990.dscx.akamaiedge.net. 20 IN A <IP_ADDRESS> ;; Query time: 6 msec ;; SERVER: <IP_ADDRESS>#53(<IP_ADDRESS>) ;; WHEN: Mon Mar 19 10:15:00 UTC 2018 ;; MSG SIZE rcvd: 158 curl https://acme-v01.api.letsencrypt.org/directory { "key-change": "https://acme-v01.api.letsencrypt.org/acme/key-change", "meta": { "terms-of-service": "https://letsencrypt.org/documents/LE-SA-v1.2-November-15-2017.pdf" }, "new-authz": "https://acme-v01.api.letsencrypt.org/acme/new-authz", "new-cert": "https://acme-v01.api.letsencrypt.org/acme/new-cert", "new-reg": "https://acme-v01.api.letsencrypt.org/acme/new-reg", "rYI1xLOBxqk": "https://community.letsencrypt.org/t/adding-random-entries-to-the-directory/33417", "revoke-cert": "https://acme-v01.api.letsencrypt.org/acme/revoke-cert" } As I said before, nothing has been changed in our k8s cluster or cert-manager configuration and everything was working a week ago. @simonkey007 cert-manager isn't trying to access letsencrypt.org, it is trying to access the domain for the certificate being issued, to check that the challenge record is in place. e.g. 'umine.farm'. I've tried today create test ingress and it works as expected. apiVersion: extensions/v1beta1 kind: Ingress metadata: name: lets-test annotations: kubernetes.io/ingress.class: "nginx" kubernetes.io/tls-acme: "true" certmanager.k8s.io/cluster-issuer: "letsencrypt-live" certmanager.k8s.io/acme-challenge-type: "dns01" certmanager.k8s.io/acme-dns01-provider: "route53" spec: tls: - secretName: test-crt-tls hosts: - lets.test.gelato.tech - letstest.gelatogroup.net rules: - host: lets.test.gelato.tech http: paths: - path: / backend: serviceName: jenkins servicePort: 80 - host: letstest.gelatogroup.net http: paths: - path: / backend: serviceName: jenkins servicePort: 80 For live ingress apiVersion: extensions/v1beta1 kind: Ingress metadata: annotations: certmanager.k8s.io/acme-challenge-type: dns01 certmanager.k8s.io/acme-dns01-provider: route53 certmanager.k8s.io/cluster-issuer: letsencrypt-live kubernetes.io/ingress.class: nginx kubernetes.io/tls-acme: "true" creationTimestamp: 2018-03-16T07:19:07Z generation: 269 name: auth-service-kubernetes-live-nginx namespace: default resourceVersion: "119221337" selfLink: /apis/extensions/v1beta1/namespaces/default/ingresses/auth-service-kubernetes-live-nginx uid: 5068e65a-28ea-11e8-94fa-0ae1ba4584f4 spec: rules: - host: auth.live.gelato.tech http: paths: - backend: serviceName: auth-service-kubernetes-live-nginx servicePort: 80 path: / - host: auth-service.gelatogroup.com http: paths: - backend: serviceName: auth-service-kubernetes-live-nginx servicePort: 80 path: / tls: - hosts: - auth.live.gelato.tech - auth-service.gelatogroup.com secretName: auth-service-kubernetes-live-nginx-tls I get next certificate like this: Name: auth-service-kubernetes-live-nginx-tls Namespace: default Labels: <none> Annotations: <none> API Version: certmanager.k8s.io/v1alpha1 Kind: Certificate Metadata: Cluster Name: Creation Timestamp: 2018-03-20T14:23:59Z Deletion Grace Period Seconds: <nil> Deletion Timestamp: <nil> Owner References: API Version: extensions/v1beta1 Block Owner Deletion: true Controller: true Kind: Ingress Name: auth-service-kubernetes-live-nginx UID: 5068e65a-28ea-11e8-94fa-0ae1ba4584f4 Resource Version: 119219236 Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/default/certificates/auth-service-kubernetes-live-nginx-tls UID: 54dba49f-2c4a-11e8-94fa-0ae1ba4584f4 Spec: Acme: Config: Dns 01: Provider: route53 Domains: auth.live.gelato.tech auth-service.gelatogroup.com Common Name: Dns Names: auth.live.gelato.tech auth-service.gelatogroup.com Issuer Ref: Kind: ClusterIssuer Name: letsencrypt-live Secret Name: auth-service-kubernetes-live-nginx-tls Status: Conditions: <nil> Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning ErrorCheckCertificate 21m cert-manager-controller Error checking existing TLS certificate: secret "auth-service-kubernetes-live-nginx-tls" not found Normal PrepareCertificate 21m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 21m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 20m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Warning ErrorCheckCertificate 9m cert-manager-controller Error checking existing TLS certificate: secret "auth-service-kubernetes-live-nginx-tls" not found Normal PrepareCertificate 9m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 9m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Normal PresentChallenge 9m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 9m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Additionally, I found today in logs next error: E0320 14:43:49.216953 1 prepare.go:167] Error cleaning up solver: Failed to change Route 53 record set: InvalidChangeBatch: Tried to delete resource record set [name='_acme-challenge.auth-service.gelatogroup.com.', type='TXT'] but it was not found status code: 400, request id: 19c8e4a0-2c4d-11e8-99b7-430c9c1f3614 I've tried create this record manually, but error again apeared. @argusua I don't think you need to create it manually as it provides a response with a long non-readable string when you query it (run test record in route53). And I am just curious - can single secret work with multiple hosts? I had an idea of one secret/certificate per host but maybe it works with several as well (I am not talking about wildcard certificates as your domain names are different.) I've created new clusterIssuer (previous I get by migrating from kube-lego as described here https://github.com/jetstack/cert-manager/blob/master/docs/user-guides/migrating-from-kube-lego.md). Now I get a certificate but only for first domain: Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal PrepareCertificate 4m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 4m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Normal PresentChallenge 4m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 3m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Normal PrepareCertificate 2m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 2m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Normal PresentChallenge 2m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 1m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Normal ObtainAuthorization 49s cert-manager-controller Obtained authorization for domain auth.live.gelato.tech Warning ErrorPrepareCertificate 3s cert-manager-controller Error preparing issuer for certificate: error presenting acme authorization for domain "auth-service.gelatogroup.com": Failed to change Route 53 record set: InvalidChangeBatch: RRSet with DNS name _acme-challenge.auth-service.gelatogroup.com. is not permitted in zone sourcelogistic.com. status code: 400, request id: 3ae889c0-2eb0-11e8-952e-ad6827c36e4f Normal RenewalScheduled 3s cert-manager-controller Certificate scheduled for renewal in 1438 hours Normal PrepareCertificate 2s (x2 over 3s) cert-manager-controller Preparing certificate with issuer Warning ErrorPrepareCertificate 2s cert-manager-controller Error preparing issuer for certificate: error presenting acme authorization for domain "auth-service.gelatogroup.com": Failed to change Route 53 record set: InvalidChangeBatch: RRSet with DNS name _acme-challenge.auth-service.gelatogroup.com. is not permitted in zone sourcelogistic.com. status code: 400, request id: 985b9304-2eb0-11e8-8ec5-9feeb1e51a10 Normal RenewalScheduled 2s cert-manager-controller Certificate scheduled for renewal in 1438 hours Normal PresentChallenge 1s (x2 over 3s) cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Looks like cert-manager uses incrorect zoneId for set dns record in route53. Finally, I found why the wrong zoneId had used. It's because of CNAME record. My current issue resolved. I had/have the same issue with http01 it seams, where the cert-manager cant curl the well-known url from inside the cluster, but externaly it is possible to access. Just the edge router missing some loopback ... spend 15h for this knowledge. The error description "error waiting for key to be available for domain" is very missleading. Should be "error waiting for collange-key resource to be available for domain ..." @Zetanova often referred to as 'hairpin NAT', needed if your egress and ingress have the same external IP address. It can be confusing to configure. @whereisaaron Yes, i disovered it too. My problem was that i just missundestud the error message. I put active/passive vm iptabales+keepalived in front of the k8s and it seams that i cant get the hairpin NAT to work. I got a similar issue on dns01 config. Log was like follows. (As @whereisaaron says, this should be shown only http01 config...) $ kubectl describe certificate ... Events: ... Warning ErrorPrepareCertificate 2m cert-manager-controller Error preparing issuer for certificate: [error waiting for key to be available for domain "{censored}": context canceled] ... In my case, it was fixed by just deleting (and restarting) the cert-manager pod. I'm going to close this now as it appears to be resolved - thanks everyone for providing support and help debugging ๐Ÿ˜„ Just tried to generate letsencrypt's wildcard cert via dns01/route53 validation and getting: E0621 14:13:18.291341 1 controller.go:186] certificates controller: Re-queuing item "namespace/xxx-net" due to error processing: Failed to change Route 53 record set: InvalidChangeBatch: Tried to delete resource record set [name='_acme-challenge.xxx.net.', type='TXT'] but it was not found Looks like it should just silently ignore that error or don't delete entries which don't exists in route53? Did you manage to resolve that issue? Using the latest cert manager v0.3.0. same problem with 0.5.0, was working fine in 0.4.1, rolling back we had same issue, it got resolved when I created the TXT record manually, then acme cert manager changed that record because it was expired and deleted successfully. I got the txt record key from cert manager logs and the value from aws cloudtrail logs. Next time I will try creating the txt record with some random value to see if that helps
2025-04-01T06:39:12.300245
2022-10-13T14:27:24
1407905390
{ "authors": [ "charlieegan3" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7265", "repo": "jetstack/jsctl", "url": "https://github.com/jetstack/jsctl/pull/38" }
gharchive/pull-request
Prompt user to check supported versions if unknown Before: $ go run main.go operator deploy --version=foo operator version foo does not exist exit status 1 After: $ jsctl operator versions v0.0.1-alpha.17 $ go run main.go operator deploy --version=foo operator version foo is unknown or not supported by this version of jsctl. Run 'jsctl operator versions' to see the supported operator versions. exit status 1 Signed-off-by: Charlie Egan<EMAIL_ADDRESS> Thanks Irbe ๐Ÿ™
2025-04-01T06:39:12.306997
2018-03-20T17:13:48
306959545
{ "authors": [ "munnerz", "wallrj" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7266", "repo": "jetstack/navigator", "url": "https://github.com/jetstack/navigator/pull/291" }
gharchive/pull-request
Use new test fixture in Cassandra controls I'm trying to get rid of the Cassandra specific test fixture in ./pkg/controllers/cassandra/testing In this branch I've switched: ServiceAccounts Roles RoleBindings I'll deal with Pilot and NodePool controls in separate branches. Release note: NONE /test e2e /test e2e v1.10 /retest /lgtm /approve
2025-04-01T06:39:12.347868
2015-04-09T03:28:10
67272239
{ "authors": [ "joshowen", "ryanhiebert", "syphar", "touilleMan" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7267", "repo": "jezdez/hirefire", "url": "https://github.com/jezdez/hirefire/issues/8" }
gharchive/issue
Celery doesn't account for running and reserved tasks This is a similar problem to #3 . I'm using the Redis backend for Celery. I tried my hand at a solution in https://github.com/aspiredu/hirefire/commit/3ae625cc527330357a66df8d8c5e2701213d5af7, but it only works for small numbers of dynos, workers, and tasks. Even when it does work, it's painfully slow. The issue is that active, reserved, and scheduled tasks don't show up in the queue. Missing the scheduled ones is fine, but missing active and reserved means that my queues, which are running a lot of long-running tasks, get shut down well before stuff is done processing, which causes the dynos to flap as they are put back in the queue, then taken out, then when the process repeats. Unfortunately I can confirm that #9 has not solved this issue, even when I'm using py-ampq. I haven't tested with librabbitmq or iron-mq, but I find it likely that I'll get the same results. An IRC conversation with @ask gave me two ideas to try. Both are somewhat involved for something as light as HireFire, but we've gotta do what we've gotta do. Use Flower to get data on the running worker queues. Use heartbeats from worker to get up-to-date data. (from me) use inspect() (that's how flower gets it's info) for all queues/procs/workers at once and somehow re-use that data for the life of the request to the HireFire info URL. 2 requires some upstream code-changes to Celery (the heartbeats don't currently include scheduled and reserved counts), and also requires a persistent monitor to receive those events. I'm honestly not sure which path to take first. I'll just have to take 1 and run with it. @ryanhiebert did you find any solution for this trouble ? No solution yet. It's been somewhat back-burnered in my current workload, but trying to use Flower to get a better picture of the running queues seems like the route I'm likely to try first. Any and all help would be welcome! Any updates on this issue? No, sorry. Haven't been able to work on it any more yet. I wrote a basic implementation that seems to work. import itertools from celery import Celery from django.conf import settings from gevent.pool import Group from hirefire.procs.celery import CeleryProc def _inspect(method): app = Celery('app', broker=settings.BROKER_URL) inspect_result = getattr(app.control.inspect(), method)() app.close() return inspect_result class OFCeleryProc(CeleryProc): def quantity(self): """ Returns the aggregated number of tasks of the proc queues. """ if hasattr(self.channel, '_size'): # Redis return sum(self.channel._size(queue) for queue in self.queues) # AMQP (Include both libs, since we have both installed) from librabbitmq import ChannelError as LCE from amqp.exceptions import ChannelError as ACE count = 0 for queue in self.queues: try: queue = self.channel.queue_declare(queue, passive=True) except (LCE, ACE): # The requested queue has not been created yet pass else: count += queue.message_count active, reserved = Group().map(_inspect, ('active', 'reserved')) num_active = sum(1 for v in itertools.chain.from_iterable(active.itervalues()) if not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues) if active else 0 num_reserved = sum(1 for v in itertools.chain.from_iterable(reserved.itervalues()) if not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues) if reserved else 0 return count + num_active + num_reserved @joshowen : That can work OK, for a small number of queues. Unfortunately, I have around 20 queues, and the combined time of running this made for timeouts in Heroku. I've got to find something faster. https://github.com/celery/celery/issues/2839 This issue with Celery also makes this issue worse. Having this fixed wouldn't be sufficient by itself, but it would make the problem less of a deal for short tasks. @ryanhiebert Even using gevent? @joshowen: Actually, that's not what I had done before. I'm going to have to dig more into it, and see if it works for me. I'm not currently using gevent. With gevent I was able to run the _inspect calls in parallel and got around the heroku timeouts Let me try it. I still do have quite a few queues (not just workers), with different procs, so even with gevent it means that I'll have to run ~20 of these to get all the stats for HireFire. Worth a try, though. @joshowen thanks for your example code, we just ran into the same issue. I build a similar solution, still slow, but less calls: the app.control.inspect() calls always return the data for every queue, not only one. We have just extended HireFireMiddleware to do these calls once (not only for active and reserved, but also for scheduled) and then the subsequent CeleryProcs will get the data from their queues out of it. So queue size does not have any effect on the calls, perhaps only on the time celery needs to call all the workers. (cc @ryanhiebert ) will try to bake this into a PR, I think this is a common problem if you want to scale to 0 :) @syphar: that sounds awesome. I'm looking forward to it. @syphar: I'm working on this now for work, and if possible I'd like to see your approach. Any chance you could let me see your work (via very raw, perhaps broken PR, or perhaps a gist of the interesting code), so that I can take inspiration at least? @ryanhiebert yes, sure. Sorry for the delay many things happening :) middleware: class HireFireMiddleware(OriginalHireFireMiddleware): """ An extended HireFireMiddleware which includes counts worker-stats. we query for running, reserved or scheduled tasks and add them to the queued tasks per queue. The related procs have to have a method called ``quantity_with_running`` which will get the inspection data, so it's queried only once per call from hirefire. """ def dump_procs(self, procs): celery_inspect = { method: getattr(app.control.inspect(), method)() for method in ('active', 'reserved', 'scheduled') } data = [ { 'name': name, 'quantity': proc.quantity_with_running(celery_inspect) or 'null', } for name, proc in procs.items() ] return json.dumps(data, cls=TimeAwareJSONEncoder, ensure_ascii=False) def info(self, request): payload = self.dump_procs(self.loaded_procs) return HttpResponse(payload, content_type='application/json') procs.py class UpdatedCeleryProc(CeleryProc): def quantity_with_running(self, celery_inspect): count = super(UpdatedCeleryProc, self).quantity() for func, tasks in celery_inspect.items(): if not tasks: continue queue_tasks = [ v for v in itertools.chain.from_iterable(tasks.values()) if ( not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues ) ] count += len(queue_tasks) return count class LowPrioWorkerProc(UpdatedCeleryProc): name = 'celeryworkerlowprio' queues = ['lowprio'] this code as expected: only works with celery not perfect: increases response time (it's around ~3s on our production system) good: the amount of procs/queues doesn't matter, since the inspect result contains everything. runs in production for 4 weeks here only tested on python 3.5 What was blocking me (when I had some time) was the question how to design this the best way to make it optional make it backend-agnostic Awesome. I'll work on figuring out those when I'm implementing it for myself. This is very, very, helpful. Thank you. @ryanhiebert we can get this into a working PR together, if you have some nice ideas on the design :) Would you be up for a pairing session, perhaps? You can hit me up on irc.freenode.net, I'm ryanhiebert, or you can email me<EMAIL_ADDRESS> If you're interested in pairing, I've set up a floobits workspace at https://floobits.com/ryanhiebert/hirefire. I do have a good idea, and that's to send a cache dictionary into the procs, so that they can share state between themselves. I'm working up an implementation. Here's my work-in-progress branch on the approach I came up with. Feel free to give me feedback. I'll open up a PR after I've done some more testing in my environment. https://github.com/aspiredu/hirefire/tree/global-cache
2025-04-01T06:39:12.351914
2022-05-28T23:44:27
1251746594
{ "authors": [ "jf2048", "thiblahute" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7268", "repo": "jf2048/gobject", "url": "https://github.com/jf2048/gobject/issues/25" }
gharchive/issue
Support gstreamer Need to write an example using the macro with an gstreamer element, and possibly create another gst_element macro to make it easier. I implemented a webrtcsrc element based on this crate, I think we have mostly everything we need tbh. I also started adding macros to ease GstElement implementation here, at this point it implements element metadata and pad templates declaration in the macros like: #[gobject::gst_element( class(final), long_name = "TheTestElement", classification = "Test/Filter", description = "Just a test", author = "Thibault Saunier<EMAIL_ADDRESS> pad_templates( src(presence="always"), // `__` is transformed to `_%` as "%" is not a valid character sink__u(direction="sink", presence="sometimes", caps="video/x-raw"), ) )] My next step will be to improve the way we interact with Pads creation and integration into the Elements, maybe adding a macros like: #[gobject::gst_pad(sink__u)] impl SrcPad { fn chain() -> Result<gst::FlowSuccess, gst::FlowError> { ... } } but I am still thinking.
2025-04-01T06:39:12.374180
2024-05-17T07:08:50
2301986153
{ "authors": [ "Soumya6Tiwari", "jfmartinz", "renunegii" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7269", "repo": "jfmartinz/ResourceHub", "url": "https://github.com/jfmartinz/ResourceHub/issues/276" }
gharchive/issue
Creation of a roadmap section Idea Contribution [X] I have read all the feature request issues. [X] I'm interested in working on this issue [X] I'm part of GSSOC organization Explain feature request Creation of a roadmap section,which will contact roadmap and their respected study material and links for various fields, Android development,web development ,ml , DSA etc Explain your solution No response Any alternative approaches/features No response Additional Context No response @jfmartinz kindly assign me this issue @jfmartinz Maybe this issue is not solved yet! I would like to resolve this issue ASAP, creating a roadmap section for almost every tech field- Data Science, Web Development, App Development, UI/UX Design, and Machine Learning... Kindly assign me this issue, and I'll resolve it soon. This is actually completed but the author forgot to reference the PR. but you can still add some resources that youd like and create a PR for that, just make sure that those resources are not duplicate, thanks @jfmartinz ok then I will do so as a separate pr
2025-04-01T06:39:12.384337
2021-05-27T15:42:43
903925621
{ "authors": [ "AnalogJ", "eyalbe4" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7270", "repo": "jfrog/jfrog-client-go", "url": "https://github.com/jfrog/jfrog-client-go/pull/357" }
gharchive/pull-request
make sure statuses are included with the buildinfo. [ ] All tests passed. If this feature is not already covered by the tests, I added new tests. [x] This pull request is on the dev branch. [x] I used gofmt for formatting the code before submitting the pull request. fixes #356 @AnalogJ, Since we're currently considering to make major changes and improvements to the entire build promotion process, I suggest that for now, we wait with changes to the build-info schema to include promotion statuses. Hi @eyalbe4 I'd really like to get this merged if possible. Any chance you could provide an update about the future of Promotion status's in this library? I've also rebased my branch on the latest from master. Hey @eyalbe4 @yahavi Any additional feedback for this PR? @eyalbe4 @yahavi any updates here? Sorry to keep bugging you @AnalogJ, Since this change may conflict with other planned changes to the build promotion process, will you be able to share more details about the need? How are you planning to use and consume this data? With this information, maybe we'll be able to come up with an alternative solution. Hey, This is a critical feature for us. Basically we use the Jfrog Go SDK to create "Builds". The builds contain a list of versioned artifacts that will be used in a deployment. These artifacts are then scanned using XRAY, and deployed to various non-production environments automatically. The "PromotionStatus" field is used as a manual flag to ensure that Production Deployments can only use vetted/production-ready artifacts. I see @AnalogJ. Would you be willing perhaps to tag the build-info artifact in Artifactory, with a property that will indicate the status of the build? Every time you publish a build-info, it is also saved as a file under the artifactory-build-info repository. You can set and get the value of this custom property on/from the build-info artifact, to support your logic. This functionality is supported by this library as well. See the Setting Properties on Files in Artifactory and Deleting Properties from Files in Artifactory sections in the README. This PR doesnt seem relevant since https://github.com/jfrog/jfrog-client-go/blob/master/artifactory/services/promote.go is merged. Closing.
2025-04-01T06:39:12.389830
2022-07-27T19:57:56
1320061483
{ "authors": [ "alexhung", "apr-1985" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7271", "repo": "jfrog/terraform-provider-artifactory", "url": "https://github.com/jfrog/terraform-provider-artifactory/issues/518" }
gharchive/issue
Some defaults in the user HCL dont match the description Describe the bug profile_updatable and disable_ui_access default to True when the docs say they should be False Requirements for and issue [x] A description of the bug [ ] A fully functioning terraform snippet that can be copy&pasted (no outside files or ENV vars unless that's part of the issue). If this is not supplied, this issue will likely be closed without any effort expended. [ ] Your version of artifactory (you can curl it at $host/artifactory/api/system/version [ ] Your version of terraform [ ] Your version of terraform provider Expected behavior Docs match actuals Additional context Add any other context about the problem here. https://github.com/jfrog/terraform-provider-artifactory/pull/517 Actually docs here say it should be true https://github.com/jfrog/terraform-provider-artifactory/blob/master/docs/resources/user.md not sure what is the correct default value and which bit is wrong :) @apr-1985 Per my comment, it's the Description that is incorrect.
2025-04-01T06:39:12.392383
2014-10-07T15:05:19
45118362
{ "authors": [ "3flex", "buckstr", "wyardley" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7272", "repo": "jfryman/puppet-nginx", "url": "https://github.com/jfryman/puppet-nginx/pull/471" }
gharchive/pull-request
issue 470 allow creation of location with only try_files This commit will fix #470. @buckstr the proposed PR won't allow you to create a location with only try_files. As is, if you only set try_files and don't set proxy, location_alias, stub_status, fastcgi or www_root then the template vhost/locations/empty.erb will be rendered. This doesn't include try_files. Hi -- There's been no further activity on this since the 'needs Rebase' and 'needs tests' labels were set. If we don't have any further comments in the next month, this PR will be closed. Addressed in #834 already.
2025-04-01T06:39:12.406462
2019-08-04T20:27:50
476590766
{ "authors": [ "gwenn", "kornelski" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7273", "repo": "jgallagher/rusqlite", "url": "https://github.com/jgallagher/rusqlite/pull/556" }
gharchive/pull-request
Conversion from FromSqlError to Error Error needs column index, but FromSqlError doesn't have it. I've fudged it by putting obviously-wrong index value, and then hiding that fact in Display. Ideally the Error should use Option<usize> for index (and type), but I assume it's better to avoid unnecessary backwards-compat break. Fixes #555 The test failure looks like a vcpkg installation issue, unlikely to be related to this change. Thank you.
2025-04-01T06:39:12.407982
2014-12-20T20:00:55
52574994
{ "authors": [ "GinkgoFJG", "akaRem", "benplum" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7274", "repo": "jgallen23/routie", "url": "https://github.com/jgallen23/routie/issues/35" }
gharchive/issue
feature request: routie.back(), routie.next() it would be great to have back and next method for navigating through history What's wrong with using window.history.back() and window.history.forward()? These should work out of the box. I haven't dug into why, but in cordova window.history.back() and window.history.forward() don't seem to work -- at least not on the emulated Nexus 4.
2025-04-01T06:39:12.457891
2016-02-02T21:02:15
130815186
{ "authors": [ "jglamine", "tylervz" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7275", "repo": "jglamine/calvinwebprint", "url": "https://github.com/jglamine/calvinwebprint/pull/49" }
gharchive/pull-request
Display an error message when file upload fails A common error that has been occurring is the failure of the /api/upload endpoint. While I'd like to get around to finding out why the error is happening so frequently, the least we can do is show an error message. Here is the console output of when it happened to me twice today: Looks good to me. See my one comment. It could be that disk space on the server is getting low. It might be worth looking into. I think that might be it. Disk space, I mean. Hiding the alert on upload is a good idea. I'll update the PR. On Feb 2, 2016 7:40 PM, "James Lamine"<EMAIL_ADDRESS>wrote: Looks good to me. See my one comment. It could be that disk space on the server is getting low. It might be worth looking into. โ€” Reply to this email directly or view it on GitHub https://github.com/jglamine/calvinwebprint/pull/49#issuecomment-178917875 .
2025-04-01T06:39:12.496373
2023-01-27T16:20:50
1560036992
{ "authors": [ "OverFloyd", "davidjgraph" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7276", "repo": "jgraph/drawio", "url": "https://github.com/jgraph/drawio/issues/3319" }
gharchive/issue
Bring "Convert labels to SVG" export option to Desktop version + conversion issue [ ] I agree to follow the Code of Conduct that this project adheres to. [ ] I have searched the issue tracker for a feature request that matches the one I want to file, without success. So, I've this big file I'm working on that has a lot (really a lot) of text labels and I'll have it to upload it to Wikimedia Commons. Since Commons doesn't like certain tags about text labels I have to convert that into vectorized text, feature that is only available on the web version. I'd really like to have it on Deskto p so that I wouldn't have to switch working environment every time I need to export my drawings. Second, since this file I'm working on is as I said pretty text intensive, the converter struggles to convert all the text and ends with an error. I'll upload the file for context so that you can replicate the issue. Carrara-Avenza rev. 3 (2023-01-23).zip Cheers! convert to SVG requires a back-end, it can't be ported to desktop.
2025-04-01T06:39:12.504075
2018-03-16T15:42:24
305977313
{ "authors": [ "jgwhite", "joe-wroten-q2" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7277", "repo": "jgwhite/ember-sortable", "url": "https://github.com/jgwhite/ember-sortable/issues/178" }
gharchive/issue
The future is ember-animated I propose to either sunset this addon and give people a transition path to ember-animated or perhaps reimagine it as a thin wrapper atop ember-animated. Thoughts welcome! ember-sortable has accomplished exactly what I attempted to create in an app I'm working on in a reliable way. I'd hate to see it go, but if the plan is to sunset the addon then I'd be happy to replace what we have with a thin wrapper atop of ember-animated should that be created. I still think this is worth exploring, but itโ€™s a long way off so closing for now.
2025-04-01T06:39:12.520004
2021-03-19T09:28:00
835750943
{ "authors": [ "avdev4j", "dieppa" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7278", "repo": "jhipster/generator-jhipster-quarkus", "url": "https://github.com/jhipster/generator-jhipster-quarkus/issues/179" }
gharchive/issue
Mongock is not compatible with the native build Describe the bug Mongock is a library to help us to handle the data migration with a MongoDb database. JHipster already uses it with Mongo and we reused the same with Quarkus. Mongock uses org.reflections that cause an error when we build it with GraalVM to generate a native executable. To Reproduce Steps to reproduce the behavior: Generate a project with MongoDb as a database. Install GraalVM openjdk 11.0.10 2021-01-19 OpenJDK Runtime Environment GraalVM CE <IP_ADDRESS> (build 11.0.10+8-jvmci-21.0-b06) OpenJDK 64-Bit Server VM GraalVM CE <IP_ADDRESS> (build 11.0.10+8-jvmci-21.0-b06, mixed mode, sharing) Run the following command: ./mvnw clean verify -Pnative -DskipTests Caused by: com.oracle.graal.pointsto.constraints.UnsupportedFeatureException: Unsupported features in 2 methods Detailed message: Error: com.oracle.graal.pointsto.constraints.UnresolvedElementException: Discovered unresolved type during parsing: org.apache.commons.vfs2.VFS. To diagnose the issue you can use the --allow-incomplete-classpath option. The missing type is then reported at run time when it is accessed the first time. Trace: at parsing org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(Vfs.java:287) Call path from entry point to org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(URL): at org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(Vfs.java:287) at org.reflections.vfs.Vfs.fromURL(Vfs.java:98) at org.reflections.vfs.Vfs.fromURL(Vfs.java:90) at org.reflections.Reflections.scan(Reflections.java:236) at org.reflections.Reflections$1.run(Reflections.java:199) at java.lang.Thread.run(Thread.java:834) at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:519) at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192) at com.oracle.svm.core.code.IsolateEnterStub.PosixJavaThreads_pthreadStartRoutine_e1f4a8c0039f8337338252cd8734f63a79b5e3df(generated:0) Error: com.oracle.graal.pointsto.constraints.UnresolvedElementException: Discovered unresolved type during parsing: org.apache.commons.vfs2.VFS. To diagnose the issue you can use the --allow-incomplete-classpath option. The missing type is then reported at run time when it is accessed the first time. Trace: at parsing org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(Vfs.java:281) Call path from entry point to org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(URL): at org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(Vfs.java:281) at org.reflections.vfs.Vfs.fromURL(Vfs.java:97) at org.reflections.vfs.Vfs.fromURL(Vfs.java:90) at org.reflections.Reflections.scan(Reflections.java:236) at org.reflections.Reflections$1.run(Reflections.java:199) at java.lang.Thread.run(Thread.java:834) at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:519) at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192) at com.oracle.svm.core.code.IsolateEnterStub.PosixJavaThreads_pthreadStartRoutine_e1f4a8c0039f8337338252cd8734f63a79b5e3df(generated:0) at com.oracle.graal.pointsto.constraints.UnsupportedFeatures.report(UnsupportedFeatures.java:129) at com.oracle.svm.hosted.NativeImageGenerator.runPointsToAnalysis(NativeImageGenerator.java:770) ... 8 more Expected behavior The native build should succeed without errors. Suggestions Move from Mongock to Liquibase with Mongo. Find a way to make Mongock compatible with the native build. Additional context Use ./mvnw dependency:tree -Dverbose to check which library is using VFS. Check this issue: https://github.com/quarkusio/quarkus/issues/9801 cc @danielpetisme closed by #180 I have just realised about this ticket and it seems that has been somehow sorted out. However is there anything that we can provide natively from Mongock to help on this or improve. Just let me know.
2025-04-01T06:39:12.577523
2019-09-17T20:51:33
494845895
{ "authors": [ "SudharakaP", "murdos" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7279", "repo": "jhipster/jhipster-core", "url": "https://github.com/jhipster/jhipster-core/issues/364" }
gharchive/issue
Add caffeine option for cacheProvider Overview of the feature request Caffeine is a new option available for cache since JHipster 6.3.0 Motivation for or Use Case It should be possible to select this option in JDL Related issues or PR https://github.com/jhipster/generator-jhipster/pull/10303 [X] Checking this box is mandatory (this is just to show you read everything) @murdos : Thanks for adding this; seems I've missed this. ๐Ÿ˜„
2025-04-01T06:39:12.587627
2024-12-05T11:11:31
2720110144
{ "authors": [ "atomfrede", "fabienpuissant", "renanfranca" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7280", "repo": "jhipster/jhipster-lite", "url": "https://github.com/jhipster/jhipster-lite/issues/11513" }
gharchive/issue
Gradle CI doesn't launch unit tests It seems that the gradle ci don't launch unit test only IT. I have tried to put a failing unit test and the ci was green Is this the behaviour or a misconfiguration on my side ? A possible working solution github-actions.yml jobs: tests: name: tests runs-on: ubuntu-latest timeout-minutes: 20 steps: - name: 'Setup: checkout project' uses: actions/checkout@v4 - name: 'Setup: environment' id: setup uses: ./.github/actions/setup - name: Setup Gradle uses: gradle/gradle-build-action@v2 - name: 'Test: run unit tests' run: ./gradlew clean test --no-daemon - name: 'Test: run integration tests' run: ./gradlew clean integrationTest --no-daemon @fabienpuissant : I looked at a random build action: feat(liquibase): async liquibase configuration should not be activated if liquibase is disabled. At least for this example, the tests were executed: > Task :test JwtReaderTest > should not authenticate user from empty token STANDARD_OUT 2024-12-05T23:23:09.354Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid empty JWT token JwtReaderTest > should not authenticate user from blank token STANDARD_OUT 2024-12-05T23:23:09.359Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid empty JWT token JwtReaderTest > should not authenticate user from invalid token STANDARD_OUT 2024-12-05T23:23:09.481Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid JWT token 2024-12-05T23:23:09.482Z TRACE --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid JWT token: Invalid compact JWT string: Compact JWSs must contain exactly 2 period characters, and compact JWEs must contain exactly 4. Found: 0 Here is the part of the github-actions.yml responsible for that: - name: 'Test: verify ${{ matrix.app }}' if: steps.tests-requirement-check.outputs.execute_tests == 'true' working-directory: /tmp/jhlite/${{ matrix.app }}/ run: | if [ -f 'mvnw' ]; then ./mvnw clean verify -Dsonar.qualitygate.wait=true sonar:sonar elif [ -f 'gradlew' ]; then ./gradlew clean build sonar --no-daemon --info else npm install npm test fi Please, could you provide an example that reproduces the not desired behavior? @renanfranca It's on a generated app See https://github.com/fabienpuissant/test-jh-lite-gradle-ci/actions/runs/12202622941 should fail because of : https://github.com/fabienpuissant/test-jh-lite-gradle-ci/blob/main/src/test/java/com/mycompany/myapp/FailingTest.java In the test-ci for jhlite it seems to be ok ๐Ÿ˜„ Not 100% sure just on my mobile, but with the soft ordering should run after for the integration test phase this does not define a dependency. So when just executing integrationtest this does not imply running unit tests. To work out of the box without Jacoco, it is going need to change the Gradle module to directly link unit and integration tests. Then, when adding Jacoco, remove that link to make Jacoco the glue. @fabienpuissant : I think your solution is better and more flexible ๐Ÿ˜‰. Even if we add the Jacoco module, the CI configuration will need to be updated to execute Jacoco instead of the IntegrationTest task. IMHO, If you want, feel free to proceed with the implementationโ€”if not, I can take care of it on Monday ๐Ÿ‘. Thanks for reporting that improvement/bug! cc: @pascalgrimaud , @murdos and @atomfrede
2025-04-01T06:39:12.602707
2021-11-16T09:14:09
1054632631
{ "authors": [ "arjung128", "zhe-juanz" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7281", "repo": "jhoffman/cycada_release", "url": "https://github.com/jhoffman/cycada_release/issues/41" }
gharchive/issue
Typo in Equation 5 in paper? Equation 5 in the paper makes use of f_S, however, the corresponding portion in Figure 2 (orange) only makes use of f_T. Additionally, f_T makes more sense since the input is in the target domain. Should the f_S in equation 5 be f_T? โ€Note that we can also consider a feature-level method Which distinguishes the features or semantics from two image sets as viewed under a task network. โ€œ I think both f _ s and f _ t can do experiments.
2025-04-01T06:39:12.608616
2024-08-04T19:18:41
2447264123
{ "authors": [ "jhon-elen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7282", "repo": "jhon-elen/Prueba3", "url": "https://github.com/jhon-elen/Prueba3/issues/8" }
gharchive/issue
"This is the solution of the issue" > "This is the solution of the issue" this is a quote Originally posted by @jhon-elen in https://github.com/jhon-elen/Prueba3/issues/7#issuecomment-2267641691 Anothe comment @jhonelentest Header Header Cell Cell Cell Cell import os os.chdir() Details This is a detail
2025-04-01T06:39:12.661119
2023-07-25T11:05:08
1820088029
{ "authors": [ "666asd", "jhy", "skylakeliu" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7283", "repo": "jhy/jsoup", "url": "https://github.com/jhy/jsoup/issues/1983" }
gharchive/issue
The document object obtained from parsing with jsoup.parse is missing one layer of elements compared to the original HTML. https://test.em.bonahl.com/#/user/login original : jsoup: https://github.com/jhy/jsoup/blob/1762412a28fa7b08ccf71d93fc4c98dc73086e03/src/main/java/org/jsoup/parser/HtmlTreeBuilderState.java#L417 It seems that nesting forms is not allowed here. However, modern frontend frameworks have implemented nested form nesting. Input: <div> <form id=1> <form id=2> </form> </form> jsoup parse: <div> <form id="1"> </form> </div> Chrome parse (Inspect Document) Spec: https://html.spec.whatwg.org/multipage/parsing.html#parsing-main-inbody A start tag whose tag name is "form" If the form element pointer is not null, and there is no template element on the stack of open elements, then this is a parse error; ignore the token. ... I believe jsoup is working per spec and to the same output as current browsers by only allowing one form on the stack at a time. Shadow DOM enables you to attach a DOM tree to an element, and have the internals of this tree hidden from JavaScript and CSS running in the page. Modern frontend frameworks have implemented neste HTML forms through Shadow DOM, and it have been widely used. jsoup should adapt to this.
2025-04-01T06:39:12.670798
2018-07-16T10:19:36
341462100
{ "authors": [ "bacongravy", "mhgharieb", "monkeydri", "ronjouch" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7284", "repo": "jiahaog/nativefier", "url": "https://github.com/jiahaog/nativefier/issues/655" }
gharchive/issue
no tabs on linux Description I assume that tabs are supported, since an issue for tabs feature-request has already been closed long time ago #116 and from some issues it seems that tabs are supported, ex in OSX : #647. However on linux (ubuntu 18.04) this is not the case, tabs open in new window. Steps to reproduce issue middle-click or ctrl-click a internal link opens a new window instead of a new tab. links are not either considered as external URLs because they do not open in web browser but in nativefier. Details Are you nativefying a public website ? => tried few websites, some hosted locally and some public websites. Feature request? Have you looked at nativefier --help to see if an existing option could fit your needs? => yes and I trired to specify --internal-urls but it did not change the behaviour. Full nativefier command used to build your app: nativefier --name "Gitlab" --icon "/path/to/gitlab/icon.png" --internal-urls "https://mydomain.comgitlab/*" "https://mydomain.com/gitlab" Version of Nativefier (run nativefier --version): v7.6.4 Version of node.js (run node --version): v9.2.0 OS: Ubuntu 18.04 The nativefier support for tabs comes from electron; support for native tabs in electron for macOS was added here: https://github.com/electron/electron/pull/9052 If electron gets updated to support tabs on more platforms then it should be relatively simple to add support for them to nativefier. It would be possible to change how middle-click and ctrl-click behave in nativefier, to open external links in an external browser when native tabs are unavailable, but I'm not convinced such a change would provide value, since external links already open in an external browser when you left-click on them. Being able to use middle-click to override the default "open externally" behavior seems useful to me. @monkeydri ๐Ÿ‘ to what @bacongravy said. See Release Notes for Nativefier v7.6.3: macOS: Add tabs, used automatically instead of windows (PR #579). Provided by Electron and (so far) no available for Windows/Linux, contributions welcome. alright, thanks fro your answer. It's a pity, as I don't think native tabs will be available on linux/windows anytime soon on Electron as there is no equivalent feature on them as macOS native tabs (ie. an option to display multiple windows from an app as tabs). @bacongravy Can the module electron-tabs help to support tabbing in Linux?
2025-04-01T06:39:12.679419
2022-10-03T06:32:19
1394210871
{ "authors": [ "ClashSAN", "jak3-taylor", "jianfch" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7285", "repo": "jianfch/stable-ts", "url": "https://github.com/jianfch/stable-ts/issues/2" }
gharchive/issue
Guidance running script Hello, very interested in the project, I would like to run the script. Supposedly there are some parameters, but I don't know how to add them. Running the script without parameters gives blank output. I also tried running the blocks of code as files, the first downloads model files and then errors at line 7: list indices must be integers or slices, not str How should one go about running this? I'm using windows Make sure you can run Whisper without any problems first. Next: import whisper from stable_whisper import modify_model model = whisper.load_model('base', 'cuda') modify_model(model) Then just use the model how you normally how you would use it. It should behave just like how whisper normally behaves, but with some additional values in the results. I'm having the same problem. I can use Whisper with no issues, but I always get this error when executing the example script. word_timestamps = results['segments']['word_timestamps'] TypeError: list indices must be integers or slices, not str I'm also using Windows 10 and I'm a complete novice, so apologies if I'm making a silly mistake. you can run like this but all the data is visible, so not very clean: make "run.py" file and put import whisper from stable_whisper import modify_model model = whisper/load_model('base', 'cuda') modify_model(model) results = model.transcribe('out.wav') print(results) type python run.py in the terminal. Maybe @jianfch can write a packaged demo script later. This is not for command line tools Yes, but that does not have single word timings, thats just generating a normal (better than youtube though w/punctuation) .srt I'd like to make .ass timed text subtitles for all videos. so I need the timestamps and word data Not sure about .ass, but you can change results_to_sentence_srt to results_to_token_srt to get the word timings. FYI - The "word" timestamps are actually token timestamps. Since token:word is not always 1:1 (varies by language), you may need to do some additional processing to get individual word timings. Thanks, that's super helpful! word_timestamps = results['segments']['word_timestamps'] TypeError: list indices must be integers or slices, not str that was a mistake on my part for not updating the execution lines in readme word_timestamps are under each segment, instead of the segments which is a list so the word_timestamps for first segments would be: first_segment_token_timestamps = results['segments'][0]['word_timestamps'] if you want a list of just the word_timestamps word_timestamps = [seg['word_timestamps'] for seg in results['segments']]
2025-04-01T06:39:12.706799
2016-04-01T06:54:55
145102350
{ "authors": [ "arun-manthani", "ecbypi" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7286", "repo": "jim/carmen-rails", "url": "https://github.com/jim/carmen-rails/issues/58" }
gharchive/issue
subregion_select dropdown has value numeric instead of alphabet code Hi there, As part of testing just observed that for country Denmark, the subregion select dropdown has value numeric. State Hovedstaden Midtjylland Nordjylland Sjรฆlland Syddanmark On saving it, the same numeric value is being saved in table. Is it normal behaviour, any fix will be really appreciated. Those are the region codes as found by the iso-codes debian package so this is expected behavior unless that codes have changed since the last time the script was run in the carmen gem. Select tag looks like this: Hovedstaden Midtjylland Nordjylland Sjรฆlland Syddanmark So on saving value, 84 saves in the database instead of state code..
2025-04-01T06:39:12.722428
2022-07-27T05:26:19
1319038763
{ "authors": [ "jimratliff" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7287", "repo": "jimratliff/pgn4people-poc-demo", "url": "https://github.com/jimratliff/pgn4people-poc-demo/issues/26" }
gharchive/issue
cm-chessboard: Figure out the status of pawn promotion Pawn promotion seems an issue with all of these โ€œjust a boardโ€ packages. See closed issue โ€œ Pawn promotion support? #19 at cm-chessboard. To the extent that Iโ€™m just replicating FEN strings, I donโ€™t think thereโ€™s any problem. Where a need would arise is either (a) when Iโ€™m stepping through a main line (rather than using FENs) or (b) when the user is adding an line that has a promotion. Closed as out of scope for this repository.
2025-04-01T06:39:12.792086
2021-05-13T05:00:39
890704247
{ "authors": [ "KaishuaiXu", "jingtaozhan" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7288", "repo": "jingtaozhan/DRhard", "url": "https://github.com/jingtaozhan/DRhard/issues/6" }
gharchive/issue
about the length of tokens Hello, I have read your paper and am quite interested in your work! There is a question about the tokens. I notice you truncat the passage tokens with 120 in MSMARCO Passage Retrieval, however, for ANCE, the original paper uses 512 tokens. So does the number of tokens have the impact on the accuracy? No We just don't have that many advanced GPUs to afford very long input :( BTW, the average passage length is about 70 tokens and truncating to 120 shouldn't be a problem. hhhhhhhhha! Thanks a lot for your reply! I have no multiple GPUs either! Lol! I agree with your settings!
2025-04-01T06:39:12.799155
2014-08-08T03:08:14
39788993
{ "authors": [ "jinzhu", "wb14123" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7289", "repo": "jinzhu/gorm", "url": "https://github.com/jinzhu/gorm/issues/190" }
gharchive/issue
Set default value while AutoMigrate If no default value, the query with gorm will get error. Hi @wb14123 I just pushed a commit, so even the data in database is NULL, gorm won't get any error now. So we don't need the default value for migrations Thank you.
2025-04-01T06:39:12.802452
2015-07-28T14:32:53
97720510
{ "authors": [ "jinzhu", "pariz" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7290", "repo": "jinzhu/gorm", "url": "https://github.com/jinzhu/gorm/pull/585" }
gharchive/pull-request
Fixed issue https://github.com/jinzhu/gorm/issues/151 where postgresql connection problems fail silently For reference: https://github.com/jinzhu/gorm/issues/151 Errors connection to a postgresql fail silently, resulting in errors. This fix sends a ping right after the connection has been made. If the ping fails, we assume the connection is broken. Accepting better solutions. This quick fix did it for me :) Should we always Ping() the database after connection? It's not the most pretty solution. But it an extra safetycheck to verify that the connection is in fact alive. Should i rewrite the pr and make the Ping() generic? Hi @pariz Yes, I think it is a better solution than just check the postgres. I agree. Code has been updated. Epic oneliner! :) picked your code to the master, thank you.
2025-04-01T06:39:12.856594
2018-03-07T20:05:38
303241067
{ "authors": [ "lyubomir", "paweldomas" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7291", "repo": "jitsi/react-native-webrtc", "url": "https://github.com/jitsi/react-native-webrtc/pull/20" }
gharchive/pull-request
[iOS] Fix uncaught NSInvalidArgumentException in RTCPeerConnection's createAnswer WebRTC appears to fail RTCPeerConnection's createAnswer without an error message i.e. nil. But NSMutableDictionary does not accept nil as a value and causes an uncaught exception. Based on Daniel Ornelas' comment that the "error" key should actually be "NSLocalizedDescription" and looking at the error.localizedDescription, I suppose it's better to morph this one further. ok, let me know when it's ready to be merged
2025-04-01T06:39:12.860669
2016-03-03T05:38:56
138081139
{ "authors": [ "jjf28", "xboi209" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7292", "repo": "jjf28/Chkdraft", "url": "https://github.com/jjf28/Chkdraft/issues/92" }
gharchive/issue
Spaces vs Tabs Why are you using spaces now? Tabs makes the code look more clean and organized in my opinion. I'm not sure how its cleanliness/organization is adversely affected by this change, could you elaborate on that? I decided to switch to spaces mainly because of the inconsistency between different programs/websites/browsers/users. Through some of these lenses tabs come out as long as 8 spaces (rather than the standard 4, in VS), requiring more scrolling or resulting in more code flowing onto subsequent lines (much less readable), can be especially painful on phones/small resolution devices. Tabs also caused problems with 'pretty spaced' code (where params/array initializers are lined up nicely for easy reading) which could easily end up ugly. Admittedly the biggest issue here was that tabs were used for more than just the start-of-line to start-of-code portions; but making this change certainly helped identify/eliminate those that were in the middle and makes it very easy to do so in the future. It's also easier to find stray tabs than stray spaces, if keeping the project on a 100% consistent convention ever became relevant (using different forms Iโ€™ve heard, can cause lotsa merge issues). Productivity wise I find the two indentation options to be quite equivalent, Visual Studios can easily be set to turn tab key-presses into 4-spaces (Tools->Options->Text Editor->C/C++->Tabs); white space can be easily jumped with ctrl+arrows or deleted with ctrl+delete and inserting or re-inserting a ; or } automatically spaces chunks of code. For tabs the only advantages I read up on and considered legitimate was the ability for users to set their size (in some programs, not all give this option) and the relative ease for newer coders - which didn't, in my opinion, outweigh their disadvantages. Some discussion on the matter: http://programmers.stackexchange.com/questions/57/tabs-versus-spaces-what-is-the-proper-indentation-character-for-everything-in-e It appears I can't link SEN shoutbox convos, so http://www.staredit.net/shoutboxsearch/ + starter phrase "like my friend and I decide for our project that indentation is to be this many spaces" then hit context.
2025-04-01T06:39:12.881072
2022-06-14T01:45:35
1270145971
{ "authors": [ "jjklemm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7293", "repo": "jjklemm/sudoku-solver", "url": "https://github.com/jjklemm/sudoku-solver/issues/1" }
gharchive/issue
Method for creating and printing board The user needs a way to put a sudoku board into the program. Boards should be printed in a way that is easy to read. A graphical user interface will be created using Swing. The GUI will allow for a user to create a sudoku board to be solved, as well as display solutions.
2025-04-01T06:39:12.889890
2023-08-28T04:23:24
1868932522
{ "authors": [ "ccolomb2", "jkitchin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7294", "repo": "jkitchin/f23-06623", "url": "https://github.com/jkitchin/f23-06623/issues/3" }
gharchive/issue
Sub listed item in numbered list in 01-jupyter.ipynb Please check the numbered list in one of the first markdown cells of the notebook. For the sub listed item I think one more space in front of the "1." is necessary, otherwise the item just becomes part of the main list. fixed. thanks.
2025-04-01T06:39:12.910528
2016-04-14T12:05:38
148340174
{ "authors": [ "elineopsommer", "jkuri", "leypoldt" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7295", "repo": "jkuri/ngTimepicker", "url": "https://github.com/jkuri/ngTimepicker/issues/3" }
gharchive/issue
Update code: issue #2 Updating model from controller does not update control Can you take this solution along in your code? Thanks scope.$watch(attrs['ngModel'], function (newTime) { scope.initTime = newTime; reinitTime(); }); Spot on! This fix was exactly what i was looking for. Can someone make a PR please? I had already made a PR: https://github.com/jkuri/ngTimepicker/pull/4 Sorry, I missed it. Thanks for your PR, it's merged.
2025-04-01T06:39:12.921152
2022-12-06T07:35:26
1478454439
{ "authors": [ "elmeister", "jlalmes" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7296", "repo": "jlalmes/trpc-openapi", "url": "https://github.com/jlalmes/trpc-openapi/issues/196" }
gharchive/issue
Lack of support for valid endpoints Consider the following code from Usage with Express getUser: t.procedure.input(z.string()).query((req) => { req.input; // string return { id: req.input, name: 'Bilbo' }; }), After adding this line .meta({ openapi: { method: 'GET', path: '/getUser' } }) trpc-openapi fail to process it TRPCError: [query.getUser] - Input parser must be a ZodObject This is correct. - z.string() + z.object({ id: z.string() }) This is correct. You have linked to the tRPC documentation. trpc-openapi has a few additional requirements so you may need to make a few changes to your procedure. - z.string() + z.object({ id: z.string() }) Please make sure you have read this ๐Ÿ‘‰ https://github.com/jlalmes/trpc-openapi/blob/master/README.md Well, it's string and not object (see the // string comment - it's there, in the documentation, not something I added), so it can't be correct. And thing is that we not only need to change the procedure, but change API requests as well, which is too high price to pay. Hi Sergey. Please read the README before commenting again. so it can't be correct trpc-openapi requires that your inputs are wrapped in an z.object. need to change the procedure Yes. change API requests I don't think you've understood REST. Consider the following procedure. getUser: t.procedure .meta({ openapi: { method: 'GET', path: '/user' } }) .input(z.object({ userId: z.string() })) .output(z.object({ id: z.string(), name: z.string() })) .query(({ input }) => { console.error(typeof req.input.userId); return { id: req.input, name: 'Bilbo' }; }), This is how it can be invoked. @trpc/server ๐Ÿ‘‰ GET /getUser?input=${encodeURI(JSON.stringify({ userId: 'abc123' }))} trpc-openapi ๐Ÿ‘‰ GET /user?userId=abc123 too high price to pay That's your decision to make.
2025-04-01T06:39:12.927371
2021-01-19T12:02:03
788965082
{ "authors": [ "fbricon", "mike-england" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7297", "repo": "jlandersen/vscode-kafka", "url": "https://github.com/jlandersen/vscode-kafka/issues/84" }
gharchive/issue
SSL support Is it possible to attach to a kafka instance with SSL enabled? For example, when I'm attaching to my cluster as a producer, I'll use the following: KafkaProducer(bootstrap_servers=kafkaServer, security_protocol="SSL", ssl_check_hostname=True) When attaching to the kafka cluster through this extension, I can specify the bootstrap server and port (443 in my case), but I can't seem to find a way to change the security protocol. Currently SSL is only enabled when using an authentication scheme, i.e. you have a username/password. So you need SSL but no auth? Currently SSL is only enabled when using an authentication scheme, i.e. you have a username/password. So you need SSL but no auth? Yes, currently we don't have any auth, just SSL. I tried to specify SASL/Plain as authentication but with a blank username password, but that doesn't seem to work. Yes, currently we don't have any auth, just SSL. I tried to specify SASL/Plain as authentication but with a blank username password, but that doesn't seem to work. Ok, so your use case seems quite simple but if we want to do SSL support right (and advertise we support SSL), we need to be able configure SSL certificate configuration, as per https://kafka.js.org/docs/configuration#ssl Currently, the ssl boolean is only set here: https://github.com/jlandersen/vscode-kafka/blob/a0257d0932263cc66bce73c5579a9d530f9655b0/src/client/client.ts#L325 Ok, so your use case seems quite simple but if we want to do SSL support right (and advertise we support SSL), we need to be able configure SSL certificate configuration, as per https://kafka.js.org/docs/configuration#ssl Currently, the ssl boolean is only set here: https://github.com/jlandersen/vscode-kafka/blob/a0257d0932263cc66bce73c5579a9d530f9655b0/src/client/client.ts#L325 Brilliant! That works well for my purposes, it connects fine and seems to be working well. Brilliant! That works well for my purposes, it connects fine and seems to be working well.
2025-04-01T06:39:12.950399
2023-08-22T18:30:20
1861984934
{ "authors": [ "Bagu", "jlesage", "walawa75" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7298", "repo": "jlesage/docker-crashplan-pro", "url": "https://github.com/jlesage/docker-crashplan-pro/issues/438" }
gharchive/issue
Cannot see files in folders since Container Manager feature Current Behavior Hello Since the appearance of Container Manager instead of Docker package I noticed that the backup was constantly stuck at 0 files found (0mb) which is not normal. So I browsed through files and noticed that folders and files were not visible. The permissions are good (USER_ID=0 and GROUP_ID=0). The problem happened after a reboot following an image update. I noticed that by restarting the container from the command line, the problem was gone. Does anyone also have the problem? What to do to continue to benefit from the evolutions of Container Manager (simple update feature) without having to restart systematically on the command line. I specify that the bug is in no way related to the good work of our dear JLesage ;-) Expected Behavior No response Steps To Reproduce No response Environment OS: OS version: CPU: Docker version: Device model: Browser/OS: Container creation Container Manager package Container log no abnormal log Container inspect No response Anything else? No response I assume you are using a Synology NAS? Looks like the issue where /volume1 is mapped to /storage (see #310). Is it the case? Yes I use a Synology NAS but the issue not seems to be the same as #310 tab. I will try to make new tests later You can also share the output of docker inspect <container name> so we can have a complete picture of the container's configuration. I deleted the container and created a new one. Issue seems gone when container started from Container Manager Hello, Now I get 1,5 years to backup 4 TB. How to increase backup speed ? My ISP speed is about 700 mbps but crashplan backup is really slow. Yes, it is known that upload speed is slow. However, because of deduplication, a lot of data don't really have be uploaded. If you look at Tools->History, you can see the "Effective rate": the value can be very high and would typically improve over time. I had the same problem. In fact, when the docker container was created (a long time ago), the storage volume was not requested. Since the last update, not only is it requested, but if it hadn't been configured originally, it points to an empty virtual folder. I installed portainer to change the configuration and set storage bind to volume1. Since then, I've been able to browse storage, but for crashplan, it's not the same folder structure. As a result, the backup started from scratch. I hope that deduplication will play its part, otherwise I'm in for several months of backups.
2025-04-01T06:39:12.952170
2019-05-09T12:38:42
442215161
{ "authors": [ "KevinZhangt", "jlesage" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7299", "repo": "jlesage/docker-nginx-proxy-manager", "url": "https://github.com/jlesage/docker-nginx-proxy-manager/issues/15" }
gharchive/issue
Use custom DB instance Hi, thanks a lot to share this ,it is really great. I'm wandering can we use our custom DB instance (like MySQL or Maria DB) instead of MySQL in container ,thanks. If you want to use your own DB instance, look at the original project: https://github.com/jc21/nginx-proxy-manager
2025-04-01T06:39:12.966448
2020-03-18T17:06:22
583876334
{ "authors": [ "TheZoker", "bookandrelease", "jlesage", "mdisieno" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7300", "repo": "jlesage/docker-nginx-proxy-manager", "url": "https://github.com/jlesage/docker-nginx-proxy-manager/issues/57" }
gharchive/issue
Let's Encrypt: Internal Error I used the hassio addon for quite some time now. Today I tried to migrate the proxy manager to my synology. For that I use the synology docker manager and this docker image. So I installed the latest version (1.7.0) and configured my port forwarding. The only issue I had was with the lets encrypt certificate. When I tried to request a certificate, I got an "Internal error": This is the content in the /config/log/letsencrypt/letsencrypt.log: 2020-03-18 17:02:07,145:DEBUG:certbot.main:certbot version: 0.30.2 2020-03-18 17:02:07,147:DEBUG:certbot.main:Arguments: ['--non-interactive', '--config', '/etc/letsencrypt.ini', '--cert-name', 'npm-15', '--agree-tos', '--email'<EMAIL_ADDRESS>'--preferred-challenges', 'dns,http', '--webroot', '--domains', 'wg.zkr.io'] 2020-03-18 17:02:07,150:DEBUG:certbot.main:Discovered plugins: PluginsRegistry(PluginEntryPoint#manual,PluginEntryPoint#null,PluginEntryPoint#standalone,PluginEntryPoint#webroot) 2020-03-18 17:02:07,202:DEBUG:certbot.log:Root logging level set at 20 2020-03-18 17:02:07,204:INFO:certbot.log:Saving debug log to /config/log/letsencrypt/letsencrypt.log 2020-03-18 17:02:07,206:DEBUG:certbot.plugins.selection:Requested authenticator webroot and installer None 2020-03-18 17:02:07,220:DEBUG:certbot.plugins.selection:Single candidate plugin: * webroot Description: Place files in webroot directory Interfaces: IAuthenticator, IPlugin Entry point: webroot = certbot.plugins.webroot:Authenticator Initialized: <certbot.plugins.webroot.Authenticator object at 0x7fded3100a90> Prep: True 2020-03-18 17:02:07,223:DEBUG:certbot.plugins.selection:Selected authenticator <certbot.plugins.webroot.Authenticator object at 0x7fded3100a90> and installer None 2020-03-18 17:02:07,224:INFO:certbot.plugins.selection:Plugins selected: Authenticator webroot, Installer None 2020-03-18 17:02:07,320:DEBUG:acme.client:Sending GET request to https://acme-v02.api.letsencrypt.org/directory. 2020-03-18 17:02:07,326:DEBUG:urllib3.connectionpool:Starting new HTTPS connection (1): acme-v02.api.letsencrypt.org 2020-03-18 17:02:12,333:DEBUG:certbot.log:Exiting abnormally: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 141, in _new_conn (self.host, self.port), self.timeout, **extra_kw) File "/usr/lib/python3.6/site-packages/urllib3/util/connection.py", line 60, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "/usr/lib/python3.6/socket.py", line 745, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno -3] Try again During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 601, in urlopen chunked=chunked) File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 346, in _make_request self._validate_conn(conn) File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 850, in _validate_conn conn.connect() File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 284, in connect conn = self._new_conn() File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 150, in _new_conn self, "Failed to establish a new connection: %s" % e) urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/requests/adapters.py", line 445, in send timeout=timeout File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 639, in urlopen _stacktrace=sys.exc_info()[2]) File "/usr/lib/python3.6/site-packages/urllib3/util/retry.py", line 388, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='acme-v02.api.letsencrypt.org', port=443): Max retries exceeded with url: /directory (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again',)) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/bin/certbot", line 11, in <module> load_entry_point('certbot==0.30.2', 'console_scripts', 'certbot')() File "/usr/lib/python3.6/site-packages/certbot/main.py", line 1364, in main return config.func(config, plugins) File "/usr/lib/python3.6/site-packages/certbot/main.py", line 1233, in certonly le_client = _init_le_client(config, auth, installer) File "/usr/lib/python3.6/site-packages/certbot/main.py", line 604, in _init_le_client acc, acme = _determine_account(config) File "/usr/lib/python3.6/site-packages/certbot/main.py", line 521, in _determine_account config, account_storage, tos_cb=_tos_cb) File "/usr/lib/python3.6/site-packages/certbot/client.py", line 181, in register acme = acme_from_config_key(config, key) File "/usr/lib/python3.6/site-packages/certbot/client.py", line 51, in acme_from_config_key return acme_client.BackwardsCompatibleClientV2(net, key, config.server) File "/usr/lib/python3.6/site-packages/acme/client.py", line 814, in __init__ directory = messages.Directory.from_json(net.get(server).json()) File "/usr/lib/python3.6/site-packages/acme/client.py", line 1152, in get self._send_request('GET', url, **kwargs), content_type=content_type) File "/usr/lib/python3.6/site-packages/acme/client.py", line 1101, in _send_request response = self.session.request(method, url, *args, **kwargs) File "/usr/lib/python3.6/site-packages/requests/sessions.py", line 512, in request resp = self.send(prep, **send_kwargs) File "/usr/lib/python3.6/site-packages/requests/sessions.py", line 622, in send r = adapter.send(request, **kwargs) File "/usr/lib/python3.6/site-packages/requests/adapters.py", line 513, in send raise ConnectionError(e, request=request) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='acme-v02.api.letsencrypt.org', port=443): Max retries exceeded with url: /directory (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again',)) 2020-03-18 17:02:12,340:ERROR:certbot.log:An unexpected error occurred: Is this an issue with this image or with the base image or is my configuration wrong? Is this issue related? https://github.com/jc21/nginx-proxy-manager/issues/180#issuecomment-599278001 Thank you! Edit: A little bit more information: I also tried version 1.6.0, but it has the same issue. For both version: Connecting to my devices trough the manager using port 80 works fine. Connecting to them via 443 leads to a empty response. Looks like the container is not able to reach the LetEncrypt server. Can you try: docker exec <container name> ping acme-v02.api.letsencrypt.org Hmm yes, something does not seem right: When I try the same with my AdGuard container it works: AdGuard is running in host mode, while ProxyManager is running in bridge mode. But bridge mode should work as well, right? Do you have any idea, how I can fix that? Are you using AdGuard as your DNS server ? If yes. this may be the cause of the issue. Are you able to ping acme-v02.api.letsencrypt.org from your Synology ? Yes I'm using it as my DNS server. Yes the ping from synology itself works: Is this DNS server also used by you Synology? You can try to run the following command on your Synology: nslookup acme-v02.api.letsencrypt.org <DNS server IP>. Also, can you confirm that this DNS server used by the container: docker exec <container name> cat /etc/resolv.conf I am having the same internal error message. I use dnsimple and I do have a certificate on my account for my domain. Pinging the address you said above is successful. If your container accessible from the internet on port 80 ? You can check with https://www.yougetsignal.com/tools/open-ports/ Same issue here. I am able to ping with: docker exec ping acme-v02.api.letsencrypt.org No internal DNS beyond a redirect to an external PiHole server. I tried both requesting a new cert (internal error) and creating my own via a LetsEncrypt container. Both no dice. Make sure the container is reachable from the internet on port 80. You can check with https://www.yougetsignal.com/tools/open-ports/ Morning. Port shows as not reachable but both my router settings and the use of the proxy manager is present over port 80. If I use the IP:80 I go directly to where I set redirect to. 443 shows open and listening. Could I just direct through that? 80 is still working fine, It appears my isp blocks port 80, though I can go it via both directly through the IP and via my duckdns dyndns. If your ISP blocks port 80, then I guess your cannot use Nginx Proxy Manager with automatic certificate generation from LetsEncrypt. Port 80 is required. This is how LetsEncrypt validates that your are the owner of the DNS name. Odd given I can use port 80 for the hosts from Nginx Proxy Manager. I'll do some more research and report back on a fix for future reference.
2025-04-01T06:39:13.001096
2017-01-07T00:08:40
199320982
{ "authors": [ "jlord", "shiftkey" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7301", "repo": "jlord/git-it-electron", "url": "https://github.com/jlord/git-it-electron/pull/180" }
gharchive/pull-request
[WIP] A Proposal For Testing Things As we talked about in https://github.com/jlord/git-it-electron/pull/178#issuecomment-270530180, it'd be nice to be able to run the verify tests automagically. This PR is my attempt to tackle the problem, which required some refactoring of how the verify scripts work. Current Limitations A couple of issues that I've had to work through: verify scripts touch the DOM as part of rendering the results - I've attempted to move this code back up to the challenge script, but it's only done for one verify script so far (the repository tests). some verify scripts are written as if they are blocking but due to using exec they need to be asynchronous - we need a way to inspect the results, so I've introduced a callback that takes an array of results. Maybe you have other ideas in mind. How This Works We're using tape to run scripts (I'm more familiar with mocha, so I'm open to other ways to get the best out of the tape API), but here's a simple test: test('verifies the hello world repository', function (t) { var folder = helper.extractFixture('hello-world') verify(folder, function (result) { var expected = 'This is a Git repository!' var first = result[0] t.ok(first.result) t.equal(expected, first.message) t.end() }) }) For scenarios where we need the repository to be in a certain state, we can splat the folders onto disk under the tests/fixtures folder - and then extract them to a temporary folder when running the tests. To get this working, you can just rename the .git folder to _git and the repository state is then version-controllable. The other notable change here is that we have a callback to poke at the results from the verify script (because exec is asynchronous), to ensure we're getting the expected results. I'm using the same shape of "message is a string" and "result is a boolean" that the existing verify scripts use. To get this working in a test harness, I needed to move the code for addToList and challengeIncomplete/markChallengeCompleted out from the verify script to where the verify method is invoked. A partial implementation of this in challenge.js works for the repository script, but it definitely needs more work and testing as I might not be understanding it correctly. What's Left To Do? A rough list: [ ] get feedback on this approach [ ] sketch out the changes to test a path-less script, update challenge.js to handle this [ ] propose a way to test request invocations (nock? something else?) [ ] ??? verify scripts touch the DOM as part of rendering the results Yeah, this is something I've wanted to improve; to separate these concerns. Which is probably a big enough project in itself to be a different PR and a thing to figure out before this? I haven't gone through all this with a fine tooth comb but in general I feel like it's what I've been thinking! ๐Ÿ‘ I have a couple things left to clean up, cause I don't think master is quite stable right now but then I'll pick up from what you've started here in moving out the DOM touching stuff and I'll get a PR going for that. โœจ :cool: Which is probably a big enough project in itself to be a different PR and a thing to figure out before this? Up to you - if you feel confident enough to do that refactoring before introducing tests then :thumbsup:
2025-04-01T06:39:13.013024
2015-05-22T13:08:57
79449040
{ "authors": [ "danielcha", "jmdobry" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7302", "repo": "jmdobry/angular-cache", "url": "https://github.com/jmdobry/angular-cache/issues/181" }
gharchive/issue
View not being update with angular-cache with localStorage I'm building an Ionic app the uses angular-cache to leverage data persistence. This data is displayed in a ng-repeat list. Either by user interaction (in another view!) or new data being loaded in the background, I need the view to update to reflect this. When developing my app everything works perfectly, however, as soon as I start using angular-cache with localStorage, views stop being updated. When I turn off localStorage (but still use in memory caching from angular-cache) everything works as expected: the data updates and so is the DOM. I was able to reproduce the issue with this CodePen http://codepen.io/anon/pen/qdayMW. To demonstrate the issue you switch tabs (from home to about). If you look in the console you can see that new items are being added to the localStorage cache. However the homescreen doesn't update. If you then refresh the page (within the 10 second cache magAge) the page shows the correct list of items... However, if we comment out line 81, the caching is no longer persistent after a refresh, however the new items are added to the DOM every time the view changes. Note: I also posted this issue to the Ionic forum as I'm not sure which exactly is the root of the problem: http://forum.ionicframework.com/t/view-not-being-update-with-angular-cache-with-localstorage/25236 This must be something weird with ionic. I've been playing with your codepen, and as I click between the home and about tabs, the home controller code is only executed once at the very beginning, so while more items are being added to the cache, the resolve code that would get the latest items isn't run again. The latest ionic implemented view caching (which is causing your "bug"), which you have to disable: http://ionicframework.com/docs/nightly/api/directive/ionNavView/
2025-04-01T06:39:13.017310
2021-06-02T14:41:20
909577463
{ "authors": [ "Reacien", "jmir1" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7303", "repo": "jmir1/aniyomi", "url": "https://github.com/jmir1/aniyomi/issues/39" }
gharchive/issue
[Bug] Kitsu anime tracking not working Device information Tachiyomi version: Latest stable and preview version Android version: 10 Device: Oneplus 5T The kitsu anime tracking only shows the manga's and not the anime. Checked it with myanimelist but that one does work, but it's not the site I use to keep my library updated. Would be awesome if it can be fixed. Keep up the great work! forgot to close. was fixed with https://github.com/jmir1/aniyomi/commit/9e2db50b77fdd99578eddc55d07cc343bacb049b.
2025-04-01T06:39:13.040101
2023-07-20T17:15:59
1814461043
{ "authors": [ "bkruger99", "jmorganca", "pdevine" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7304", "repo": "jmorganca/ollama", "url": "https://github.com/jmorganca/ollama/issues/143" }
gharchive/issue
error loading model: unexpectedly reached end of file On a couple of models I am receiving this error: llama.cpp: loading model from /Users/REDACTED/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 error loading model: unexpectedly reached end of file llama_load_model_from_file: failed to load model This happens with a couple of the larger models: nous-hermes:latest llama2:13b If I do ollama pull against them, the manifests match up and it doesn't re-pull anything. Since this looks like docker under the hood, are the models corrupt? or? Any thoughts? FWIW, llama2:latest and wizard-vicuna:latest work fine. M2 Macbook Pro 32 Gigs of ram. Thanks @bkruger99, will check out why this is happening Let me know if you need any additional debugging data from my side. You'll have to tell me how to enable other than running server via cli :) Great! @bkruger99 is this on Mac? Thanks! Yes! Hardware: Model Name: MacBook Pro Model Identifier: Mac14,10 Model Number: Z174000EBLL/A Chip: Apple M2 Pro Total Number of Cores: 12 (8 performance and 4 efficiency) Memory: 32 GB OS: Ventura 13.4.1 (c) @bkruger99 can you run: sha2 -256 ~/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 Check to see that sha sum matches, and if it doesn't you can rm ~/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 and then re-pull the image. There's a fix that I think was just merged that will make certain the sha sum is verified correctly when you're pulling the layers. Yeah. there's something w/ manifest not verifying the sha256 when pulling. These two models did have a network interruption as the laptop went to sleep. โฏ shasum -a 256 sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 f2a1788633ddf3edef0ee4d9d4e93c399bfeeeb7363015d7c1b630ff268cdcf5 sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 I re-pulled llama2:12b and it's happy, I'll do the same with the rest of 'em. The next version will check the SHAs; the re-pull is pretty tolerant of network interruptions, but wondering if the buffer wrote garbage onto the end of the partial file somehow. I haven't (yet) tested with sleeping the machine though, so that could have been the reason. I'm going to go ahead and close the issue. Feel free to re-open it though.
2025-04-01T06:39:13.102597
2012-01-09T12:20:18
2768101
{ "authors": [ "jdelStrother", "jnicklas" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7305", "repo": "jnicklas/capybara", "url": "https://github.com/jnicklas/capybara/issues/602" }
gharchive/issue
Raise ArgumentError when passing invalid keys to has_selector? It's a relatively common misconception that you can use :content as a key - e.g. has_selector?('h1', :content=>'wooo!') which would silently pass, even when 'wooo!' is nowhere to be found. This change should make that error much more obvious. Related issues : https://github.com/jnicklas/capybara/issues/525 It would be nice to raise an error on passing a block in - eg @session.should have_selector("//p") do |p_tag| p_tag.should have_selector("//a") end which seems to be a common error on people switching to capybara via webrat. However, that's a bit more tricky and would probably need fixing at the rspec-matcher level, since the block doesn't make it as far as has_selector? and gets silently discarded. ValidKeys on Selector could be replaced with something like this : diff --git a/lib/capybara/selector.rb b/lib/capybara/selector.rb index 1bff91d..0b85a04 100644 --- a/lib/capybara/selector.rb +++ b/lib/capybara/selector.rb @@ -15,6 +15,10 @@ module Capybara def remove(name) all.delete(name.to_sym) end + + def valid_keys + all.map{|selector| selector.custom_filters.keys}.flatten.uniq + end end def initialize(name, &block) ? I'm not sure what you mean about the implementation for checking query keys, though. There's 3 places that keys passed to has_selector?() are used - node/matchers.rb, query.rb, and selector.rb. Whichever one is responsible for checking the keys will have to also check the keys of the other 2. I could move assert_valid_options to a class method to Query if you like? Closing this pull request, we already have an open issue for this in #525, and I don't like this implementation
2025-04-01T06:39:13.151373
2021-04-19T00:31:09
860806476
{ "authors": [ "CaptainHalflife", "ImRyley", "jobobby04" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7311", "repo": "jobobby04/TachiyomiSY", "url": "https://github.com/jobobby04/TachiyomiSY/issues/277" }
gharchive/issue
[Feature Request] Change blue loading and badge accent colors in Midnight Dusk theme to something similar to J2K Midnight Dusk? Why/User Benefit/User Problem Improves the theme by keeping the color scheme. What/Requirements The blue accent (and possibly black background on loading) would be changed to better fit the Midnight Dusk theme. SY Midnight Dusk J2K Midnight Dusk Why/User Benefit/User Problem Improves the theme by keeping the color scheme. What/Requirements The blue accent would be changed to better fit the Midnight Dusk theme. Someone can do that for the other themes too and why the text in black-red are red instead of white ? I think its better to just replace everything that blue to red and change the colour of the download badge to one that look good for each theme Already addressed by main tachi
2025-04-01T06:39:13.165262
2016-05-29T09:46:12
157378264
{ "authors": [ "dvdobrovolskiy", "joeaudette" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7312", "repo": "joeaudette/cloudscribe.Web.Pagination", "url": "https://github.com/joeaudette/cloudscribe.Web.Pagination/issues/12" }
gharchive/issue
async requests Is any chance an option to use async requests? like asp.net core rc2 standard template public async Task<IActionResult> Index() { return View(await _context.Reason.ToListAsync()); } _context.Reason.ToPagedList(0, DefaultPageSize) does not have ToListAsync method honestly PagedList is a completely separate thing from the pagertaghelper, pagertaghelper does not depend on PagedList and really the only reason pagedlist exists in this library is because it was based on Martijn Boland's MVCPaging project and it had a pagedlist which I copied. PagedList is kind of a hack for when you are retrieving more than a page of data but want to do paging against the result set. I don't use PagedList at all myself and don't really care to try to support it or update it. There are other implementations of PagedList out there that anyone could also use, so it is completely independent from the pagertaghelper. From a brief search I found this pagedlist which has async support, but not sure it has a compatible nuget for aspnetcore rc2. But you can search and maybe find a better compatible one, I did not spend much time looking. Or if anyone is interested in implementing that and submitting a pull request I'm open to that as well. I do async methods to get data in my own use of pagertaghelper, and I just return one page of data at a time from the database so I don't need any pagedlist. maybe could do it like this: public async Task<IActionResult> Index() { var data = await _context.Reason.ToListAsync(); return View(data.ToPagedList(0, DefaultPageSize)); } but again that looks like it returns more than a page of data and then uses pagedlist to page over it which is what PagedList is for but also why PagedList is a hack compared to just return a page of data from the database using .Skip .Take etc only thing I've ever used PagedList for is the demo pages because it was convenient to make a demo without any database, and much of the demo code was also copied from Martijn's project and that is how PagedList ended up in this project. I never use that PagedList in my own projects, I only use the taghelpers hm, sorry I a bit confused here. if you dont use PagedList what you use in your Pagination project? I mean then you retrieve data from DB. you can see an example here a controller which has methods for user admin and paging of users with search, and the corresponding partial view here as I said the use of PagedList in the pager demo pages was mainly because I copied existing demo pages and that makes it easy to have demo without any db, you just make a big list of data and use pagedlist to page over it. but again PagedList is only for if you already retrieved (or created as in the demo code) more than a page of data and now you want to page over that data. It is much better to not use PagedList and only retrieve data one page at a time from the db. PageTagHelper doesn't know about PagedList at all. not that in the example I linked, you will see model.Paging.* properties being set, in that model Paging is an instance of PaginationSettings it is simply a model to pass settings on to the pagertaghelper this makes me think I should get rid of PagedList and not use it in the demos because it makes people think they need to use that when in fact it is best not to fyi, I've reworked the demo app and removed all use of PagedList from the demo so as not to encourage use of that Many tnks. checking
2025-04-01T06:39:13.176382
2018-03-01T08:17:05
301313799
{ "authors": [ "hirnschmalz", "sseide", "tburschka" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7313", "repo": "joeferner/redis-commander", "url": "https://github.com/joeferner/redis-commander/issues/236" }
gharchive/issue
Multiple connections to the same instance with different DBs not working Hello, I'm try to run the docker container with the REDIS_HOSTS env variable witch looks like this: REDIS_HOSTS='dev:redis:6379:0,test:redis:6379:10' but it seems that the given database is ignored, since the UI shows me for both connections the database 0. This leads to an instable system, since after you click on one of the two connections the the view stucks on "loading ...." instead of showing the connection parameters. I also tried to configure multiple connections via REDIS_HOSTS environment variable. No luck for me too. The undefined entries at the bottom are also a bit irritating. @hirnschmalz You are right, i've seen this too but i forgot to mentioned the undifined fields @tburschka The latest image solved this issue @joeferner This issue is back in the latest docker image. ๐Ÿ‘Ž Seems to be an issue with the with the change of the namespace @awearsolutions/redis-commander. It's a bit annoying to create a PR to fix a problem which is back one version later. Should be working with latest github/docker versions. At least we use it without any problem... Can issue be closed? This problem should be fixed in current version. If you think its not feel free to reopen issue with updated informations based on latest version.
2025-04-01T06:39:13.182106
2020-02-28T21:49:43
573023714
{ "authors": [ "joegasewicz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7314", "repo": "joegasewicz/pytask-io", "url": "https://github.com/joegasewicz/pytask-io/issues/28" }
gharchive/issue
Docker Container image for PytaskIO Make a PyTaskIO Docker image available to download from the Docker registry. Also update the README with the docker registry install details.
2025-04-01T06:39:13.190726
2021-12-28T23:15:57
1090171830
{ "authors": [ "joepio" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7315", "repo": "joepio/atomic-data-rust", "url": "https://github.com/joepio/atomic-data-rust/issues/264" }
gharchive/issue
v0.30.1 won't publish - issue with dependency When running cargo publish, the pin-project fails to build, because of some issue with a macro: error[E0659]: `parse_quote_spanned` is ambiguous (`macro_rules` vs non-`macro_rules` from other module) --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pin_project/derive.rs:859:67 | 859 | proj_generics.make_where_clause().predicates.push(parse_quote_spanned! { span => | ^^^^^^^^^^^^^^^^^^^ ambiguous name | note: `parse_quote_spanned` could refer to the macro defined here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/utils.rs:22:1 | 22 | / macro_rules! parse_quote_spanned { 23 | | ($span:expr => $($tt:tt)*) => { 24 | | syn::parse2(quote::quote_spanned!($span => $($tt)*)).unwrap_or_else(|e| panic!("{}", e)) 25 | | }; 26 | | } | |_^ note: `parse_quote_spanned` could also refer to the macro imported here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pin_project/derive.rs:7:5 | 7 | *, | ^ = help: use `self::parse_quote_spanned` to refer to this macro unambiguously error[E0659]: `parse_quote_spanned` is ambiguous (`macro_rules` vs non-`macro_rules` from other module) --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pinned_drop.rs:108:21 | 108 | *path = parse_quote_spanned! { path.span() => | ^^^^^^^^^^^^^^^^^^^ ambiguous name | note: `parse_quote_spanned` could refer to the macro defined here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/utils.rs:22:1 | 22 | / macro_rules! parse_quote_spanned { 23 | | ($span:expr => $($tt:tt)*) => { 24 | | syn::parse2(quote::quote_spanned!($span => $($tt)*)).unwrap_or_else(|e| panic!("{}", e)) 25 | | }; 26 | | } | |_^ note: `parse_quote_spanned` could also refer to the macro imported here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pinned_drop.rs:3:50 | 3 | use syn::{spanned::Spanned, visit_mut::VisitMut, *}; | ^ = help: use `self::parse_quote_spanned` to refer to this macro unambiguously For more information about this error, try `rustc --explain E0659`. error: could not compile `pin-project-internal` due to 2 previous errors warning: build failed, waiting for other jobs to finish... error: failed to verify package tarball This seems to be caused by this: https://github.com/taiki-e/pin-project/blob/eedd0885afc4bebad728f0adb1acffaaaed0d31a/pin-project-internal/src/pin_project/derive.rs Which seems to have been changed just days ago: https://github.com/taiki-e/pin-project/commit/ba7d5fd13f81092e64c04a30fd0dd7dd70cb87a8 And was mentioned in this issue: https://github.com/taiki-e/pin-project/issues/337 cargo update did not work. cargo publish --lockfile dit not work. I don't know what I can try next. Removing Cargo.lockl and running publish again worked.
2025-04-01T06:39:13.202385
2016-10-08T01:41:10
181796481
{ "authors": [ "joev", "lkj61" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7316", "repo": "joev/SVUI-Temp", "url": "https://github.com/joev/SVUI-Temp/issues/80" }
gharchive/issue
Getting this error with my rogue... Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value) Ok. Will look into it (will be later this evening as I'm out and about at present :)) -- Joe Vaughan On 8 October 2016 at 03:41:10, lkj61<EMAIL_ADDRESS>wrote: Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value) โ€” You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/joev/SVUI-Temp/issues/80, or mute the thread https://github.com/notifications/unsubscribe-auth/AATrWRguVwSt-gouFgw6WMK71PE4WqrFks5qxvS2gaJpZM4KRlUg . Thanks! Much appreciate all the work you are doing for us fans of SVUI. ๐Ÿ˜€ On Saturday, October 8, 2016, Joe Vaughan<EMAIL_ADDRESS>wrote: Ok. Will look into it (will be later this evening as I'm out and about at present :)) -- Joe Vaughan On 8 October 2016 at 03:41:10, lkj61<EMAIL_ADDRESS><javascript:_e(%7B%7D,'cvml','notifications@github.com');>) wrote: Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value) โ€” You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/joev/SVUI-Temp/issues/80, or mute the thread <https://github.com/notifications/unsubscribe-auth/AATrWRguVwSt- gouFgw6WMK71PE4WqrFks5qxvS2gaJpZM4KRlUg> . โ€” You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/joev/SVUI-Temp/issues/80#issuecomment-252416152, or mute the thread https://github.com/notifications/unsubscribe-auth/AJEkVZg5YQOMoY-j_UDeUkY3O-EwXyhGks5qx2q4gaJpZM4KRlUg .
2025-04-01T06:39:13.255104
2020-11-21T07:10:21
747964988
{ "authors": [ "johannrichard", "qx54" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7318", "repo": "johannrichard/homebridge-dingz", "url": "https://github.com/johannrichard/homebridge-dingz/issues/123" }
gharchive/issue
[BUG] No up-to-date light and temperature values in Homebridge anymore Describe the bug Temperature and light don't get updated in Homebridge and Home app anymore, although http://<PIR-IP>/api/v1/sensors shows up-to-date temperature and light values. To Reproduce Steps to reproduce the behavior: In the Homebridge GUI navigate to Accessories Light will have an old value Open http://<PIR-IP>/api/v1/sensors in browser Light value will be different, but correct Expected behavior The light and temperature values in Homebridge should have the same values as the ones provided by the API. Plugin environment: myStrom PIR Firmware Version: 3.8.2 Homebridge Server: Raspberry Pi 4 Plugin Version: 1.8.0 Node.js Version: 14.15.1 Homebridge Version: 1.1.6 Additional context The issue appears only after some time. Upon re-starting Homebridge it works again, but will stop working after a couple of hours. The change related to reachability (73298a58f2cfbf3dff2a12e146ab1d137b2aac64) introduced in v1.8.0 seems to break quite a few things, but only after some time running. ๐Ÿคฆ A downgrade to v1.7.1 should work around this while I work on a fix. I've released<EMAIL_ADDRESS>(homebridge-dingz@nightly) and am running it now. If you want to help with testing, feel free to install it and report back. the nightly reverts some of the changes that might have caused some of the problems encountered here and elsewhere ๐Ÿค•๐Ÿค’ I've installed 1.8.3 nightly and will report back ๐Ÿ‘ In the meantime I've worked on a v2.0.0-nightly of the plug-in. If you reinstall the nightly w/ homebridge-dingz@nightly you should be upgraded to this version. The functionality remains largely the same but the code is less cluttered and better structured so should generally run better. There might be bugs though but you can always downgrade to v1.8.3-nightly (although I think the newest nightlies are much more robust). I've had it running over longer periods of time now on my production dingz. So far I could not observe any problems with temperature/brightness values not updating. (The easiest way to test if it still works after a few hours is to use your phone's flashlight to illuminate or your hand to darken the illumination: the LUX value should change significantly ๐Ÿ˜„) It's apparently still not fixed though. ๐Ÿคฆ ๐Ÿ˜ž But I now have a clue where it happened (unreleased mutex locks on the request queue) in the code. :octocat: https://github.com/johannrichard/homebridge-dingz/blob/dee30b26ab348adb4db62c7f90f47c4ae107b684/src/dingzAccessory.ts#L1585-L1597 So far v2.0.0-nightly is working for me, but it has been running less than 24 hours.
2025-04-01T06:39:13.277595
2015-06-02T13:49:44
84032138
{ "authors": [ "johnculviner", "mike192020", "shustinho" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7319", "repo": "johnculviner/jquery.fileDownload", "url": "https://github.com/johnculviner/jquery.fileDownload/issues/98" }
gharchive/issue
Warning in Google Chrome console when generating pdf/excel document There is a warning in Google Chrome console when generating document: Resource interpreted as Document but transferred with MIME type application/pdf. Request Headers: Accept:text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,/;q=0.8 Accept-Encoding:gzip, deflate Accept-Language:en-US,en;q=0.8,ru;q=0.6 Cache-Control:max-age=0 Connection:keep-alive Content-Length:305 Content-Type:application/x-www-form-urlencoded User-Agent:Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/43.0.2357.81 Safari/537.36 Response Headers: Content-Disposition:attachment; filename=report.pdf Content-Type:application/pdf;charset=windows-1251 Date:Tue, 02 Jun 2015 13:43:47 GMT Expires:Thu, 02 Jun 2016 06:50:44 GMT Server:Apache-Coyote/1.1 Transfer-Encoding:chunked Yeah I get that too, it still works though so I think its okay IMO I get this error also for PDF's. But it will not even try to download. I change the file to download to something else (a non-pdf) but get the error and it still downloads. Also do you know of a way to suppress this error. I would like to not put the download URL out in the open. This isn't a warning in chrome its logged at info. If you are worried about your download URL being logged that's not good because anyone can view it in the chrome debugger or by monitoring http traffic. Would suggest using single use url/token/cookie if you are concerned about URL re-use. The web is not inherently secure: http://johnculviner.com/dont-trust-the-client/ understandable that there will always be a way to view the url, but I just dont want to hand it to them, would rather make them work just a little. If you have any thoughts as to suppress it that would be great. But there was still the issue of pdf's not downloading. do you happen to get this in your last push of the script? On your dont trust the client. Believe me I do not really. I treat the client as if they have some knowledge on web development so I close those holes when I can. But then I am not a security expert and cannot get everything. Though that would be nice. BTW I am in chrome.
2025-04-01T06:39:13.283145
2016-02-22T20:16:07
135531609
{ "authors": [ "billfranklin", "johnpapa" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7320", "repo": "johnpapa/lite-server", "url": "https://github.com/johnpapa/lite-server/issues/39" }
gharchive/issue
Errors Installing I have Python2.7.11 for Windows x86/x64 installed on Windows 7 x64. When I run npm install lite-server --save-dev, I get the following errors: TypeError: Request path contains unescaped characters. And a few 404 status codes downloading 32-bit node.libs. anyone else on windows able to test this? moving to https://github.com/johnpapa/lite-server/issues/54
2025-04-01T06:39:13.287994
2016-02-10T16:38:38
132752012
{ "authors": [ "cgmartin", "johnpapa", "valorkin" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7321", "repo": "johnpapa/lite-server", "url": "https://github.com/johnpapa/lite-server/pull/32" }
gharchive/pull-request
feat(bin): added run time option to config file path and name Hey, It would be really good to have an option for custom path to configuration especially if you hate dozen of files in your project root folder (like me) or have to run different configuration for different needs Thanks in advance Regards, Dima @johnpapa please review i like the idea. @cgmartin is this how you would like to do it? i prefer yargs myself. This looks good to me. I like the fact that it matches names with the browsersync --config option. Yargs is what I'm most comfortable with, but minimist is fine. There are several popular args modules out there that all score pretty high. What we're doing here is pretty simple/straightforward in terms of arg options. Easy to replace if need be. @johnpapa good to merge? Thanks for the PR @valorkin ! i havent had a chance to test it ... have you Chris? If so, Im cool with it as long as it ups the version number by a minor. Tested it, works fine. Merging and bumping minor version. Since it changes the API (not breaking), Shouldn't this be a minor bump to 2.1.0 ? (not a patch) i bumped it to 2.1.0 and published yes, minor, I fubar. Thanks for fix no worries. thanks for testing
2025-04-01T06:39:13.310288
2022-04-02T20:50:42
1190736386
{ "authors": [ "JTinkers", "yaegassy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7322", "repo": "johnsoncodehk/volar", "url": "https://github.com/johnsoncodehk/volar/issues/1139" }
gharchive/issue
Disable automatic quotation marks for attrs I would like to an option in settings to disable automatic quotation marks - we use ' but it keeps putting " and it just grinds our gears :) Duplicate of #840 and #917. Try it with HTML: Auto Create Quotes disabled.
2025-04-01T06:39:13.317865
2021-12-15T15:19:11
1081165871
{ "authors": [ "Zenthae", "johnsoncodehk", "wendyprogramic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7323", "repo": "johnsoncodehk/volar", "url": "https://github.com/johnsoncodehk/volar/issues/793" }
gharchive/issue
Absolute import i've seen that there has been an issue marked as resolved but it might have broke again because it doesn't seems to work for me. when volar auto-import a component, it's always using the relative path my config : tsconfig.json { "compilerOptions": { "baseUrl": ".", "esModuleInterop": true, "experimentalDecorators": true, "jsx": "preserve", "lib": ["esnext", "dom"], "module": "esnext", "moduleResolution": "node", "paths": { "@/*": ["src/*"], "~/*": ["src/components/*"] }, "resolveJsonModule": true, "sourceMap": true, "strict": true, "target": "esnext", "useDefineForClassFields": false, "types": ["quasar"] }, "exclude": ["node_modules"], "include": ["src/**/*.ts", "src/**/*.d.ts", "src/**/*.tsx", "src/**/*.vue"] } vite.config.ts import { quasar, transformAssetUrls } from '@quasar/vite-plugin'; import vue from '@vitejs/plugin-vue'; import { resolve } from 'path'; import { defineConfig } from 'vite'; // https://vitejs.dev/config/ export default defineConfig({ server: { fs: { allow: ['.'], }, }, resolve: { alias: [ { find: '@', replacement: resolve(__dirname, './src'), }, { find: '~', replacement: resolve(__dirname, './src/components'), }, ], }, plugins: [ vue({ template: { transformAssetUrls }, }), quasar({ autoImportComponentCase: 'pascal', sassVariables: 'src/assets/scss/variable.scss', }), ], }); Versions : "typescript": "^4.3.2", "vite": "^2.5.4", "vue-tsc": "^0.29.0", "Volar extension": "v0.29.8" You can config "typescript.preferences.importModuleSpecifier": "non-relative" in vscode. that already what i did, both in user and workspace settings, and it's still not importing using non-relative path, with or without the built-in typescript extension enabled I have the same problem, also type hinting is not working in VS Code when using an absolute import.
2025-04-01T06:39:13.346183
2022-01-24T17:22:05
1112931625
{ "authors": [ "dergigi", "dnlggr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7324", "repo": "joinmarket-webui/joinmarket-webui", "url": "https://github.com/joinmarket-webui/joinmarket-webui/pull/11" }
gharchive/pull-request
Update balance UI Updates the balance text a bit to more closely resemble the screens in Figma and the Bitcoin Design guidelines on units and symbols and on protecting a wallet. Basically, I turned on some font features according to the Figma designs and added a better symbol for when the balance is hidden. It's not perfect yet, I also want to add: The possibility to cycle through states (BTC, sats, hidden) by clicking on a balance. For that we need some global state and I didn't want to touch that quite yet. Formatting based on the userโ€™s locale. Currently I hardcoded the BTC symbol to be prefixed to the number (as it is done in the U.S. with USD amounts. However, I didn't know what to do with the sats symbol as it looked strange when prefixed to a number so I kept it as a postfix. Open for suggestions on this one! For a first version it should be fine though. Let me know what you think! Not sure how we want to organize custom CSS going forward. I used a CSS Module but am open to do it differently or just add everything to index.css for now. ๐Ÿ“ธ See also https://github.com/joinmarket-webui/joinmarket-webui/issues/19. Wonderful! Regarding the sat symbol: so far most interfaces still use "sat" or "sats," i.e. no symbol at all. Thunderhubโ€”which looks quite neat in general I have to sayโ€”uses the sat symbol, and simply puts it at the end using a slightly dimmer font color: Most other interfaces, including bitcoin and lightning wallets, simply use "sat" or "sats" either before or after the number (usually after). 1ML: LightningNetwork.plus: Updated to use the sat symbol (resolves #19). The bitcoin design guidelines don't really use a specific sat symbol so maybe we should let the user pick if they want a symbol or just "sat(s)" in the upcoming settings page. Maybe @GBKS has some input on that? For now, I'd say let's keep it simple and just go with the symbol. I also added digit groups using Intl.NumberFormat as suggested by the design guides. For now I hardcoded the formatting to the en-US locale but we could do more location specific things in the future. Regarding browser compatibility of Intl.NumberFormat: It should be available on all somewhat modern browsers. Does someone have any experience with that? And a more broad question: What browser compatibility are we targeting in general? Great! Let's keep it as is for now. We can do some polishing for v4 later on, I'm sure we can improve things here and there. Different font color for the symbol would also be easier on the eyes, for example.
2025-04-01T06:39:13.375714
2016-02-19T14:16:48
134876078
{ "authors": [ "joelwurtz", "sagikazarmark" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7325", "repo": "jolicode/jane", "url": "https://github.com/jolicode/jane/pull/23" }
gharchive/pull-request
Add support for nullable property Other way for #21 WDYT @sagikazarmark ? How do you know if a multiple type is nullable? It's done in the ObjectGuesser I don't see any change in denormalization. You use isset in denormalization, but that won't work with null values. Right, should be good now Looks good to me. Although it lacks tests, doesn't it? Yes will add a better test system following this Cool.
2025-04-01T06:39:13.392741
2022-11-25T10:59:35
1464378660
{ "authors": [ "Killajoedotcom", "jon-harper" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7326", "repo": "jon-harper/OmniBox", "url": "https://github.com/jon-harper/OmniBox/pull/71" }
gharchive/pull-request
[MCU Tray] MKS Skipr Made an MKS Skipr tray for my personal use, but wanted to share with the community. I hope this can make it into the project. MCU Tray - MKS Skipr - Heatserts.zip I made a total hash of merging this PR to the right branch, so I just cherry-picked the commits to 0.9.9-unstable instead. Let me know how the fit goes. Thanks!
2025-04-01T06:39:13.397455
2018-02-23T12:07:02
299690929
{ "authors": [ "Matrixcoffee", "goldfndr", "truedat101" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7327", "repo": "jonan/ForkHub", "url": "https://github.com/jonan/ForkHub/issues/345" }
gharchive/issue
TLS v1.2 for KitKat (and below?) With GitHub having removed support for TLS v1.0 and v1.1, ForkHub no longer works, it gives an error toast (Loading account & organizations failed). https://stackoverflow.com/questions/28943660/how-to-enable-tls-1-2-support-in-an-android-application-running-on-android-4-1 shows how to enable TLS v1.2 support for pre-Lollipop. There must be something working, since I can access GitHub on both Chrome Mobile and Lightning Browser. In the meantime, if anyone can direct me to how to get the "News" look for an arbitrary repo via web browser, I'd appreciate it. The closest I can find is Insights, but that doesn't list individual comments. square/okhttp#2372 may be instructive here. Noting also, this affects the version of Forkhub found in the Blackberry World App store, which is an Android app, I think maybe v1.2.0 (tagged release of apk), and the Android runtime provided in BB10 is 4.x era. It was a sad day when I opened the app and found that I couldn't get in. I will give a compliment, and say this is THE best performing Android app in terms of speed that I've found on the Blackberry BB10 platform. I'm willing to dig in here and help find a work around, maybe submit a pull request, since I don't have the option to upgrade my Android runtime, and find Forkhub to be essential to my daily work and my ability to be away from my desk but deal with the number of projects and companies I work with that have their stuff on GH.
2025-04-01T06:39:13.409051
2022-07-07T16:45:14
1297756534
{ "authors": [ "XiNaru", "jonas-peeters" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7328", "repo": "jonas-peeters/EasyScoreboards", "url": "https://github.com/jonas-peeters/EasyScoreboards/issues/47" }
gharchive/issue
same scoreboard for all players.... sry but a new problem. I have made a scoreboard with placeholders and for me it works. BUT my friend joined and has MY scoreboard, my name in scoreboard, my ranks, my money....not his own. sry this is a discord screenshot Ok, I have released a new version. Could you check out if the issue is still present there? no the bug isn fixed. NOW a new player joined the game, he has his own scoreboard but NOW I see his scoreboard iam xinaru he is Zeus or must i delete the old config? okay the bug isn fixed i have delete the old config and add the lines new. but the scoreboard dont show the own scoreboard always. Ok, I believe it has to do with the order the players join. I will have another look at it Alright, I have updated some logic concerning the buffering of scoreboards. In addition, this version will log if placeholder usage is detected. You should not need to delete the config file. If the issue persists, please check in the server logs if it is says "Detected use of placeholders" or "Detected no use of placeholders" EasyScoreboard-2.5.3.jar.zip sry but the problem ist there with the new version too. 08.07 17:27:59 [Server] Server thread/ERROR [Sponge]: The Scheduler tried to run the task de_yottaflops_easyscoreboard-S-3 owned by Plugin{id=de_yottaflops_easyscoreboard, name=EasyScoreboards, version=2.5.3, description=A plugin to easily create scoreboards for lobbys, etc., authors=[YottaFLOPS], source=./mods/EasyScoreboard-2.5.3.jar}, but an error occured. now the plugin is crashed Ok, this was probably due to another change I made. I apologize. Try with this version, if you could. EasyScoreboard-2.5.3.jar.zip รคhm...i use 2.5.3? you have uplรถoad this above? this version i use and the error comes It should be a new build with the same version number ah okay i try it okay tried. scoreboard there but my fried has MY scoreboard 08.07 19:52:04 [Server] Server thread/INFO [EasyScoreboard: Config]: Detected use of placeholders I believe that I have finally fixed the problem and in my testing it worked. Let me know if the problem persists for you. EasyScoreboard-2.5.3-2.jar.zip This time it looks like it's finally going to work. Just a friend and I on the server, everyone had their own sb. we are both separately leased and jointed it was still each his own sb. the only thing i forgot to test is the /esb reload command. Amazing. Thanks for your help debugging this issue and your patience. I have created a new release here but it should be identical (apart from the version number) to the .jar posted above. Thanks again
2025-04-01T06:39:13.416900
2024-07-15T10:34:30
2408403753
{ "authors": [ "inyong1", "zenzzenpl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7329", "repo": "jonataslaw/getx", "url": "https://github.com/jonataslaw/getx/issues/3146" }
gharchive/issue
Navigating to New Product Details Page with Different Product ID Fails to Load New Data Question I am developing an e-commerce app using Flutter with GetX for state management. I have a ProductDetailsView that displays product details using a named route and arguments. In the ProductDetailsController, I load product details based on the product ID passed through arguments. Here is my controller code: import 'package:e_commerce/api/send_request.dart'; import 'package:e_commerce/app/data/color_collections.dart'; import 'package:e_commerce/app/widgets/theme_mode.dart'; import 'package:get/get.dart'; import 'package:webview_flutter/webview_flutter.dart'; class ProductDetailsController extends GetxController { WebViewController? webViewController; var webViewHeight = 100.0.obs; var productDetailsLoading = true.obs; var productDetails = <String, dynamic>{}.obs; var productId = 0.obs; @override void onInit() { super.onInit(); productId.value = Get.arguments['product_id']; loadPage(); } @override void onClose() { super.onClose(); productId.close(); } void loadPage() async { productDetailsLoading.value = true; var resp = await SendRequest("/Products/Manager/CRUD/GetProductDetails", data: { "product_id": productId.value, }); productDetails.value = resp["product_details"]; productDetailsLoading.value = false; if (webViewController == null) { webViewController = WebViewController() ..setJavaScriptMode(JavaScriptMode.unrestricted) ..setBackgroundColor(isDarkMode() ? DynamicColorsCollection.backgroundColorDark : DynamicColorsCollection.backgroundColorLight) ..setNavigationDelegate( NavigationDelegate( onProgress: (int progress) {}, onPageStarted: (String url) {}, onPageFinished: (String url) async { final result = await webViewController?.runJavaScriptReturningResult( "document.documentElement.scrollHeight.toString();"); String heightStr = result as String; double height = double.parse(heightStr.replaceAll('"', '')); updateHeight(height); }, onHttpError: (HttpResponseError error) {}, onWebResourceError: (WebResourceError error) {}, onNavigationRequest: (NavigationRequest request) { if (request.url.startsWith('https://www.youtube.com/')) { return NavigationDecision.navigate; } return NavigationDecision.prevent; }, ), ) ..loadRequest(Uri.parse( "http://some-domain.com/products/description/${resp["product_details"]["product_url"]}")); } else { webViewController?.loadRequest(Uri.parse( "http://some-domain.com/products/description/${resp["product_details"]["product_url"]}")); } } void updateHeight(double height) { webViewHeight.value = height; } } Here is the start of my view code: class ProductDetailsView extends GetView<ProductDetailsController> { const ProductDetailsView({Key? key}) : super(key: key); @override Widget build(BuildContext context) { Get.create<ProductDetailsController>(() => ProductDetailsController()); return Obx(() { if (controller.productDetailsLoading.value) { return Material( color: isDarkMode() ? DynamicColorsCollection.backgroundColorSecondaryDark : DynamicColorsCollection.backgroundColorSecondaryLight, child: LoadingShimmer(), ); } return CustomScaffold( backgroundColor: isDarkMode() ? DynamicColorsCollection.backgroundColorDark : DynamicColorsCollection.backgroundColorLight, The issue arises when I navigate to a new product details page from the current product details page using a button. The button's code is as follows: Get.toNamed( Routes.PRODUCT_DETAILS, arguments: { "product_id": id, }, preventDuplicates: false, ); When I navigate like this, the new product details page does not load the new product data. Instead, it shows the data of the previous product. How can I ensure that navigating to a new product details page with a different product ID loads the new product's data correctly? I tried navigating to the new product details page from the current product details page by using the Get.toNamed method with the new product ID as an argument. I expected the new product details page to load and display the details of the new product corresponding to the new product ID. However, instead of loading the new product's data, the page continued to display the data of the previously viewed product. I expected the ProductDetailsController to fetch and show the details of the new product based on the updated product ID passed in the arguments, but this did not happen. Because onInit olny called once, Mayby you can use GetBuilder and create new controller instance each the page is called try add global: false, to the GetBuilder Thanks, That worked ๐Ÿ‘๐Ÿป๐Ÿ‘๐Ÿป๐Ÿ‘๐Ÿป๐Ÿ‘๐Ÿป๐Ÿ‘๐Ÿป๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ