id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
|---|---|---|---|---|---|
179613632
|
Triangle subdivide
Also contains several fixes in the kdtree. Therefore lets just merge it onto master and continue from there with radiosity and subdivision.
Coverage decreased (-1.8%) to 46.033% when pulling 336b657c9688c05d9955cf9858efea78b0012551 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Coverage increased (+1.5%) to 49.365% when pulling 3c692fc1febc132889896250b508f498c82fbc50 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Coverage decreased (-1.8%) to 46.033% when pulling 21452b4ed6144a3a83c14069051445b5dc3fb054 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Coverage decreased (-1.05%) to 46.779% when pulling f8ed7d1cda90710966a28282d8b55675233e7bb7 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Coverage decreased (-1.8%) to 46.033% when pulling 71915703b578fd40ccbb0a29587aac297e58c4ff on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Keine Ahnung wie man die nicht abgedeckten Zeilen mit Test zu covern sind.
KDTree::build kannst du ja mit einer kleinen Box aufrufen https://coveralls.io/builds/8119862/source?filename=lib%2Fkdtree.cpp#L101
So coverst du box.surface_area() == 0.
Das hier würde ich eine Hilfsmethode packen, da die Methode schon sehr lang ist.
Coverage decreased (-1.7%) to 46.082% when pulling df22071f2f522f1af3433c326cc8590b0a5b6ff6 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Coverage decreased (-1.1%) to 46.708% when pulling e0ff81e96e487819891ff520c4c524cd7b741839 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Coverage decreased (-0.7%) to 47.126% when pulling 01d71371778cdff9edf123721bb478ffa13aaa09 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master.
Tests done. Coverage decreased weil wir radiohit erweitert haben und da nichts testen.
Awesome work!
|
gharchive/pull-request
| 2016-09-27T21:05:24
|
2025-04-01T06:39:12.190102
|
{
"authors": [
"boxdot",
"coveralls",
"jeschkies"
],
"repo": "jeschkies/renderer",
"url": "https://github.com/jeschkies/renderer/pull/60",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2486632383
|
Project group and some optimization for faster retrieval of workitems.
I do this PR to add my personal use case (project group) as it can be useful.
It's only here as a workitem retriever for now but can be largely modified.
It's implemented so it's not more complicated to use than a project.
I believe that my code is clean, but any review is very welcomed as I'm not an expert.
Looks good to me, never used it but seems like a nice addition. Also thanks for the cleanup.
I'll merge it if you could resolve the merge conflict please
|
gharchive/pull-request
| 2024-08-26T11:29:41
|
2025-04-01T06:39:12.194048
|
{
"authors": [
"WH-Yoshi",
"jesper-raemaekers"
],
"repo": "jesper-raemaekers/python-polarion",
"url": "https://github.com/jesper-raemaekers/python-polarion/pull/181",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
101102747
|
Allow a notation that's just "This belongs to a certain class". Partial types, I suppose.
So the original idea here was to take generics, and impose a constraint on them. It wouldn't be a complete constraint, only something that would say "Anything of this particular kind of a class". It was something like A > function specification for generics.
There are two problems with this approach:
How Lily distinguish between someone that wants a partial type constraint, and someone who goofed. I'd like to avoid a symbol soup here, so A > ~function doesn't set well.
How does one create a list of partially typed values, since the constraint is against a generic? Uhh...You kinda don't. Because you can't refer to generics outside of a generic function.
So, on to the new idea. Let's say you're trying to make a list of those partially a function things. It's gonna look like this:
var v: list[~function] = [list::select, integer::to_s, string::split]
So now you've got a list of things that are a function. That's what the tilde is for: It says "these things are all just a function. I don't know any more about them." Since there's an annotation for ~function supplied, Lily will create a list of kinda-functions. Otherwise, it will retain the same behavior. The reason for this, is that I don't like that sort of behavior (unifying in a way that reduces away all subtype information) without explicit intent.
Lily understands variance, and will also allow functions to specify that they want a partial function. Also, with that in mind, Lily should have no problem determining that any particular kind of a function with arguments or a return type of any sort is a function. Lily having a better understanding of variance was also a motivation for this change.
Initially, only functions will be able to be specified as being partial (no subtype information). However, in the future, I'll extend this if it turns out there's a use for it on other types.
I feel like an idiot when reading this back to myself now. What I'm trying to solve can easily be solved by having a callable type which would act as a parent type to what function is. There would be no need for any of the above to be done, as Lily understands when doing variance to follow the parent's generic count. So just adding a callable type and maybe refining a couple areas is all that's necessary. I'll get around to it when I add the doc thing specified by #127. No need for a whole bug and a new syntax. Besides, I can only see adding what this asks for as being a problem later (Valid questions like "Why not partial lists, partial hashes, partial higher kinds begin showing up). So there's no need for a special bug for this, I think.
|
gharchive/issue
| 2015-08-14T21:01:25
|
2025-04-01T06:39:12.205623
|
{
"authors": [
"jesserayadkins"
],
"repo": "jesserayadkins/lily",
"url": "https://github.com/jesserayadkins/lily/issues/126",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
175508657
|
How to set Image and text in same Bubble cell
New issue checklist
[x] I have read all of the README, documentation, and FAQ.
[x] I have reviewed the contributing guidelines. Confirmation: ____
[x] I have searched existing issues and this is not a duplicate.
General information
Library version(s):
iOS version(s):
Devices/Simulators affected:
Reproducible in the demo project? (Yes/No):
Related issues:
Bug report
Expected behavior
Actual behavior
Steps to reproduce
Crash log? Screenshots? Videos? Sample project?
Thanks! :smile:
Please provide all the information in the issue template so we can help!
And don't forget to read our contribution guidelines. :blush:
|
gharchive/issue
| 2016-09-07T13:54:24
|
2025-04-01T06:39:12.211502
|
{
"authors": [
"Komal2905",
"jessesquires"
],
"repo": "jessesquires/JSQMessagesViewController",
"url": "https://github.com/jessesquires/JSQMessagesViewController/issues/1820",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2101635068
|
🛑 Boucherie La Lienne is down
In 48f4091, Boucherie La Lienne (https://boucherielalienne.ch) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Boucherie La Lienne is back up in dc653b3 after 5 minutes.
|
gharchive/issue
| 2024-01-26T05:53:24
|
2025-04-01T06:39:12.216821
|
{
"authors": [
"jessica-habaneros"
],
"repo": "jessica-habaneros/uptime",
"url": "https://github.com/jessica-habaneros/uptime/issues/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1868560569
|
[Bug]: node18 memory consumption growth up to25%
Version
29.6.4
Steps to reproduce
This happens on a NX powered private repo - on demand I can produce a reproduction repo if needed
Expected behavior
with node16 our test suite takes about 2GB memory in run in band mode
Actual behavior
with node18 our test suite takes about 2.6GB memory and has hard time to finish on CircleCI
Additional context
During memory leak hunt I can see that string allocation loads the same modules several times leading to significant growth, shouldn't this in --runInBand mode never happen with the given cache/resolver?
Environment
System:
OS: macOS 13.3.1
CPU: (8) arm64 Apple M1 Pro
Binaries:
Node: 18.17.0 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/node
Yarn: 3.6.2 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/yarn
npm: 9.6.7 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/npm
Update:
I've tried --no-cache and this had no effect on loaded modules
Probably related to https://github.com/jestjs/jest/issues/11956
Yeah, dupe of that one
|
gharchive/issue
| 2023-08-27T16:53:05
|
2025-04-01T06:39:12.220662
|
{
"authors": [
"Delapouite",
"SimenB",
"vire"
],
"repo": "jestjs/jest",
"url": "https://github.com/jestjs/jest/issues/14454",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
341296543
|
failing beforeAll() causes even passing tests in the scope to fail
🐛 Bug Report
I'm looking for the best practice for aborting a describe block if the test setup fails. For example, when testing an API, if authenticating fails, it's pointless to run any other tests.
As suggested in the issue template, I searched StackOverflow first, where I found this incomplete answer, advising to place the initialization code in a beforeAll block. Hence, raising the issue here because
I haven't seen anything documenting the practice of returning failure from beforeAll
If a beforeAll block fails, tests in that describe block are still run, and they fail, even if otherwise they would pass!
To Reproduce
describe('test that a 3rd party API remains consistent', () => {
beforeAll(() => expect('login').toBe('successful')); // this will fail
test('API function 1', () => expect(1).toBe(1)); // each...
test('API function 2', () => expect(2).toBe(2)); // ...of these...
test('API function 3', () => expect(3).toBe(3)); // ...will fail too
});
Expected behavior
Jest should report that the beforeAll() failed, and bail the describe scope without executing further tests in it. If it's somehow intended behavior that all tests should still be executed, and still marked as failed (which I find odd), this should be documented under beforeAll and in the Setup and Teardown guide. Other tests in the file, outside of the failing block, should still be executed.
Link to repl or repo (highly encouraged)
https://repl.it/@DanDascalescu/beforeAll-failure-should-bail-the-test
Run npx envinfo --preset jest
System:
OS: Linux 4.15 Ubuntu 16.04.4 LTS (Xenial Xerus)
CPU: x64 Intel(R) Core(TM) i7-7500U CPU @ 2.70GHz
Binaries:
Node: 10.5.0 - /usr/local/bin/node
Yarn: 1.7.0 - /usr/bin/yarn
npm: 6.2.0 - /usr/local/bin/npm
npmPackages:
jest: ^23.4.1 => 23.4.1
See also
#6527
@aaronabramov have we changed the behaviour here for circus? Bailing after a failing lifecycle hook makes sense to me
yes! this was a bug in jasmine (or a feature). We changed this behavior in jest-circus.
@dandv jest-circus isn't shipped with jest by default yet, but you can test it by yarn add jest-circus and adding "testRunner": "jest-circus/runner" to your jest config
I would like to add to this a request for named hooks. The test is marked as a failure and not run, but named hooks (especially when the hooks reside outside of the test file) would aid in debugging.
@alycda feel free to open up a separate issue for that 🙂
Closing this as the behavior is fixed in jest-circus
Closing this as the behavior is fixed in jest-circus
It doesn't seem to be fixed in latest Jest (v23.6.0). So it's fixed for what users exactly? Facebook internal?
See the comment above: https://github.com/facebook/jest/issues/6695#issuecomment-405326998
Install circus (which will become the default in Jest at some point) and you should be good. Also see #7198
Install circus (which will become the default in Jest at some point)
Thanks I'll follow. Still why then it's not set as default? Are there any other issues we should be aware of? Is jest-circus considered stable?
You can follow #6295 for state of making it the default.
jest-circus seems to not fix this. All 3 tests still run and fail.
Here's the command I ran:
$ yarn run jest --testRunner=jest-circus/runner --bail
Am I missing a configuration value somewhere?
This is still as broken as ~1yr ago, even with circus... Please consider reopening.
To my eyes this just looks like an issue with the reporting. beforeAll is only run once and none of the other tests actually execute, but they all fail with the same reason, and they all print the same error.
We should probably avoid printing the individual tests if beforeAll fails. Right now the implementation just checks if we've had an error and if yes, print it and fail the test
I've seen a similar issue with reporting. It's hard to disentangle from proprietary code, but what happens is that a test.todo('description', callback); reports the MongoDB error thrown by the previously executed test (a bulk write failure) instead of the correct error that only a description should be specified for .todo tests. I tried throwing an error directly in that failing test but that doesn't reproduce the behavior.
Happy to see if the .todo issue reproduces once this is fixed.
@SimenB the issue seems to be only partially fixed in circus. since nested hooks are still being executed.
To my eyes this just looks like an issue with the reporting. beforeAll is only run once and none of the other tests actually execute, but they all fail with the same reason, and they all print the same error.
We should probably avoid printing the individual tests if beforeAll fails. Right now the implementation just checks if we've had an error and if yes, print it and fail the test without actually executing it
@SimenB Should we create a separate issue for this? Or was the intention to re-open this to continue the work in this one?
Sure, a new issue about the state today and what we want the end state to be makes sense to me
Is it a duplicate of #577?
@pkuczynski I think so, yes.
Still not working as expected... Not via throwing an error or calling done.fail(err)
Definitely not stale, still see this issue in latest jest.
Undigging topic - imo. test suite should interrupt tests execution with appropriate message when one of hooks will fail or timeout (beforeEach/All, afterEach/All).
Unstale
As a present for this issue's 5th birthday, I would like to ask a maintainer to reopen #2713, because throwing from beforeAll should really stop executing any further tests. @SimenB?
It seems this framework is totally abandonned... RIP!
|
gharchive/issue
| 2018-07-15T06:36:23
|
2025-04-01T06:39:12.238095
|
{
"authors": [
"JaneJeon",
"SimenB",
"aaronabramov",
"alycda",
"chrismwendt",
"dandv",
"denissabramovs",
"fider",
"medikoo",
"pkuczynski",
"svyandun",
"thernstig",
"vtgn"
],
"repo": "jestjs/jest",
"url": "https://github.com/jestjs/jest/issues/6695",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2475719495
|
compile frontend in oldSchool, update task info
In #69 there was uncompiled frontend, also task-info.yaml files didn't reflect the new changes. This is fixed
@Daoortor could you please fix Qodana?
|
gharchive/pull-request
| 2024-08-20T13:53:07
|
2025-04-01T06:39:12.247011
|
{
"authors": [
"Daoortor",
"nbirillo"
],
"repo": "jetbrains-academy/kotlin-onboarding-collections",
"url": "https://github.com/jetbrains-academy/kotlin-onboarding-collections/pull/70",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
256651489
|
Supporting auto-generation of CA issuer keypair
#79 adds a basic CA issuer that reads a signing keypair from a Secret in the Kubernetes API server in order to issue certificates.
For convenience, it may be desirable to support an 'automatically generate a signing keypair' mode.
ref #85 #84
Instead of adding a 'generate CA' field to the Issuer, I think we should look at adding a selfsigned certificate issuer and adding an IsCA field to the Certificate resource type (as described in #85 #84)
Closing in favour of #84
|
gharchive/issue
| 2017-09-11T10:15:59
|
2025-04-01T06:39:12.262339
|
{
"authors": [
"munnerz"
],
"repo": "jetstack-experimental/cert-manager",
"url": "https://github.com/jetstack-experimental/cert-manager/issues/83",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
443287523
|
Cert manager stuck at attempting to acquire leader lease
Describe the bug:
Have been running with cert manager for a year now, but suddenly, it is being stuck at "attempting to acquire leader lease", with no error messages. It is working fine on a similar cluster, with the same settings.
Environment details::
Kubernetes version (e.g. v1.10.2): 1.11.8-gke.6
Cloud-provider/provisioner (e.g. GKE, kops AWS, etc): GKE
cert-manager version (e.g. v0.4.0): v0.6.2
Install method (e.g. helm or static manifests): helm
/kind bug
After reinstall of helm chart, and some patience, then it worked.
|
gharchive/issue
| 2019-05-13T09:11:49
|
2025-04-01T06:39:12.265136
|
{
"authors": [
"jonasdkhansen"
],
"repo": "jetstack/cert-manager",
"url": "https://github.com/jetstack/cert-manager/issues/1669",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
305999446
|
Issue certificate using dns01 via route53 stuck on SelfCheck status
Is this a BUG REPORT or FEATURE REQUEST?:
Uncomment only one, leave it on its own line:
/kind bug
/kind feature
What happened:
Trying to configure dns01 route53 provider and it works using staging letsencrypt ClusterIssuer. When changing ClusterIssuer to live, it stuck on status
Normal PrepareCertificate 15m cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 15m cert-manager-controller Presenting dns-01 challenge for domain auth-service-trunk.gel.net
Normal PresentChallenge 15m cert-manager-controller Presenting dns-01 challenge for domain auth.test.gel.tech
Normal SelfCheck 14m cert-manager-controller Performing self-check for domain auth-service-trunk.gel.net
Normal SelfCheck 14m cert-manager-controller Performing self-check for domain auth.test.geo.tech
I've checked route 53 and I see there _acme-challenge. TXT records for both domains.
Same live ClusterIssuer works as expected with http01 provider.
In log no errors. How to understand what wrong? Can I enable debug mode in some way?
What you expected to happen:
Successfully issued certificate
How to reproduce it (as minimally and precisely as possible):
Anything else we need to know?:
Environment:
Kubernetes version (use kubectl version):
Server Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.8", GitCommit:"2f73858c9e6ede659d6828fe5a1862a48034a0fd", GitTreeState:"clean", BuildDate:"2018-02-09T21:23:25Z", GoVersion:"go1.8.3", Compiler:"gc", Platform:"linux/amd64"}
Cloud provider or hardware configuration**:
AWS
Install tools:
Installed via helm from https://github.com/kubernetes/charts/tree/master/stable/cert-manager
Others:
Hi @argusua is your cluster DNS able to resolve external domain names?
cert-manager with be trying to fetch the _acme-challenge' using the DNS provided by the cluster to the cert-manager` pod. If that DNS does not allow Internet DNS look-ups, that will fail.
If you have this problem, then from k8s 1.9 you can specify custom DNS servers for each Pod, so you could add an Internet DNS server like Google 8.8.8.8 to the cert-manager pod.
https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pods-dns-config
I've noticed the similar issue. Everything was working 4 days ago but when I applied a new certificate resource I got the following error.
Error preparing issuer for certificate: error waiting for key to be available for domain
cert-manager pod logs
I0316 14:00:52.147870 1 server.go:68] Listening on http://0.0.0.0:9402
I0316 14:00:52.153768 1 leaderelection.go:174] attempting to acquire leader lease...
I0316 14:00:52.230488 1 leaderelection.go:184] successfully acquired lease kube-system/cert-manager-controller
I0316 14:03:07.518985 1 controller.go:138] clusterissuers controller: syncing item 'letsencrypt-production'
I0316 14:03:10.273963 1 helpers.go:122] Setting lastTransitionTime for ClusterIssuer "letsencrypt-production" condition "Ready" to 2018-03-16 14:03:10.273947464 +0000 UTC m=+138.299805182
I0316 14:03:10.279701 1 controller.go:152] clusterissuers controller: Finished processing work item "letsencrypt-production"
I0316 14:03:10.279766 1 controller.go:138] clusterissuers controller: syncing item 'letsencrypt-production'
I0316 14:03:10.991498 1 controller.go:152] clusterissuers controller: Finished processing work item "letsencrypt-production"
I0316 14:04:19.614244 1 controller.go:187] certificates controller: syncing item 'test/test-certificate'
I0316 14:04:19.615050 1 sync.go:107] Error checking existing TLS certificate: secret "test-certificate" not found
I0316 14:04:19.615205 1 sync.go:209] Preparing certificate with issuer
I0316 14:04:19.616105 1 prepare.go:230] Compare "" with "https://acme-v01.api.letsencrypt.org/acme/reg/31265273"
I0316 14:08:14.093642 1 helpers.go:165] Setting lastTransitionTime for Certificate "test-certificate" condition "Ready" to 2018-03-16 14:08:14.093626634 +0000 UTC m=+442.119484352
I0316 14:08:14.093733 1 sync.go:213] Error preparing issuer for certificate: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded
E0316 14:08:14.110479 1 sync.go:190] [test/test-certificate] Error getting certificate 'test-certificate': secret "test-certificate" not found
Certificate
apiVersion: certmanager.k8s.io/v1alpha1
kind: Certificate
metadata:
name: test-certificate
namespace: test
spec:
secretName: test-certificate
issuerRef:
name: letsencrypt-production
kind: ClusterIssuer
dnsNames:
- test.mydomain.com
acme:
config:
- dns01:
provider: route53
domains:
- test.mydomain.com
Cluster Issuer
apiVersion: certmanager.k8s.io/v1alpha1
kind: ClusterIssuer
metadata:
name: letsencrypt-production
spec:
acme:
server: https://acme-v01.api.letsencrypt.org/directory
email: myemail@mydomain.com
privateKeySecretRef:
name: letsencrypt-production
dns01:
providers:
- name: route53
route53:
accessKeyID: AKIAXXXXXXXXXXXXX
secretAccessKeySecretRef:
name: route53-config
key: secret-access-key
region: us-east-1
Certificate detailed description
Name: test-certificate
Namespace: test
Labels: <none>
Annotations: <none>
API Version: certmanager.k8s.io/v1alpha1
Kind: Certificate
Metadata:
Cluster Name:
Creation Timestamp: 2018-03-16T14:08:31Z
Generation: 0
Resource Version: 33688876
Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/test/certificates/test-certificate
UID: 9rf2b603-98uj-11e8-8521-1239c7046d88
Spec:
Acme:
Config:
Dns 01:
Provider: route53
Domains:
test.mydomain.com
Common Name:
Dns Names:
test.mydomain.com
Issuer Ref:
Kind: ClusterIssuer
Name: letsencrypt-production
Secret Name: test-certificate
Status:
Acme:
Authorizations: <nil>
Conditions:
Last Transition Time: 2018-03-16T14:19:33Z
Message: Error checking ACME domain validation: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded
Reason: ErrCheckAuthorization
Status: False
Type: Ready
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning ErrorPrepareCertificate 26m (x29 over 5h) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded
Warning ErrorCheckCertificate 10m (x30 over 5h) cert-manager-controller Error checking existing TLS certificate: secret "test-certificate" not found
Normal PrepareCertificate 10m (x30 over 5h) cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 10m (x30 over 5h) cert-manager-controller Presenting dns-01 challenge for domain test.mydomain.com
Normal SelfCheck 9m (x30 over 5h) cert-manager-controller Performing self-check for domain test.mydomain.com
TXT record was created in Route53 hosted zone and it was publicly accessible
cert-manager: v0.2.3
kubernetes: v1.7.9
cloud provider: AWS
Hi @whereisaaron
Thanks for fast response. Looks like yes, dns resolution works inside cert-manager container:
Also, today status changed for certificate to:
Status:
Acme:
Authorizations: <nil>
Conditions:
Last Transition Time: 2018-03-16T15:12:40Z
Message: Error checking ACME domain validation: [error waiting for authorization for domain "auth.test.gel.tech": context canceled, error waiting for authorization for domain "auth-service-trunk.gel.net": context canceled]
Reason: ErrCheckAuthorization
Status: False
Type: Ready
Same here:
$ kubectl get clusterissuer -oyaml
apiVersion: v1
items:
- apiVersion: certmanager.k8s.io/v1alpha1
kind: ClusterIssuer
metadata:
clusterName: ""
creationTimestamp: 2018-03-18T11:00:27Z
generation: 0
name: letsencrypt-prod
namespace: ""
resourceVersion: "3499055"
selfLink: /apis/certmanager.k8s.io/v1alpha1/letsencrypt-prod
uid: 90d32342-2a9b-11e8-ae12-4ccc6ad2499b
spec:
acme:
dns01:
providers:
- name: aws
route53:
accessKeyID: AK232AHSKGRLPPABQ
hostedZoneID: ""
region: eu-central-1
secretAccessKeySecretRef:
key: secret-access-key
name: route53-credentials-secret
email: info@example.com
privateKeySecretRef:
key: ""
name: ingress-key-pair
server: https://acme-staging.api.letsencrypt.org/directory
status:
acme:
uri: https://acme-staging.api.letsencrypt.org/acme/reg/232323
conditions:
- lastTransitionTime: 2018-03-18T11:00:28Z
message: The ACME account was registered with the ACME server
reason: ACMEAccountRegistered
status: "True"
type: Ready
kind: List
metadata:
resourceVersion: ""
selfLink: ""
kubectl describe certificate
Name: rig2-tls
Namespace: default
Labels: <none>
Annotations: <none>
API Version: certmanager.k8s.io/v1alpha1
Kind: Certificate
Metadata:
Cluster Name:
Creation Timestamp: 2018-03-18T11:02:58Z
Generation: 0
Owner References:
API Version: extensions/v1beta1
Block Owner Deletion: true
Controller: true
Kind: Ingress
Name: rig2-zcash-miner
UID: 95c1ed1c-2a2a-11e8-ae12-4ccc6ad2499b
Resource Version: 3499961
Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/default/certificates/rig2-tls
UID: eb1acec6-2a9b-11e8-ae12-4ccc6ad2499b
Spec:
Acme:
Config:
Dns 01:
Provider: aws
Domains:
rigs.umine.farm
Common Name:
Dns Names:
rigs.umine.farm
Issuer Ref:
Kind: ClusterIssuer
Name: letsencrypt-prod
Secret Name: rig2-tls
Status:
Acme:
Authorizations: <nil>
Conditions:
Last Transition Time: 2018-03-18T11:06:38Z
Message: Error checking ACME domain validation: error waiting for key to be available for domain "rigs.umine.farm": context deadline exceeded
Reason: ErrCheckAuthorization
Status: False
Type: Ready
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal SelfCheck 7m (x5 over 21m) cert-manager-controller Performing self-check for domain rigs.umine.farm
Warning ErrorCheckCertificate 4m (x6 over 22m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-tls" not found
Normal PrepareCertificate 4m (x6 over 22m) cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 4m (x6 over 22m) cert-manager-controller Presenting dns-01 challenge for domain rigs.umine.farm
Warning ErrorPrepareCertificate 39s (x6 over 18m) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "rigs.umine.farm": context deadline exceeded
and route53 record is created and working well
My ingress:
ingress:
enabled: true
annotations: #{}
kubernetes.io/ingress.class: nginx
nginx.ingress.kubernetes.io/rewrite-target: '/'
kubernetes.io/tls-acme: "true"
certmanager.k8s.io/acme-challenge-type: 'dns01'
certmanager.k8s.io/acme-dns01-provider: 'aws'
# nginx.ingress.kubernetes.io/ssl-redirect: "false"
path: /rig2
hosts:
- rigs.umine.farm
tls: #[]
- secretName: rig2-tls
hosts:
- rigs.umine.farm
@whereisaaron thanks for the tip regarding dns. I can confirm where I exec into cert manager and try pining that address it fails. External addresses are ok, but I configured my cluster woith the same name as dns address, so I guess that's why the query doesn't go to upstream server which is 8.8.8.8
Not fully fixed. Got it working for rigs.umine.farm and got fake certificate from staging acme server. Changed server to prod and was hoping to get a new certificate (proper one) but that didn't work. Changed DNS to another name and started getting errors during self-check again.
Warning ErrorCheckCertificate 17m (x2 over 20m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-cert" not found
Normal PrepareCertificate 17m (x2 over 20m) cert-manager-controller Preparing certificate with issuer
Warning ErrorPrepareCertificate 17m cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "monitor.umine.farm": context deadline exceeded
Normal PresentChallenge 17m (x2 over 20m) cert-manager-controller Presenting dns-01 challenge for domain monitor.umine.farm
Normal SelfCheck 16m (x2 over 19m) cert-manager-controller Performing self-check for domain monitor.umine.farm
Warning ErrorCheckCertificate 1m (x5 over 15m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-cert" not found
Normal PrepareCertificate 1m (x5 over 15m) cert-manager-controller Preparing certificate with issuer
Warning ErrorPrepareCertificate 1m (x4 over 11m) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "monitor.umine.farm": context deadline exceeded
Normal PresentChallenge 1m (x5 over 15m) cert-manager-controller Presenting dns-01 challenge for domain monitor.umine.farm
Normal SelfCheck 1m (x5 over 14m) cert-manager-controller Performing self-check for domain monitor.umine.farm
'context deadline exceeded' means the HTTP request timed out. With the range of stuff you have reported my guess would be you've having some connectivity issues with Internet requests from your cluster. It looks like your cert-manager is fine, but it has, perhaps intermittent, trouble making DNS requests and making HTTP requests to the Internet to check the challenges are in place.
Looks like DNS and HTTP request are correct
Request performed from the cert-manager container:
dig acme-v01.api.letsencrypt.org
; <<>> DiG 9.11.2-P1 <<>> acme-v01.api.letsencrypt.org
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 48485
;; flags: qr rd ra; QUERY: 1, ANSWER: 3, AUTHORITY: 0, ADDITIONAL: 1
;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 4096
;; QUESTION SECTION:
;acme-v01.api.letsencrypt.org. IN A
;; ANSWER SECTION:
acme-v01.api.letsencrypt.org. 60 IN CNAME api.letsencrypt.org-ng.edgekey.net.
api.letsencrypt.org-ng.edgekey.net. 60 IN CNAME e14990.dscx.akamaiedge.net.
e14990.dscx.akamaiedge.net. 20 IN A 104.95.238.173
;; Query time: 6 msec
;; SERVER: 10.3.0.10#53(10.3.0.10)
;; WHEN: Mon Mar 19 10:15:00 UTC 2018
;; MSG SIZE rcvd: 158
curl https://acme-v01.api.letsencrypt.org/directory
{
"key-change": "https://acme-v01.api.letsencrypt.org/acme/key-change",
"meta": {
"terms-of-service": "https://letsencrypt.org/documents/LE-SA-v1.2-November-15-2017.pdf"
},
"new-authz": "https://acme-v01.api.letsencrypt.org/acme/new-authz",
"new-cert": "https://acme-v01.api.letsencrypt.org/acme/new-cert",
"new-reg": "https://acme-v01.api.letsencrypt.org/acme/new-reg",
"rYI1xLOBxqk": "https://community.letsencrypt.org/t/adding-random-entries-to-the-directory/33417",
"revoke-cert": "https://acme-v01.api.letsencrypt.org/acme/revoke-cert"
}
As I said before, nothing has been changed in our k8s cluster or cert-manager configuration and everything was working a week ago.
@simonkey007 cert-manager isn't trying to access letsencrypt.org, it is trying to access the domain for the certificate being issued, to check that the challenge record is in place. e.g. 'umine.farm'.
Still can't get it work for my live ingress. I've tried today create test ingress and it works as expected.
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
name: lets-test
annotations:
kubernetes.io/ingress.class: "nginx"
kubernetes.io/tls-acme: "true"
certmanager.k8s.io/cluster-issuer: "letsencrypt-live"
certmanager.k8s.io/acme-challenge-type: "dns01"
certmanager.k8s.io/acme-dns01-provider: "route53"
spec:
tls:
- secretName: test-crt-tls
hosts:
- lets.test.gelato.tech
- letstest.gelatogroup.net
rules:
- host: lets.test.gelato.tech
http:
paths:
- path: /
backend:
serviceName: jenkins
servicePort: 80
- host: letstest.gelatogroup.net
http:
paths:
- path: /
backend:
serviceName: jenkins
servicePort: 80
For live ingress
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
annotations:
certmanager.k8s.io/acme-challenge-type: dns01
certmanager.k8s.io/acme-dns01-provider: route53
certmanager.k8s.io/cluster-issuer: letsencrypt-live
kubernetes.io/ingress.class: nginx
kubernetes.io/tls-acme: "true"
creationTimestamp: 2018-03-16T07:19:07Z
generation: 269
name: auth-service-kubernetes-live-nginx
namespace: default
resourceVersion: "119221337"
selfLink: /apis/extensions/v1beta1/namespaces/default/ingresses/auth-service-kubernetes-live-nginx
uid: 5068e65a-28ea-11e8-94fa-0ae1ba4584f4
spec:
rules:
- host: auth.live.gelato.tech
http:
paths:
- backend:
serviceName: auth-service-kubernetes-live-nginx
servicePort: 80
path: /
- host: auth-service.gelatogroup.com
http:
paths:
- backend:
serviceName: auth-service-kubernetes-live-nginx
servicePort: 80
path: /
tls:
- hosts:
- auth.live.gelato.tech
- auth-service.gelatogroup.com
secretName: auth-service-kubernetes-live-nginx-tls
I get next certificate like this:
Name: auth-service-kubernetes-live-nginx-tls
Namespace: default
Labels: <none>
Annotations: <none>
API Version: certmanager.k8s.io/v1alpha1
Kind: Certificate
Metadata:
Cluster Name:
Creation Timestamp: 2018-03-20T14:23:59Z
Deletion Grace Period Seconds: <nil>
Deletion Timestamp: <nil>
Owner References:
API Version: extensions/v1beta1
Block Owner Deletion: true
Controller: true
Kind: Ingress
Name: auth-service-kubernetes-live-nginx
UID: 5068e65a-28ea-11e8-94fa-0ae1ba4584f4
Resource Version: 119219236
Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/default/certificates/auth-service-kubernetes-live-nginx-tls
UID: 54dba49f-2c4a-11e8-94fa-0ae1ba4584f4
Spec:
Acme:
Config:
Dns 01:
Provider: route53
Domains:
auth.live.gelato.tech
auth-service.gelatogroup.com
Common Name:
Dns Names:
auth.live.gelato.tech
auth-service.gelatogroup.com
Issuer Ref:
Kind: ClusterIssuer
Name: letsencrypt-live
Secret Name: auth-service-kubernetes-live-nginx-tls
Status:
Conditions: <nil>
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning ErrorCheckCertificate 21m cert-manager-controller Error checking existing TLS certificate: secret "auth-service-kubernetes-live-nginx-tls" not found
Normal PrepareCertificate 21m cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 21m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech
Normal SelfCheck 20m cert-manager-controller Performing self-check for domain auth.live.gelato.tech
Warning ErrorCheckCertificate 9m cert-manager-controller Error checking existing TLS certificate: secret "auth-service-kubernetes-live-nginx-tls" not found
Normal PrepareCertificate 9m cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 9m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com
Normal PresentChallenge 9m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech
Normal SelfCheck 9m cert-manager-controller Performing self-check for domain auth.live.gelato.tech
Additionally, I found today in logs next error:
E0320 14:43:49.216953 1 prepare.go:167] Error cleaning up solver: Failed to change Route 53 record set: InvalidChangeBatch: Tried to delete resource record set [name='_acme-challenge.auth-service.gelatogroup.com.', type='TXT'] but it was not found
status code: 400, request id: 19c8e4a0-2c4d-11e8-99b7-430c9c1f3614
I've tried create this record manually, but error again apeared.
@argusua I don't think you need to create it manually as it provides a response with a long non-readable string when you query it (run test record in route53). And I am just curious - can single secret work with multiple hosts? I had an idea of one secret/certificate per host but maybe it works with several as well (I am not talking about wildcard certificates as your domain names are different.)
Get some progress on my issue. I've created new clusterIssuer (previous I get by migrating from kube-lego as described here https://github.com/jetstack/cert-manager/blob/master/docs/user-guides/migrating-from-kube-lego.md). Now I get a certificate but only for first domain:
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal PrepareCertificate 4m cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 4m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com
Normal PresentChallenge 4m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech
Normal SelfCheck 3m cert-manager-controller Performing self-check for domain auth.live.gelato.tech
Normal PrepareCertificate 2m cert-manager-controller Preparing certificate with issuer
Normal PresentChallenge 2m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com
Normal PresentChallenge 2m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech
Normal SelfCheck 1m cert-manager-controller Performing self-check for domain auth.live.gelato.tech
Normal ObtainAuthorization 49s cert-manager-controller Obtained authorization for domain auth.live.gelato.tech
Warning ErrorPrepareCertificate 3s cert-manager-controller Error preparing issuer for certificate: error presenting acme authorization for domain "auth-service.gelatogroup.com": Failed to change Route 53 record set: InvalidChangeBatch: RRSet with DNS name _acme-challenge.auth-service.gelatogroup.com. is not permitted in zone sourcelogistic.com.
status code: 400, request id: 3ae889c0-2eb0-11e8-952e-ad6827c36e4f
Normal RenewalScheduled 3s cert-manager-controller Certificate scheduled for renewal in 1438 hours
Normal PrepareCertificate 2s (x2 over 3s) cert-manager-controller Preparing certificate with issuer
Warning ErrorPrepareCertificate 2s cert-manager-controller Error preparing issuer for certificate: error presenting acme authorization for domain "auth-service.gelatogroup.com": Failed to change Route 53 record set: InvalidChangeBatch: RRSet with DNS name _acme-challenge.auth-service.gelatogroup.com. is not permitted in zone sourcelogistic.com.
status code: 400, request id: 985b9304-2eb0-11e8-8ec5-9feeb1e51a10
Normal RenewalScheduled 2s cert-manager-controller Certificate scheduled for renewal in 1438 hours
Normal PresentChallenge 1s (x2 over 3s) cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com
Looks like cert-manager uses incrorect zoneId for set dns record in route53.
Finally, I found why the wrong zoneId had used. It's because of CNAME record. My current issue resolved.
I had/have the same issue with http01 it seams, where the cert-manager cant curl the well-known url from inside the cluster, but externaly it is possible to access.
Just the edge router missing some loopback ... spend 15h for this knowledge.
The error description "error waiting for key to be available for domain" is very missleading.
Should be "error waiting for collange-key resource to be available for domain ..."
@Zetanova often referred to as 'hairpin NAT', needed if your egress and ingress have the same external IP address. It can be confusing to configure.
@whereisaaron Yes, i disovered it too. My problem was that i just missundestud the error message.
I put active/passive vm iptabales+keepalived in front of the k8s and it seams that i cant get the hairpin NAT to work.
I got a similar issue on dns01 config. Log was like follows.
(As @whereisaaron says, this should be shown only http01 config...)
$ kubectl describe certificate
...
Events:
...
Warning ErrorPrepareCertificate 2m cert-manager-controller Error preparing issuer for certificate: [error waiting for key to be available for domain "{censored}": context canceled]
...
In my case, it was fixed by just deleting (and restarting) the cert-manager pod.
I'm going to close this now as it appears to be resolved - thanks everyone for providing support and help debugging 😄
Just tried to generate letsencrypt's wildcard cert via dns01/route53 validation and getting:
E0621 14:13:18.291341 1 controller.go:186] certificates controller: Re-queuing item "namespace/xxx-net" due to error processing: Failed to change Route 53 record set: InvalidChangeBatch: Tried to delete resource record set [name='_acme-challenge.xxx.net.', type='TXT'] but it was not found
Looks like it should just silently ignore that error or don't delete entries which don't exists in route53?
Did you manage to resolve that issue?
Using the latest cert manager v0.3.0.
same problem with 0.5.0, was working fine in 0.4.1, rolling back
we had same issue, it got resolved when I created the TXT record manually, then acme cert manager changed that record because it was expired and deleted successfully.
I got the txt record key from cert manager logs and the value from aws cloudtrail logs.
Next time I will try creating the txt record with some random value to see if that helps
|
gharchive/issue
| 2018-03-16T16:43:48
|
2025-04-01T06:39:12.295897
|
{
"authors": [
"Zetanova",
"argusua",
"cryptcoin-junkey",
"grebois",
"infa-ddeore",
"mlushpenko",
"munnerz",
"simonkey007",
"urbaniak",
"whereisaaron"
],
"repo": "jetstack/cert-manager",
"url": "https://github.com/jetstack/cert-manager/issues/398",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1407905390
|
Prompt user to check supported versions if unknown
Before:
$ go run main.go operator deploy --version=foo
operator version foo does not exist
exit status 1
After:
$ jsctl operator versions
v0.0.1-alpha.17
$ go run main.go operator deploy --version=foo
operator version foo is unknown or not supported by this version of jsctl. Run 'jsctl operator versions' to see the supported operator versions.
exit status 1
Signed-off-by: Charlie Egan charlieegan3@users.noreply.github.com
Thanks Irbe 🙏
|
gharchive/pull-request
| 2022-10-13T14:27:24
|
2025-04-01T06:39:12.300245
|
{
"authors": [
"charlieegan3"
],
"repo": "jetstack/jsctl",
"url": "https://github.com/jetstack/jsctl/pull/38",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
306959545
|
Use new test fixture in Cassandra controls
I'm trying to get rid of the Cassandra specific test fixture in ./pkg/controllers/cassandra/testing
In this branch I've switched:
ServiceAccounts
Roles
RoleBindings
I'll deal with Pilot and NodePool controls in separate branches.
Release note:
NONE
/test e2e
/test e2e v1.10
/retest
/lgtm
/approve
|
gharchive/pull-request
| 2018-03-20T17:13:48
|
2025-04-01T06:39:12.306997
|
{
"authors": [
"munnerz",
"wallrj"
],
"repo": "jetstack/navigator",
"url": "https://github.com/jetstack/navigator/pull/291",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
376957886
|
Document how to get the root key after init
It would be good to have a few lines in the documentation about how to retrieve the root key after init, to make it possible to login to vault.
if you use AWS KMS you can do it with this script:
export REGION=us-west-2
export AWS_PROFILE=dev
echo "Fetching Vault unseal keys and root token from AWS..."
aws ssm get-parameters --names kubernetes-vault-root --region us-west-2 | jq -r '.Parameters[].Value' | base64 -D> /tmp/root-token
ROOT_TOKEN=$(aws kms decrypt --region $REGION --ciphertext-blob fileb:///tmp/root-token --encryption-context Tool=vault-unsealer | jq -r '.Plaintext' | base64 -D)
for i in {0..4}; do
aws ssm get-parameters --names kubernetes-vault-unseal-${i} --region us-west-2 | jq -r '.Parameters[].Value' | base64 -D > /tmp/unseal-${i}
echo -n "UNSEAL_KEY_$((i+1)): $(aws kms decrypt --region $REGION --ciphertext-blob fileb:///tmp/unseal-${i} --encryption-context Tool=vault-unsealer | jq -r '.Plaintext' | base64 -D)
"
done
echo "Initial Root token: $ROOT_TOKEN"
#31
THanks for typing that out! Yes, my question was related to AWS KMS, sorry for not clarifying
|
gharchive/issue
| 2018-11-02T20:31:29
|
2025-04-01T06:39:12.308811
|
{
"authors": [
"innovia",
"trondhindenes"
],
"repo": "jetstack/vault-unsealer",
"url": "https://github.com/jetstack/vault-unsealer/issues/29",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
67272239
|
Celery doesn't account for running and reserved tasks
This is a similar problem to #3 .
I'm using the Redis backend for Celery. I tried my hand at a solution in https://github.com/aspiredu/hirefire/commit/3ae625cc527330357a66df8d8c5e2701213d5af7, but it only works for small numbers of dynos, workers, and tasks. Even when it does work, it's painfully slow.
The issue is that active, reserved, and scheduled tasks don't show up in the queue. Missing the scheduled ones is fine, but missing active and reserved means that my queues, which are running a lot of long-running tasks, get shut down well before stuff is done processing, which causes the dynos to flap as they are put back in the queue, then taken out, then when the process repeats.
Unfortunately I can confirm that #9 has not solved this issue, even when I'm using py-ampq. I haven't tested with librabbitmq or iron-mq, but I find it likely that I'll get the same results.
An IRC conversation with @ask gave me two ideas to try. Both are somewhat involved for something as light as HireFire, but we've gotta do what we've gotta do.
Use Flower to get data on the running worker queues.
Use heartbeats from worker to get up-to-date data.
(from me) use inspect() (that's how flower gets it's info) for all queues/procs/workers at once and somehow re-use that data for the life of the request to the HireFire info URL.
2 requires some upstream code-changes to Celery (the heartbeats don't currently include scheduled and reserved counts), and also requires a persistent monitor to receive those events.
I'm honestly not sure which path to take first. I'll just have to take 1 and run with it.
@ryanhiebert did you find any solution for this trouble ?
No solution yet. It's been somewhat back-burnered in my current workload, but trying to use Flower to get a better picture of the running queues seems like the route I'm likely to try first. Any and all help would be welcome!
Any updates on this issue?
No, sorry. Haven't been able to work on it any more yet.
I wrote a basic implementation that seems to work.
import itertools
from celery import Celery
from django.conf import settings
from gevent.pool import Group
from hirefire.procs.celery import CeleryProc
def _inspect(method):
app = Celery('app', broker=settings.BROKER_URL)
inspect_result = getattr(app.control.inspect(), method)()
app.close()
return inspect_result
class OFCeleryProc(CeleryProc):
def quantity(self):
"""
Returns the aggregated number of tasks of the proc queues.
"""
if hasattr(self.channel, '_size'):
# Redis
return sum(self.channel._size(queue) for queue in self.queues)
# AMQP (Include both libs, since we have both installed)
from librabbitmq import ChannelError as LCE
from amqp.exceptions import ChannelError as ACE
count = 0
for queue in self.queues:
try:
queue = self.channel.queue_declare(queue, passive=True)
except (LCE, ACE):
# The requested queue has not been created yet
pass
else:
count += queue.message_count
active, reserved = Group().map(_inspect, ('active', 'reserved'))
num_active = sum(1 for v in itertools.chain.from_iterable(active.itervalues()) if not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues) if active else 0
num_reserved = sum(1 for v in itertools.chain.from_iterable(reserved.itervalues()) if not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues) if reserved else 0
return count + num_active + num_reserved
@joshowen : That can work OK, for a small number of queues. Unfortunately, I have around 20 queues, and the combined time of running this made for timeouts in Heroku. I've got to find something faster.
https://github.com/celery/celery/issues/2839
This issue with Celery also makes this issue worse. Tasks that are running are immediately shut down, instead of being given the time that Heroku usually gives to let the dynos shut down gracefully. Having this fixed wouldn't be sufficient by itself, but it would make the problem less of a deal for short tasks.
@ryanhiebert Even using gevent?
@joshowen: Actually, that's not what I had done before. I'm going to have to dig more into it, and see if it works for me.
I'm not currently using gevent.
With gevent I was able to run the _inspect calls in parallel and got around the heroku timeouts
Let me try it. I still do have quite a few queues (not just workers), with different procs, so even with gevent it means that I'll have to run ~20 of these to get all the stats for HireFire. Worth a try, though.
@joshowen thanks for your example code, we just ran into the same issue.
I build a similar solution, still slow, but less calls:
the app.control.inspect() calls always return the data for every queue, not only one.
We have just extended HireFireMiddleware to do these calls once (not only for active and reserved, but also for scheduled) and then the subsequent CeleryProcs will get the data from their queues out of it. So queue size does not have any effect on the calls, perhaps only on the time celery needs to call all the workers.
(cc @ryanhiebert )
will try to bake this into a PR, I think this is a common problem if you want to scale to 0 :)
@syphar: that sounds awesome. I'm looking forward to it.
@syphar: I'm working on this now for work, and if possible I'd like to see your approach. Any chance you could let me see your work (via very raw, perhaps broken PR, or perhaps a gist of the interesting code), so that I can take inspiration at least?
@ryanhiebert yes, sure. Sorry for the delay many things happening :)
middleware:
class HireFireMiddleware(OriginalHireFireMiddleware):
"""
An extended HireFireMiddleware which includes counts worker-stats.
we query for running, reserved or scheduled tasks and add them to the
queued tasks per queue.
The related procs have to have a method called ``quantity_with_running``
which will get the inspection data, so it's queried only once per
call from hirefire.
"""
def dump_procs(self, procs):
celery_inspect = {
method: getattr(app.control.inspect(), method)()
for method in ('active', 'reserved', 'scheduled')
}
data = [
{
'name': name,
'quantity': proc.quantity_with_running(celery_inspect) or 'null',
}
for name, proc in procs.items()
]
return json.dumps(data, cls=TimeAwareJSONEncoder, ensure_ascii=False)
def info(self, request):
payload = self.dump_procs(self.loaded_procs)
return HttpResponse(payload, content_type='application/json')
procs.py
class UpdatedCeleryProc(CeleryProc):
def quantity_with_running(self, celery_inspect):
count = super(UpdatedCeleryProc, self).quantity()
for func, tasks in celery_inspect.items():
if not tasks:
continue
queue_tasks = [
v
for v in itertools.chain.from_iterable(tasks.values())
if (
not v.get('acknowledged', False) and
v.get('delivery_info', {}).get('exchange') in self.queues
)
]
count += len(queue_tasks)
return count
class LowPrioWorkerProc(UpdatedCeleryProc):
name = 'celeryworkerlowprio'
queues = ['lowprio']
this code
as expected: only works with celery
not perfect: increases response time (it's around ~3s on our production system)
good: the amount of procs/queues doesn't matter, since the inspect result contains everything.
runs in production for 4 weeks here
only tested on python 3.5
What was blocking me (when I had some time) was the question how to design this the best way to
make it optional
make it backend-agnostic
Awesome. I'll work on figuring out those when I'm implementing it for myself. This is very, very, helpful. Thank you.
@ryanhiebert we can get this into a working PR together, if you have some nice ideas on the design :)
Would you be up for a pairing session, perhaps? You can hit me up on irc.freenode.net, I'm ryanhiebert, or you can email me ryan@ryanhiebert.com
If you're interested in pairing, I've set up a floobits workspace at https://floobits.com/ryanhiebert/hirefire.
I do have a good idea, and that's to send a cache dictionary into the procs, so that they can share state between themselves. I'm working up an implementation.
Here's my work-in-progress branch on the approach I came up with. Feel free to give me feedback. I'll open up a PR after I've done some more testing in my environment.
https://github.com/aspiredu/hirefire/tree/global-cache
|
gharchive/issue
| 2015-04-09T03:28:10
|
2025-04-01T06:39:12.347868
|
{
"authors": [
"joshowen",
"ryanhiebert",
"syphar",
"touilleMan"
],
"repo": "jezdez/hirefire",
"url": "https://github.com/jezdez/hirefire/issues/8",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1251746594
|
Support gstreamer
Need to write an example using the macro with an gstreamer element, and possibly create another gst_element macro to make it easier.
I implemented a webrtcsrc element based on this crate, I think we have mostly everything we need tbh.
I also started adding macros to ease GstElement implementation here, at this point it implements element metadata and pad templates declaration in the macros like:
#[gobject::gst_element(
class(final),
long_name = "TheTestElement",
classification = "Test/Filter",
description = "Just a test",
author = "Thibault Saunier <tsaunier@igalia.com>",
pad_templates(
src(presence="always"),
// `__` is transformed to `_%` as "%" is not a valid character
sink__u(direction="sink", presence="sometimes", caps="video/x-raw"),
)
)]
My next step will be to improve the way we interact with Pads creation and integration into the Elements, maybe adding a macros like:
#[gobject::gst_pad(sink__u)]
impl SrcPad {
fn chain() -> Result<gst::FlowSuccess, gst::FlowError> {
...
}
}
but I am still thinking.
|
gharchive/issue
| 2022-05-28T23:44:27
|
2025-04-01T06:39:12.351914
|
{
"authors": [
"jf2048",
"thiblahute"
],
"repo": "jf2048/gobject",
"url": "https://github.com/jf2048/gobject/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2301986153
|
Creation of a roadmap section
Idea Contribution
[X] I have read all the feature request issues.
[X] I'm interested in working on this issue
[X] I'm part of GSSOC organization
Explain feature request
Creation of a roadmap section,which will contact roadmap and their respected study material and links for various fields, Android development,web development ,ml , DSA etc
Explain your solution
No response
Any alternative approaches/features
No response
Additional Context
No response
@jfmartinz kindly assign me this issue
@jfmartinz Maybe this issue is not solved yet!
I would like to resolve this issue ASAP, creating a roadmap section for almost every tech field- Data Science, Web Development, App Development, UI/UX Design, and Machine Learning...
Kindly assign me this issue, and I'll resolve it soon.
This is actually completed but the author forgot to reference the PR. but you can still add some resources that youd like and create a PR for that, just make sure that those resources are not duplicate, thanks
@jfmartinz ok then I will do so as a separate pr
|
gharchive/issue
| 2024-05-17T07:08:50
|
2025-04-01T06:39:12.374180
|
{
"authors": [
"Soumya6Tiwari",
"jfmartinz",
"renunegii"
],
"repo": "jfmartinz/ResourceHub",
"url": "https://github.com/jfmartinz/ResourceHub/issues/276",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
903925621
|
make sure statuses are included with the buildinfo.
[ ] All tests passed. If this feature is not already covered by the tests, I added new tests.
[x] This pull request is on the dev branch.
[x] I used gofmt for formatting the code before submitting the pull request.
fixes #356
@AnalogJ,
Since we're currently considering to make major changes and improvements to the entire build promotion process, I suggest that for now, we wait with changes to the build-info schema to include promotion statuses.
Hi @eyalbe4
I'd really like to get this merged if possible.
Any chance you could provide an update about the future of Promotion status's in this library?
I've also rebased my branch on the latest from master.
Hey @eyalbe4 @yahavi
Any additional feedback for this PR?
@eyalbe4 @yahavi any updates here? Sorry to keep bugging you
@AnalogJ,
Since this change may conflict with other planned changes to the build promotion process, will you be able to share more details about the need? How are you planning to use and consume this data?
With this information, maybe we'll be able to come up with an alternative solution.
Hey,
This is a critical feature for us. Basically we use the Jfrog Go SDK to create "Builds".
The builds contain a list of versioned artifacts that will be used in a deployment. These artifacts are then scanned using XRAY, and deployed to various non-production environments automatically. The "PromotionStatus" field is used as a manual flag to ensure that Production Deployments can only use vetted/production-ready artifacts.
I see @AnalogJ.
Would you be willing perhaps to tag the build-info artifact in Artifactory, with a property that will indicate the status of the build? Every time you publish a build-info, it is also saved as a file under the artifactory-build-info repository. You can set and get the value of this custom property on/from the build-info artifact, to support your logic.
This functionality is supported by this library as well. See the Setting Properties on Files in Artifactory and Deleting Properties from Files in Artifactory sections in the README.
This PR doesnt seem relevant since https://github.com/jfrog/jfrog-client-go/blob/master/artifactory/services/promote.go is merged. Closing.
|
gharchive/pull-request
| 2021-05-27T15:42:43
|
2025-04-01T06:39:12.384337
|
{
"authors": [
"AnalogJ",
"eyalbe4"
],
"repo": "jfrog/jfrog-client-go",
"url": "https://github.com/jfrog/jfrog-client-go/pull/357",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1320061483
|
Some defaults in the user HCL dont match the description
Describe the bug
profile_updatable and disable_ui_access default to True when the docs say they should be False
Requirements for and issue
[x] A description of the bug
[ ] A fully functioning terraform snippet that can be copy&pasted (no outside files or ENV vars unless that's part of the issue). If this is not supplied, this issue will likely be closed without any effort expended.
[ ] Your version of artifactory (you can curl it at $host/artifactory/api/system/version
[ ] Your version of terraform
[ ] Your version of terraform provider
Expected behavior
Docs match actuals
Additional context
Add any other context about the problem here.
https://github.com/jfrog/terraform-provider-artifactory/pull/517
Actually docs here say it should be true https://github.com/jfrog/terraform-provider-artifactory/blob/master/docs/resources/user.md not sure what is the correct default value and which bit is wrong :)
@apr-1985 Per my comment, it's the Description that is incorrect.
|
gharchive/issue
| 2022-07-27T19:57:56
|
2025-04-01T06:39:12.389830
|
{
"authors": [
"alexhung",
"apr-1985"
],
"repo": "jfrog/terraform-provider-artifactory",
"url": "https://github.com/jfrog/terraform-provider-artifactory/issues/518",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
45118362
|
issue 470 allow creation of location with only try_files
This commit will fix #470.
@buckstr the proposed PR won't allow you to create a location with only try_files.
As is, if you only set try_files and don't set proxy, location_alias, stub_status, fastcgi or www_root then the template vhost/locations/empty.erb will be rendered. This doesn't include try_files.
Hi --
There's been no further activity on this since the 'needs Rebase' and 'needs tests' labels were set. If we don't have any further comments in the next month, this PR will be closed.
Addressed in #834 already.
|
gharchive/pull-request
| 2014-10-07T15:05:19
|
2025-04-01T06:39:12.392383
|
{
"authors": [
"3flex",
"buckstr",
"wyardley"
],
"repo": "jfryman/puppet-nginx",
"url": "https://github.com/jfryman/puppet-nginx/pull/471",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
476590766
|
Conversion from FromSqlError to Error
Error needs column index, but FromSqlError doesn't have it. I've fudged it by putting obviously-wrong index value, and then hiding that fact in Display.
Ideally the Error should use Option<usize> for index (and type), but I assume it's better to avoid unnecessary backwards-compat break.
Fixes #555
The test failure looks like a vcpkg installation issue, unlikely to be related to this change.
Thank you.
|
gharchive/pull-request
| 2019-08-04T20:27:50
|
2025-04-01T06:39:12.406462
|
{
"authors": [
"gwenn",
"kornelski"
],
"repo": "jgallagher/rusqlite",
"url": "https://github.com/jgallagher/rusqlite/pull/556",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
52574994
|
feature request: routie.back(), routie.next()
it would be great to have back and next method for navigating through history
What's wrong with using window.history.back() and window.history.forward()? These should work out of the box.
I haven't dug into why, but in cordova window.history.back() and window.history.forward() don't seem to work -- at least not on the emulated Nexus 4.
|
gharchive/issue
| 2014-12-20T20:00:55
|
2025-04-01T06:39:12.407982
|
{
"authors": [
"GinkgoFJG",
"akaRem",
"benplum"
],
"repo": "jgallen23/routie",
"url": "https://github.com/jgallen23/routie/issues/35",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
130815186
|
Display an error message when file upload fails
A common error that has been occurring is the failure of the /api/upload endpoint. While I'd like to get around to finding out why the error is happening so frequently, the least we can do is show an error message.
Here is the console output of when it happened to me twice today:
Looks good to me. See my one comment.
It could be that disk space on the server is getting low. It might be worth looking into.
I think that might be it. Disk space, I mean.
Hiding the alert on upload is a good idea. I'll update the PR.
On Feb 2, 2016 7:40 PM, "James Lamine" notifications@github.com wrote:
Looks good to me. See my one comment.
It could be that disk space on the server is getting low. It might be
worth looking into.
—
Reply to this email directly or view it on GitHub
https://github.com/jglamine/calvinwebprint/pull/49#issuecomment-178917875
.
|
gharchive/pull-request
| 2016-02-02T21:02:15
|
2025-04-01T06:39:12.457891
|
{
"authors": [
"jglamine",
"tylervz"
],
"repo": "jglamine/calvinwebprint",
"url": "https://github.com/jglamine/calvinwebprint/pull/49",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1560036992
|
Bring "Convert labels to SVG" export option to Desktop version + conversion issue
[ ] I agree to follow the Code of Conduct that this project adheres to.
[ ] I have searched the issue tracker for a feature request that matches the one I want to file, without success.
So, I've this big file I'm working on that has a lot (really a lot) of text labels and I'll have it to upload it to Wikimedia Commons. Since Commons doesn't like certain tags about text labels I have to convert that into vectorized text, feature that is only available on the web version. I'd really like to have it on Deskto
p so that I wouldn't have to switch working environment every time I need to export my drawings.
Second, since this file I'm working on is as I said pretty text intensive, the converter struggles to convert all the text and ends with an error. I'll upload the file for context so that you can replicate the issue.
Carrara-Avenza rev. 3 (2023-01-23).zip
Cheers!
convert to SVG requires a back-end, it can't be ported to desktop.
|
gharchive/issue
| 2023-01-27T16:20:50
|
2025-04-01T06:39:12.496373
|
{
"authors": [
"OverFloyd",
"davidjgraph"
],
"repo": "jgraph/drawio",
"url": "https://github.com/jgraph/drawio/issues/3319",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
305977313
|
The future is ember-animated
I propose to either sunset this addon and give people a transition path to ember-animated or perhaps reimagine it as a thin wrapper atop ember-animated.
Thoughts welcome!
ember-sortable has accomplished exactly what I attempted to create in an app I'm working on in a reliable way. I'd hate to see it go, but if the plan is to sunset the addon then I'd be happy to replace what we have with a thin wrapper atop of ember-animated should that be created.
I still think this is worth exploring, but it’s a long way off so closing for now.
|
gharchive/issue
| 2018-03-16T15:42:24
|
2025-04-01T06:39:12.504075
|
{
"authors": [
"jgwhite",
"joe-wroten-q2"
],
"repo": "jgwhite/ember-sortable",
"url": "https://github.com/jgwhite/ember-sortable/issues/178",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
835750943
|
Mongock is not compatible with the native build
Describe the bug
Mongock is a library to help us to handle the data migration with a MongoDb database.
JHipster already uses it with Mongo and we reused the same with Quarkus.
Mongock uses org.reflections that cause an error when we build it with GraalVM to generate a native executable.
To Reproduce
Steps to reproduce the behavior:
Generate a project with MongoDb as a database.
Install GraalVM
openjdk 11.0.10 2021-01-19
OpenJDK Runtime Environment GraalVM CE 21.0.0.2 (build 11.0.10+8-jvmci-21.0-b06)
OpenJDK 64-Bit Server VM GraalVM CE 21.0.0.2 (build 11.0.10+8-jvmci-21.0-b06, mixed mode, sharing)
Run the following command: ./mvnw clean verify -Pnative -DskipTests
Caused by: com.oracle.graal.pointsto.constraints.UnsupportedFeatureException: Unsupported features in 2 methods
Detailed message:
Error: com.oracle.graal.pointsto.constraints.UnresolvedElementException: Discovered unresolved type during parsing: org.apache.commons.vfs2.VFS. To diagnose the issue you can use the --allow-incomplete-classpath option. The missing type is then reported at run time when it is accessed the first time.
Trace:
at parsing org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(Vfs.java:287)
Call path from entry point to org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(URL):
at org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(Vfs.java:287)
at org.reflections.vfs.Vfs.fromURL(Vfs.java:98)
at org.reflections.vfs.Vfs.fromURL(Vfs.java:90)
at org.reflections.Reflections.scan(Reflections.java:236)
at org.reflections.Reflections$1.run(Reflections.java:199)
at java.lang.Thread.run(Thread.java:834)
at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:519)
at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192)
at com.oracle.svm.core.code.IsolateEnterStub.PosixJavaThreads_pthreadStartRoutine_e1f4a8c0039f8337338252cd8734f63a79b5e3df(generated:0)
Error: com.oracle.graal.pointsto.constraints.UnresolvedElementException: Discovered unresolved type during parsing: org.apache.commons.vfs2.VFS. To diagnose the issue you can use the --allow-incomplete-classpath option. The missing type is then reported at run time when it is accessed the first time.
Trace:
at parsing org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(Vfs.java:281)
Call path from entry point to org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(URL):
at org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(Vfs.java:281)
at org.reflections.vfs.Vfs.fromURL(Vfs.java:97)
at org.reflections.vfs.Vfs.fromURL(Vfs.java:90)
at org.reflections.Reflections.scan(Reflections.java:236)
at org.reflections.Reflections$1.run(Reflections.java:199)
at java.lang.Thread.run(Thread.java:834)
at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:519)
at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192)
at com.oracle.svm.core.code.IsolateEnterStub.PosixJavaThreads_pthreadStartRoutine_e1f4a8c0039f8337338252cd8734f63a79b5e3df(generated:0)
at com.oracle.graal.pointsto.constraints.UnsupportedFeatures.report(UnsupportedFeatures.java:129)
at com.oracle.svm.hosted.NativeImageGenerator.runPointsToAnalysis(NativeImageGenerator.java:770)
... 8 more
Expected behavior
The native build should succeed without errors.
Suggestions
Move from Mongock to Liquibase with Mongo.
Find a way to make Mongock compatible with the native build.
Additional context
Use ./mvnw dependency:tree -Dverbose to check which library is using VFS.
Check this issue: https://github.com/quarkusio/quarkus/issues/9801
cc @danielpetisme
closed by #180
I have just realised about this ticket and it seems that has been somehow sorted out.
However is there anything that we can provide natively from Mongock to help on this or improve.
Just let me know.
|
gharchive/issue
| 2021-03-19T09:28:00
|
2025-04-01T06:39:12.520004
|
{
"authors": [
"avdev4j",
"dieppa"
],
"repo": "jhipster/generator-jhipster-quarkus",
"url": "https://github.com/jhipster/generator-jhipster-quarkus/issues/179",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
494845895
|
Add caffeine option for cacheProvider
Overview of the feature request
Caffeine is a new option available for cache since JHipster 6.3.0
Motivation for or Use Case
It should be possible to select this option in JDL
Related issues or PR
https://github.com/jhipster/generator-jhipster/pull/10303
[X] Checking this box is mandatory (this is just to show you read everything)
@murdos : Thanks for adding this; seems I've missed this. 😄
|
gharchive/issue
| 2019-09-17T20:51:33
|
2025-04-01T06:39:12.577523
|
{
"authors": [
"SudharakaP",
"murdos"
],
"repo": "jhipster/jhipster-core",
"url": "https://github.com/jhipster/jhipster-core/issues/364",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2720110144
|
Gradle CI doesn't launch unit tests
It seems that the gradle ci don't launch unit test only IT.
I have tried to put a failing unit test and the ci was green
Is this the behaviour or a misconfiguration on my side ?
A possible working solution
github-actions.yml
jobs:
tests:
name: tests
runs-on: ubuntu-latest
timeout-minutes: 20
steps:
- name: 'Setup: checkout project'
uses: actions/checkout@v4
- name: 'Setup: environment'
id: setup
uses: ./.github/actions/setup
- name: Setup Gradle
uses: gradle/gradle-build-action@v2
- name: 'Test: run unit tests'
run: ./gradlew clean test --no-daemon
- name: 'Test: run integration tests'
run: ./gradlew clean integrationTest --no-daemon
@fabienpuissant : I looked at a random build action: feat(liquibase): async liquibase configuration should not be activated if liquibase is disabled. At least for this example, the tests were executed:
> Task :test
JwtReaderTest > should not authenticate user from empty token STANDARD_OUT
2024-12-05T23:23:09.354Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid empty JWT token
JwtReaderTest > should not authenticate user from blank token STANDARD_OUT
2024-12-05T23:23:09.359Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid empty JWT token
JwtReaderTest > should not authenticate user from invalid token STANDARD_OUT
2024-12-05T23:23:09.481Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid JWT token
2024-12-05T23:23:09.482Z TRACE --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid JWT token: Invalid compact JWT string: Compact JWSs must contain exactly 2 period characters, and compact JWEs must contain exactly 4. Found: 0
Here is the part of the github-actions.yml responsible for that:
- name: 'Test: verify ${{ matrix.app }}'
if: steps.tests-requirement-check.outputs.execute_tests == 'true'
working-directory: /tmp/jhlite/${{ matrix.app }}/
run: |
if [ -f 'mvnw' ]; then
./mvnw clean verify -Dsonar.qualitygate.wait=true sonar:sonar
elif [ -f 'gradlew' ]; then
./gradlew clean build sonar --no-daemon --info
else
npm install
npm test
fi
Please, could you provide an example that reproduces the not desired behavior?
@renanfranca It's on a generated app
See https://github.com/fabienpuissant/test-jh-lite-gradle-ci/actions/runs/12202622941
should fail because of : https://github.com/fabienpuissant/test-jh-lite-gradle-ci/blob/main/src/test/java/com/mycompany/myapp/FailingTest.java
In the test-ci for jhlite it seems to be ok 😄
Not 100% sure just on my mobile, but with the soft ordering should run after for the integration test phase this does not define a dependency. So when just executing integrationtest this does not imply running unit tests.
To work out of the box without Jacoco, it is going need to change the Gradle module to directly link unit and integration tests. Then, when adding Jacoco, remove that link to make Jacoco the glue.
@fabienpuissant : I think your solution is better and more flexible 😉. Even if we add the Jacoco module, the CI configuration will need to be updated to execute Jacoco instead of the IntegrationTest task. IMHO, If you want, feel free to proceed with the implementation—if not, I can take care of it on Monday 👍.
Thanks for reporting that improvement/bug!
cc: @pascalgrimaud , @murdos and @atomfrede
|
gharchive/issue
| 2024-12-05T11:11:31
|
2025-04-01T06:39:12.587627
|
{
"authors": [
"atomfrede",
"fabienpuissant",
"renanfranca"
],
"repo": "jhipster/jhipster-lite",
"url": "https://github.com/jhipster/jhipster-lite/issues/11513",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1054632631
|
Typo in Equation 5 in paper?
Equation 5 in the paper makes use of f_S, however, the corresponding portion in Figure 2 (orange) only makes use of f_T. Additionally, f_T makes more sense since the input is in the target domain. Should the f_S in equation 5 be f_T?
”Note that we can also consider a feature-level method Which distinguishes the features or semantics from two image sets as viewed under a task network. “
I think both f _ s and f _ t can do experiments.
|
gharchive/issue
| 2021-11-16T09:14:09
|
2025-04-01T06:39:12.602707
|
{
"authors": [
"arjung128",
"zhe-juanz"
],
"repo": "jhoffman/cycada_release",
"url": "https://github.com/jhoffman/cycada_release/issues/41",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2447264123
|
"This is the solution of the issue"
> "This is the solution of the issue"
this is a quote
Originally posted by @jhon-elen in https://github.com/jhon-elen/Prueba3/issues/7#issuecomment-2267641691
Anothe comment
@jhonelentest
Header
Header
Cell
Cell
Cell
Cell
import os
os.chdir()
Details
This is a detail
|
gharchive/issue
| 2024-08-04T19:18:41
|
2025-04-01T06:39:12.608616
|
{
"authors": [
"jhon-elen"
],
"repo": "jhon-elen/Prueba3",
"url": "https://github.com/jhon-elen/Prueba3/issues/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1503778256
|
Abrocoma bennetti is mentioned in MSW1/2 but Abrocoma bennettii was mentioned in original publication
via @myrmoteras -
[...] For example, Abrocoma bennetti does not match COL because it is spelled with out a ..ii in MSW2 (and also in MSW1, and 3). The original in has a ...ii though. https://doi.org/10.5281/zenodo.7457193 or https://github.com/plazi/treatments-xml/blob/main/data/03/FA/64/03FA64167865FF8AFF34F96C4ECCA3D4.xml and with access to the original publication. now also in BLR. [...]
also via @myrmoteras
@myrmoteras in running
preston ls | preston grep -o --no-line "(Abrocoma bennetti)([^i])"
after cloning
Plazi Community. (2022). Plazi Treatments XML Archive hash://sha256/3cfd60b8b19e76d208377537835de92efdb5b945a6a71765b74ed2fe22298b42 hash://md/594923284e3eb9965b8cbad149c76cd0f (hash://sha256/3cfd60b8b19e76d208377537835de92efdb5b945a6a71765b74ed2fe22298b42) [Data set]. Zenodo. https://doi.org/10.5281/zenodo.7443343
I was able to cite all locations of Abrocoma bennetti without the ...ii suffix in a recent plazi/treatments-xml corpus and appended them to annotations.csv via https://github.com/jhpoelen/msw-plazi/commit/ea24ff964ea1373302ac01d406a0ac9226f65c68 .
The associated treatment metadata headers in which the name (without ii suffix) occurs are:
<document ID-DOI="http://doi.org/10.5281/zenodo.6228755" ID-GBIF-Dataset="aa7a2f2b-ea55-4ae2-863a-93bf3bdde535" ID-GBIF-Taxon="119379023" ID-ISSN="1175-5326" ID-Zenodo-Dep="6228755" checkinTime="1459984091607" checkinUser="plazi" docAuthor="Valim, Michel P. & Linardi, Pedro Marcos" docDate="2008" docId="03948456FFB1FFED26E95090DEA41A59" docLanguage="en" docName="zt01899p024.pdf" docOrigin="Zootaxa 1899" docStyle="DocumentStylede.uka.ipd.idaho.easyIO.settings.Settings@1effd53c" docStyleName="zootaxa.2007.monograph" docTitle="Gyropus scalaris Werneck 1942" docType="treatment" docVersion="6" lastPageId="16" lastPageNumber="17" masterDocId="FFADFC2EFFBFFFFD267E5649DD2D1865" masterDocTitle="A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae)" masterLastPageNumber="24" masterPageNumber="1" pageId="14" pageNumber="15" updateTime="1645556127074" updateUser="ExternalLinkService">
<document ID-DOI="http://doi.org/10.5281/zenodo.4592306" ID-GBIF-Dataset="e61aa52f-2e17-4afc-b30a-6cd468eea36b" ID-GBIF-Taxon="180302034" ID-ISSN="1545-1410" ID-Zenodo-Dep="4592306" ID-ZooBank="B8876E05-78A8-45FA-A34A-9FFB56741A20" checkinTime="1614695199489" checkinUser="felipe" docAuthor="Pérez, M. Julieta & Díaz, M. Mónica" docDate="2018" docId="03970124E858FFE6FF5BB08A2EBC056C" docLanguage="en" docName="MammalSpecies.50.963.74-83.pdf" docOrigin="Mammalian Species 50 (963)" docSource="http://dx.doi.org/10.1093/mspecies/sey010" docStyle="DocumentStyle:E260F2FEBB038053A7447521347BED36.2:MammalSpecies.2016-.journal_article" docStyleId="E260F2FEBB038053A7447521347BED36" docStyleName="MammalSpecies.2016-.journal_article" docStyleVersion="2" docTitle="Octodontomys gliroides" docType="treatment" docVersion="4" lastPageId="6" lastPageNumber="80" masterDocId="FFAE795CE859FFE0FF8AB0412A6B0312" masterDocTitle="Octodontomys gliroides (Rodentia: Octodontidae)" masterLastPageNumber="83" masterPageNumber="74" pageId="1" pageNumber="75" updateTime="1643466530690" updateUser="ExternalLinkService" zenodo-license-document="CLOSED">
<document ID-DOI="http://doi.org/10.5281/zenodo.6228735" ID-GBIF-Dataset="aa7a2f2b-ea55-4ae2-863a-93bf3bdde535" ID-GBIF-Taxon="119379012" ID-ISSN="1175-5326" ID-Zenodo-Dep="6228735" checkinTime="1459984091607" checkinUser="plazi" docAuthor="Valim, Michel P. & Linardi, Pedro Marcos" docDate="2008" docId="03948456FFB8FFFA26E952E1DE221F7B" docLanguage="en" docName="zt01899p024.pdf" docOrigin="Zootaxa 1899" docStyle="DocumentStylede.uka.ipd.idaho.easyIO.settings.Settings@1effd53c" docStyleName="zootaxa.2007.monograph" docTitle="Gyropus longus Neumann 1912" docType="treatment" docVersion="6" lastPageNumber="8" masterDocId="FFADFC2EFFBFFFFD267E5649DD2D1865" masterDocTitle="A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae)" masterLastPageNumber="24" masterPageNumber="1" pageId="7" pageNumber="8" updateTime="1645556127074" updateUser="ExternalLinkService">
<document ID-DOI="http://doi.org/10.5281/zenodo.7353039" ID-ISBN="0-89327-235-3" ID-Zenodo-Dep="7353039" approvalRequired="41" approvalRequired_for_document="2" approvalRequired_for_matCits="35" approvalRequired_for_originalDoi="1" approvalRequired_for_treatments="3" checkinTime="1665792792562" checkinUser="GgServerImporter" docAuthor="James H. Honacki, Kenneth E. Kinman & James W. Koeppl" docDate="1982" docId="03BC87A94624FFB823F2F6CB530FF624" docLanguage="en" docName="MammalSpeciesofTheWorld.1982.Rodentia.560-594.pdf.imd" docOrigin="Mammal Species of the World (1 st Edition), Lawrence, Kansas, USA: Alien Press, Inc. & The Association of Systematics Collections" docTitle="Abrocoma bennetti Waterhouse 1837" docType="treatment" docVersion="3" lastPageNumber="585" masterDocId="FF85FFD1463DFFA12306FFF55074FFFA" masterDocTitle="Order Rodentia (Part 6)" masterLastPageNumber="594" masterPageNumber="560" pageNumber="585" updateTime="1669255589438" updateUser="ExternalLinkService">
<document ID-DOI="http://doi.org/10.5281/zenodo.7353083" ID-ISBN="1-56098-217-9" ID-Zenodo-Dep="7353083" approvalRequired="3" approvalRequired_for_document="2" approvalRequired_for_originalDoi="1" checkinTime="1667534037376" checkinUser="GgServerImporter" docAuthor="Charles A. Woods" docDate="1993" docId="E11F878EFFCAFFC5FF3DF9DBFC54D984" docLanguage="en" docName="MammalSpeciesofTheWorld.1993.Hystricognathi.771-806.pdf.imd" docOrigin="Mammal Species of the World (2 nd Edition), Washington and London: Smithsonian Institution Press" docTitle="Abrocoma bennetti Waterhouse 1837" docType="treatment" docVersion="3" lastPageNumber="789" masterDocId="1D26FFF6FFD8FFD7FFE1FC69FFB0DF05" masterDocTitle="Order Rodentia - Suborder Hystricognathi" masterLastPageNumber="806" masterPageNumber="771" pageNumber="789" updateTime="1669256602727" updateUser="ExternalLinkService">
<document ID-DOI="http://doi.org/10.5281/zenodo.7353083" ID-ISBN="1-56098-217-9" ID-Zenodo-Dep="7353083" approvalRequired="3" approvalRequired_for_document="2" approvalRequired_for_originalDoi="1" checkinTime="1667534037376" checkinUser="GgServerImporter" docAuthor="Charles A. Woods" docDate="1993" docId="E11F878EFFCAFFC5FF50F949FDC9DA9C" docLanguage="en" docName="MammalSpeciesofTheWorld.1993.Hystricognathi.771-806.pdf.imd" docOrigin="Mammal Species of the World (2 nd Edition), Washington and London: Smithsonian Institution Press" docTitle="Abrocoma Waterhouse 1837" docType="treatment" docVersion="3" lastPageNumber="789" masterDocId="1D26FFF6FFD8FFD7FFE1FC69FFB0DF05" masterDocTitle="Order Rodentia - Suborder Hystricognathi" masterLastPageNumber="806" masterPageNumber="771" pageNumber="789" updateTime="1669256602727" updateUser="ExternalLinkService">
generated via
cat ../msw-plazi2/annotations.csv\
| grep Donat\
| cut -d ',' -f6\
| sed 's/cut://g'\
| sed 's+!/b.*$++g'\
| sort\
| uniq\
| grep hash\
| parallel "preston cat {1} | head -n1"
So,according to Plazi's recent treatments-xml corpus, the assumed invalid name (i.e., Abrocoma bennetti without ii prefix) appears in:
Valim, Michel P., & Linardi, Pedro Marcos. (2008). Gyropus scalaris Werneck 1942. In A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae), pp. 1-24 in Zootaxa (Vol. 1899, pp. 15–17). Zenodo. https://doi.org/10.5281/zenodo.6228755
Pérez, M. Julieta, & Díaz, M. Mónica. (2018). Octodontomys gliroides. In Octodontomys gliroides (Rodentia: Octodontidae), pp. 74-83 in Mammalian Species (Vol. 50, Number 963, pp. 75–80). Zenodo. https://doi.org/10.5281/zenodo.4592306
Valim, Michel P., & Linardi, Pedro Marcos. (2008). Gyropus longus Neumann 1912. In A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae), pp. 1-24 in Zootaxa (Vol. 1899, p. 8). Zenodo. https://doi.org/10.5281/zenodo.6228735
James H. Honacki, Kenneth E. Kinman, & James W. Koeppl. (1982). Order Rodentia (Part 6). In Mammal Species of the World (1st Edition) (pp. 560–594). Alien Press, Inc. & The Association of Systematics Collections. https://doi.org/10.5281/zenodo.7353039
Charles A. Woods. (1993). Order Rodentia - Suborder Hystricognathi. In Mammal Species of the World (2nd Edition) (pp. 771–806). Smithsonian Institution Press. https://doi.org/10.5281/zenodo.7353083
and . . . annotations now exist to point to exactly which position of the texts they occur. You should actually be able to generate a little web widget that places the annotation exact on top of the location the suspected text occurs. . .
|
gharchive/issue
| 2022-12-19T23:39:55
|
2025-04-01T06:39:12.621723
|
{
"authors": [
"jhpoelen"
],
"repo": "jhpoelen/msw-plazi",
"url": "https://github.com/jhpoelen/msw-plazi/issues/2",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1820088029
|
The document object obtained from parsing with jsoup.parse is missing one layer of elements compared to the original HTML.
https://test.em.bonahl.com/#/user/login
original :
jsoup:
https://github.com/jhy/jsoup/blob/1762412a28fa7b08ccf71d93fc4c98dc73086e03/src/main/java/org/jsoup/parser/HtmlTreeBuilderState.java#L417
It seems that nesting forms is not allowed here. However, modern frontend frameworks have implemented nested form nesting.
Input:
<div>
<form id=1>
<form id=2>
</form>
</form>
jsoup parse:
<div>
<form id="1">
</form>
</div>
Chrome parse (Inspect Document)
Spec:
https://html.spec.whatwg.org/multipage/parsing.html#parsing-main-inbody
A start tag whose tag name is "form"
If the form element pointer is not null, and there is no template element on the stack of open elements, then this is a parse error; ignore the token.
...
I believe jsoup is working per spec and to the same output as current browsers by only allowing one form on the stack at a time.
Shadow DOM enables you to attach a DOM tree to an element, and have the internals of this tree hidden from JavaScript and CSS running in the page. Modern frontend frameworks have implemented neste HTML forms through Shadow DOM, and it have been widely used. jsoup should adapt to this.
|
gharchive/issue
| 2023-07-25T11:05:08
|
2025-04-01T06:39:12.661119
|
{
"authors": [
"666asd",
"jhy",
"skylakeliu"
],
"repo": "jhy/jsoup",
"url": "https://github.com/jhy/jsoup/issues/1983",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
341462100
|
no tabs on linux
Description
I assume that tabs are supported, since an issue for tabs feature-request has already been closed long time ago #116 and from some issues it seems that tabs are supported, ex in OSX : #647.
However on linux (ubuntu 18.04) this is not the case, tabs open in new window.
Steps to reproduce issue
middle-click or ctrl-click a internal link opens a new window instead of a new tab.
links are not either considered as external URLs because they do not open in web browser but in nativefier.
Details
Are you nativefying a public website ? => tried few websites, some hosted locally and some public websites.
Feature request? Have you looked at nativefier --help to see if an existing option could fit your needs? => yes and I trired to specify --internal-urls but it did not change the behaviour.
Full nativefier command used to build your app: nativefier --name "Gitlab" --icon "/path/to/gitlab/icon.png" --internal-urls "https://mydomain.comgitlab/*" "https://mydomain.com/gitlab"
Version of Nativefier (run nativefier --version): v7.6.4
Version of node.js (run node --version): v9.2.0
OS: Ubuntu 18.04
The nativefier support for tabs comes from electron; support for native tabs in electron for macOS was added here: https://github.com/electron/electron/pull/9052
If electron gets updated to support tabs on more platforms then it should be relatively simple to add support for them to nativefier.
It would be possible to change how middle-click and ctrl-click behave in nativefier, to open external links in an external browser when native tabs are unavailable, but I'm not convinced such a change would provide value, since external links already open in an external browser when you left-click on them. Being able to use middle-click to override the default "open externally" behavior seems useful to me.
@monkeydri 👍 to what @bacongravy said. See Release Notes for Nativefier v7.6.3:
macOS: Add tabs, used automatically instead of windows (PR #579).
Provided by Electron and (so far) no available for Windows/Linux, contributions welcome.
alright, thanks fro your answer. It's a pity, as I don't think native tabs will be available on linux/windows anytime soon on Electron as there is no equivalent feature on them as macOS native tabs (ie. an option to display multiple windows from an app as tabs).
@bacongravy Can the module electron-tabs help to support tabbing in Linux?
|
gharchive/issue
| 2018-07-16T10:19:36
|
2025-04-01T06:39:12.670798
|
{
"authors": [
"bacongravy",
"mhgharieb",
"monkeydri",
"ronjouch"
],
"repo": "jiahaog/nativefier",
"url": "https://github.com/jiahaog/nativefier/issues/655",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1394210871
|
Guidance running script
Hello, very interested in the project, I would like to run the script. Supposedly there are some parameters, but I don't know how to add them.
Running the script without parameters gives blank output.
I also tried running the blocks of code as files, the first downloads model files and then errors at line 7: list indices must be integers or slices, not str
How should one go about running this? I'm using windows
Make sure you can run Whisper without any problems first.
Next:
import whisper
from stable_whisper import modify_model
model = whisper.load_model('base', 'cuda')
modify_model(model)
Then just use the model how you normally how you would use it. It should behave just like how whisper normally behaves, but with some additional values in the results.
I'm having the same problem. I can use Whisper with no issues, but I always get this error when executing the example script.
word_timestamps = results['segments']['word_timestamps']
TypeError: list indices must be integers or slices, not str
I'm also using Windows 10 and I'm a complete novice, so apologies if I'm making a silly mistake.
you can run like this but all the data is visible, so not very clean:
make "run.py" file and put
import whisper
from stable_whisper import modify_model
model = whisper/load_model('base', 'cuda')
modify_model(model)
results = model.transcribe('out.wav')
print(results)
type python run.py in the terminal.
Maybe @jianfch can write a packaged demo script later. This is not for command line tools
Yes, but that does not have single word timings, thats just generating a normal (better than youtube though w/punctuation) .srt
I'd like to make .ass timed text subtitles for all videos. so I need the timestamps and word data
Not sure about .ass, but you can change results_to_sentence_srt to results_to_token_srt to get the word timings. FYI -
The "word" timestamps are actually token timestamps. Since token:word is not always 1:1 (varies by language), you may need to do some additional processing to get individual word timings.
Thanks, that's super helpful!
word_timestamps = results['segments']['word_timestamps'] TypeError: list indices must be integers or slices, not str
that was a mistake on my part for not updating the execution lines in readme
word_timestamps are under each segment, instead of the segments which is a list
so the word_timestamps for first segments would be:
first_segment_token_timestamps = results['segments'][0]['word_timestamps']
if you want a list of just the word_timestamps
word_timestamps = [seg['word_timestamps'] for seg in results['segments']]
|
gharchive/issue
| 2022-10-03T06:32:19
|
2025-04-01T06:39:12.679419
|
{
"authors": [
"ClashSAN",
"jak3-taylor",
"jianfch"
],
"repo": "jianfch/stable-ts",
"url": "https://github.com/jianfch/stable-ts/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
224355012
|
在uxcore中使用jsonp方式获取数据会报错
SyntaxError: Unexpected token :
具体什么错?
不好意思,忘了jsonp要求的返回数据格式不是json。已经通过修改header实现跨域访问。
|
gharchive/issue
| 2017-04-26T06:16:50
|
2025-04-01T06:39:12.687635
|
{
"authors": [
"isoloist",
"jias"
],
"repo": "jias/natty-fetch",
"url": "https://github.com/jias/natty-fetch/issues/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2350364198
|
✨ チーム開発の資料を追加
概要
インターンで使用する資料を追加しました
変更点
README.mdの追加
docs.mdの追加
drawi.o integrationをrecommendationsに追加
スクラムマスターがlinterにかかってしまったのでfilterを追加した
@dicenull 演習と目次を修正完了しました
|
gharchive/pull-request
| 2024-06-13T07:22:21
|
2025-04-01T06:39:12.693977
|
{
"authors": [
"kimflip"
],
"repo": "jigintern/intern-dev-tutorial",
"url": "https://github.com/jigintern/intern-dev-tutorial/pull/41",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
145102350
|
subregion_select dropdown has value numeric instead of alphabet code
Hi there,
As part of testing just observed that for country Denmark, the subregion select dropdown has value numeric.
State
Hovedstaden
Midtjylland
Nordjylland
Sjælland
Syddanmark
On saving it, the same numeric value is being saved in table.
Is it normal behaviour, any fix will be really appreciated.
Those are the region codes as found by the iso-codes debian package so this is expected behavior unless that codes have changed since the last time the script was run in the carmen gem.
Select tag looks like this:
Hovedstaden
Midtjylland
Nordjylland
Sjælland
Syddanmark
So on saving value, 84 saves in the database instead of state code..
|
gharchive/issue
| 2016-04-01T06:54:55
|
2025-04-01T06:39:12.706799
|
{
"authors": [
"arun-manthani",
"ecbypi"
],
"repo": "jim/carmen-rails",
"url": "https://github.com/jim/carmen-rails/issues/58",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1319038763
|
cm-chessboard: Figure out the status of pawn promotion
Pawn promotion seems an issue with all of these “just a board” packages. See closed issue “ Pawn promotion support? #19 at cm-chessboard.
To the extent that I’m just replicating FEN strings, I don’t think there’s any problem. Where a need would arise is either (a) when I’m stepping through a main line (rather than using FENs) or (b) when the user is adding an line that has a promotion.
Closed as out of scope for this repository.
|
gharchive/issue
| 2022-07-27T05:26:19
|
2025-04-01T06:39:12.722428
|
{
"authors": [
"jimratliff"
],
"repo": "jimratliff/pgn4people-poc-demo",
"url": "https://github.com/jimratliff/pgn4people-poc-demo/issues/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
890704247
|
about the length of tokens
Hello,
I have read your paper and am quite interested in your work! There is a question about the tokens.
I notice you truncat the passage tokens with 120 in MSMARCO Passage Retrieval, however, for ANCE, the original paper uses 512 tokens. So does the number of tokens have the impact on the accuracy?
No
We just don't have that many advanced GPUs to afford very long input :(
BTW, the average passage length is about 70 tokens and truncating to 120 shouldn't be a problem.
hhhhhhhhha!
Thanks a lot for your reply! I have no multiple GPUs either! Lol!
I agree with your settings!
|
gharchive/issue
| 2021-05-13T05:00:39
|
2025-04-01T06:39:12.792086
|
{
"authors": [
"KaishuaiXu",
"jingtaozhan"
],
"repo": "jingtaozhan/DRhard",
"url": "https://github.com/jingtaozhan/DRhard/issues/6",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
39788993
|
Set default value while AutoMigrate
If no default value, the query with gorm will get error.
Hi @wb14123
I just pushed a commit, so even the data in database is NULL, gorm won't get any error now.
So we don't need the default value for migrations
Thank you.
|
gharchive/issue
| 2014-08-08T03:08:14
|
2025-04-01T06:39:12.799155
|
{
"authors": [
"jinzhu",
"wb14123"
],
"repo": "jinzhu/gorm",
"url": "https://github.com/jinzhu/gorm/issues/190",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
97720510
|
Fixed issue https://github.com/jinzhu/gorm/issues/151 where postgresql connection problems fail silently
For reference: https://github.com/jinzhu/gorm/issues/151
Errors connection to a postgresql fail silently, resulting in errors. This fix sends a ping right after the connection has been made. If the ping fails, we assume the connection is broken.
Accepting better solutions. This quick fix did it for me :)
Should we always Ping() the database after connection?
It's not the most pretty solution. But it an extra safetycheck to verify that the connection is in fact alive.
Should i rewrite the pr and make the Ping() generic?
Hi @pariz
Yes, I think it is a better solution than just check the postgres.
I agree. Code has been updated. Epic oneliner! :)
picked your code to the master, thank you.
|
gharchive/pull-request
| 2015-07-28T14:32:53
|
2025-04-01T06:39:12.802452
|
{
"authors": [
"jinzhu",
"pariz"
],
"repo": "jinzhu/gorm",
"url": "https://github.com/jinzhu/gorm/pull/585",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
200065337
|
datas on ajax call not work correctly
Problemes
1- The filter don't work.
2- After filtering with a range then the datas are correctly loaded.
I don't have the probleme when in load datas on external files data (places.js)
Please help me:
Here is a part of my code:
$(document).ready(function(){
var places = [];
$.ajax({
type: 'GET',
contentType: "json",
url: 'http://www.mydomaine.com/client/places',
async: false,
success:function(data){
places = data;
}
});
renderCategories();
initInputs();
var place = places[0];
GoogleMap.init(place.coordinate.latitude, place.coordinate.longitude, places);
var afterFilter = function(result){
$('#total_places').text(result.length);
GoogleMap.updateMarkers(result);
}
afterFilter(places);
//search: {ele: '#searchbox', fields: ['runtime']}, // With specific fields
var FJS = FilterJS(places, '#places', {
template: '#place-template',
search: {ele: '#searchbox', fields: ['name', 'categories','disponible','regionId','departementId']},
callbacks: {
afterFilter: afterFilter
}
});
FJS.addCriteria({field: 'rating', ele: '#rating_filter', type: 'range'});
FJS.addCriteria({field: 'disponible', ele: '#is_closed_criteria input:checkbox'});
FJS.addCriteria({field: 'categories', ele: '#categories_criteria input:checkbox', all: 'all'});
FJS.addCriteria({field: 'regionId', ele: '#region_criteria input:checkbox', all: 'all'});
FJS.addCriteria({field: 'departementId', ele: '#region_criteria input:checkbox', all: 'all'});
window.FJS = FJS;
});
I'm having trouble understanding your problem.
The Filter doesn't seem to give any results until you actually change the range?
Are you initializing it with the min and max values from the start?
Code looks correct, but can you send me your site url or HTML. It will help me to debug i.e How you initializing your range filters.
I correct the problems : it's comes from the initiating of the range.
Thanks a lot.
|
gharchive/issue
| 2017-01-11T11:11:26
|
2025-04-01T06:39:12.810632
|
{
"authors": [
"Acen",
"diouf83",
"jiren"
],
"repo": "jiren/filter.js",
"url": "https://github.com/jiren/filter.js/issues/143",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1308399302
|
Consider adding compute unit limit to bundles
Problem
Proposed Solution
@buffalu What should the limit be?
|
gharchive/issue
| 2022-07-18T19:12:55
|
2025-04-01T06:39:12.822956
|
{
"authors": [
"anoushk1234",
"buffalu"
],
"repo": "jito-foundation/jito-solana",
"url": "https://github.com/jito-foundation/jito-solana/issues/81",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
303241067
|
[iOS] Fix uncaught NSInvalidArgumentException in RTCPeerConnection's createAnswer
WebRTC appears to fail RTCPeerConnection's createAnswer without an
error message i.e. nil. But NSMutableDictionary does not accept nil as
a value and causes an uncaught exception.
Based on Daniel Ornelas' comment that the "error" key should actually be "NSLocalizedDescription" and looking at the error.localizedDescription, I suppose it's better to morph this one further.
ok, let me know when it's ready to be merged
|
gharchive/pull-request
| 2018-03-07T20:05:38
|
2025-04-01T06:39:12.856594
|
{
"authors": [
"lyubomir",
"paweldomas"
],
"repo": "jitsi/react-native-webrtc",
"url": "https://github.com/jitsi/react-native-webrtc/pull/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
138081139
|
Spaces vs Tabs
Why are you using spaces now? Tabs makes the code look more clean and organized in my opinion.
I'm not sure how its cleanliness/organization is adversely affected by this change, could you elaborate on that?
I decided to switch to spaces mainly because of the inconsistency between different programs/websites/browsers/users. Through some of these lenses tabs come out as long as 8 spaces (rather than the standard 4, in VS), requiring more scrolling or resulting in more code flowing onto subsequent lines (much less readable), can be especially painful on phones/small resolution devices.
Tabs also caused problems with 'pretty spaced' code (where params/array initializers are lined up nicely for easy reading) which could easily end up ugly. Admittedly the biggest issue here was that tabs were used for more than just the start-of-line to start-of-code portions; but making this change certainly helped identify/eliminate those that were in the middle and makes it very easy to do so in the future. It's also easier to find stray tabs than stray spaces, if keeping the project on a 100% consistent convention ever became relevant (using different forms I’ve heard, can cause lotsa merge issues).
Productivity wise I find the two indentation options to be quite equivalent, Visual Studios can easily be set to turn tab key-presses into 4-spaces (Tools->Options->Text Editor->C/C++->Tabs); white space can be easily jumped with ctrl+arrows or deleted with ctrl+delete and inserting or re-inserting a ; or } automatically spaces chunks of code.
For tabs the only advantages I read up on and considered legitimate was the ability for users to set their size (in some programs, not all give this option) and the relative ease for newer coders - which didn't, in my opinion, outweigh their disadvantages.
Some discussion on the matter:
http://programmers.stackexchange.com/questions/57/tabs-versus-spaces-what-is-the-proper-indentation-character-for-everything-in-e
It appears I can't link SEN shoutbox convos, so http://www.staredit.net/shoutboxsearch/ + starter phrase "like my friend and I decide for our project that indentation is to be this many spaces" then hit context.
|
gharchive/issue
| 2016-03-03T05:38:56
|
2025-04-01T06:39:12.860669
|
{
"authors": [
"jjf28",
"xboi209"
],
"repo": "jjf28/Chkdraft",
"url": "https://github.com/jjf28/Chkdraft/issues/92",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1270145971
|
Method for creating and printing board
The user needs a way to put a sudoku board into the program. Boards should be printed in a way that is easy to read.
A graphical user interface will be created using Swing. The GUI will allow for a user to create a sudoku board to be solved, as well as display solutions.
|
gharchive/issue
| 2022-06-14T01:45:35
|
2025-04-01T06:39:12.881072
|
{
"authors": [
"jjklemm"
],
"repo": "jjklemm/sudoku-solver",
"url": "https://github.com/jjklemm/sudoku-solver/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1868932522
|
Sub listed item in numbered list in 01-jupyter.ipynb
Please check the numbered list in one of the first markdown cells of the notebook. For the sub listed item I think one more space in front of the "1." is necessary, otherwise the item just becomes part of the main list.
fixed. thanks.
|
gharchive/issue
| 2023-08-28T04:23:24
|
2025-04-01T06:39:12.889890
|
{
"authors": [
"ccolomb2",
"jkitchin"
],
"repo": "jkitchin/f23-06623",
"url": "https://github.com/jkitchin/f23-06623/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
148340174
|
Update code: issue #2 Updating model from controller does not update control
Can you take this solution along in your code? Thanks
scope.$watch(attrs['ngModel'], function (newTime) {
scope.initTime = newTime;
reinitTime();
});
Spot on! This fix was exactly what i was looking for.
Can someone make a PR please?
I had already made a PR: https://github.com/jkuri/ngTimepicker/pull/4
Sorry, I missed it. Thanks for your PR, it's merged.
|
gharchive/issue
| 2016-04-14T12:05:38
|
2025-04-01T06:39:12.910528
|
{
"authors": [
"elineopsommer",
"jkuri",
"leypoldt"
],
"repo": "jkuri/ngTimepicker",
"url": "https://github.com/jkuri/ngTimepicker/issues/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1478454439
|
Lack of support for valid endpoints
Consider the following code from Usage with Express
getUser: t.procedure.input(z.string()).query((req) => {
req.input; // string
return { id: req.input, name: 'Bilbo' };
}),
After adding this line
.meta({ openapi: { method: 'GET', path: '/getUser' } })
trpc-openapi fail to process it
TRPCError: [query.getUser] - Input parser must be a ZodObject
This is correct.
- z.string()
+ z.object({ id: z.string() })
This is correct.
You have linked to the tRPC documentation.
trpc-openapi has a few additional requirements so you may need to make a few changes to your procedure.
- z.string()
+ z.object({ id: z.string() })
Please make sure you have read this 👉 https://github.com/jlalmes/trpc-openapi/blob/master/README.md
Well, it's string and not object (see the // string comment - it's there, in the documentation, not something I added), so it can't be correct.
And thing is that we not only need to change the procedure, but change API requests as well, which is too high price to pay.
Hi Sergey. Please read the README before commenting again.
so it can't be correct
trpc-openapi requires that your inputs are wrapped in an z.object.
need to change the procedure
Yes.
change API requests
I don't think you've understood REST.
Consider the following procedure.
getUser: t.procedure
.meta({ openapi: { method: 'GET', path: '/user' } })
.input(z.object({ userId: z.string() }))
.output(z.object({ id: z.string(), name: z.string() }))
.query(({ input }) => {
console.error(typeof req.input.userId);
return { id: req.input, name: 'Bilbo' };
}),
This is how it can be invoked.
@trpc/server 👉 GET /getUser?input=${encodeURI(JSON.stringify({ userId: 'abc123' }))}
trpc-openapi 👉 GET /user?userId=abc123
too high price to pay
That's your decision to make.
|
gharchive/issue
| 2022-12-06T07:35:26
|
2025-04-01T06:39:12.921152
|
{
"authors": [
"elmeister",
"jlalmes"
],
"repo": "jlalmes/trpc-openapi",
"url": "https://github.com/jlalmes/trpc-openapi/issues/196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
788965082
|
SSL support
Is it possible to attach to a kafka instance with SSL enabled? For example, when I'm attaching to my cluster as a producer, I'll use the following:
KafkaProducer(bootstrap_servers=kafkaServer, security_protocol="SSL", ssl_check_hostname=True)
When attaching to the kafka cluster through this extension, I can specify the bootstrap server and port (443 in my case), but I can't seem to find a way to change the security protocol.
Currently SSL is only enabled when using an authentication scheme, i.e. you have a username/password.
So you need SSL but no auth?
Currently SSL is only enabled when using an authentication scheme, i.e. you have a username/password.
So you need SSL but no auth?
Yes, currently we don't have any auth, just SSL. I tried to specify SASL/Plain as authentication but with a blank username password, but that doesn't seem to work.
Yes, currently we don't have any auth, just SSL. I tried to specify SASL/Plain as authentication but with a blank username password, but that doesn't seem to work.
Ok, so your use case seems quite simple but if we want to do SSL support right (and advertise we support SSL), we need to be able configure SSL certificate configuration, as per https://kafka.js.org/docs/configuration#ssl
Currently, the ssl boolean is only set here: https://github.com/jlandersen/vscode-kafka/blob/a0257d0932263cc66bce73c5579a9d530f9655b0/src/client/client.ts#L325
Ok, so your use case seems quite simple but if we want to do SSL support right (and advertise we support SSL), we need to be able configure SSL certificate configuration, as per https://kafka.js.org/docs/configuration#ssl
Currently, the ssl boolean is only set here: https://github.com/jlandersen/vscode-kafka/blob/a0257d0932263cc66bce73c5579a9d530f9655b0/src/client/client.ts#L325
Brilliant! That works well for my purposes, it connects fine and seems to be working well.
Brilliant! That works well for my purposes, it connects fine and seems to be working well.
|
gharchive/issue
| 2021-01-19T12:02:03
|
2025-04-01T06:39:12.927371
|
{
"authors": [
"fbricon",
"mike-england"
],
"repo": "jlandersen/vscode-kafka",
"url": "https://github.com/jlandersen/vscode-kafka/issues/84",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1861984934
|
Cannot see files in folders since Container Manager feature
Current Behavior
Hello
Since the appearance of Container Manager instead of Docker package I noticed that the backup was constantly stuck at 0 files found (0mb) which is not normal. So I browsed through files and noticed that folders and files were not visible. The permissions are good (USER_ID=0 and GROUP_ID=0).
The problem happened after a reboot following an image update.
I noticed that by restarting the container from the command line, the problem was gone.
Does anyone also have the problem?
What to do to continue to benefit from the evolutions of Container Manager (simple update feature) without having to restart systematically on the command line.
I specify that the bug is in no way related to the good work of our dear JLesage ;-)
Expected Behavior
No response
Steps To Reproduce
No response
Environment
OS:
OS version:
CPU:
Docker version:
Device model:
Browser/OS:
Container creation
Container Manager package
Container log
no abnormal log
Container inspect
No response
Anything else?
No response
I assume you are using a Synology NAS?
Looks like the issue where /volume1 is mapped to /storage (see #310). Is it the case?
Yes I use a Synology NAS but the issue not seems to be the same as #310 tab.
I will try to make new tests later
You can also share the output of docker inspect <container name> so we can have a complete picture of the container's configuration.
I deleted the container and created a new one. Issue seems gone when container started from Container Manager
Hello,
Now I get 1,5 years to backup 4 TB. How to increase backup speed ? My ISP speed is about 700 mbps but crashplan backup is really slow.
Yes, it is known that upload speed is slow. However, because of deduplication, a lot of data don't really have be uploaded. If you look at Tools->History, you can see the "Effective rate": the value can be very high and would typically improve over time.
I had the same problem.
In fact, when the docker container was created (a long time ago), the storage volume was not requested.
Since the last update, not only is it requested, but if it hadn't been configured originally, it points to an empty virtual folder.
I installed portainer to change the configuration and set storage bind to volume1.
Since then, I've been able to browse storage, but for crashplan, it's not the same folder structure. As a result, the backup started from scratch. I hope that deduplication will play its part, otherwise I'm in for several months of backups.
|
gharchive/issue
| 2023-08-22T18:30:20
|
2025-04-01T06:39:12.950399
|
{
"authors": [
"Bagu",
"jlesage",
"walawa75"
],
"repo": "jlesage/docker-crashplan-pro",
"url": "https://github.com/jlesage/docker-crashplan-pro/issues/438",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
442215161
|
Use custom DB instance
Hi, thanks a lot to share this ,it is really great.
I'm wandering can we use our custom DB instance (like MySQL or Maria DB) instead of MySQL in container ,thanks.
If you want to use your own DB instance, look at the original project:
https://github.com/jc21/nginx-proxy-manager
|
gharchive/issue
| 2019-05-09T12:38:42
|
2025-04-01T06:39:12.952170
|
{
"authors": [
"KevinZhangt",
"jlesage"
],
"repo": "jlesage/docker-nginx-proxy-manager",
"url": "https://github.com/jlesage/docker-nginx-proxy-manager/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
583876334
|
Let's Encrypt: Internal Error
I used the hassio addon for quite some time now. Today I tried to migrate the proxy manager to my synology. For that I use the synology docker manager and this docker image.
So I installed the latest version (1.7.0) and configured my port forwarding. The only issue I had was with the lets encrypt certificate. When I tried to request a certificate, I got an "Internal error":
This is the content in the /config/log/letsencrypt/letsencrypt.log:
2020-03-18 17:02:07,145:DEBUG:certbot.main:certbot version: 0.30.2
2020-03-18 17:02:07,147:DEBUG:certbot.main:Arguments: ['--non-interactive', '--config', '/etc/letsencrypt.ini', '--cert-name', 'npm-15', '--agree-tos', '--email', 'mail@zoker.me', '--preferred-challenges', 'dns,http', '--webroot', '--domains', 'wg.zkr.io']
2020-03-18 17:02:07,150:DEBUG:certbot.main:Discovered plugins: PluginsRegistry(PluginEntryPoint#manual,PluginEntryPoint#null,PluginEntryPoint#standalone,PluginEntryPoint#webroot)
2020-03-18 17:02:07,202:DEBUG:certbot.log:Root logging level set at 20
2020-03-18 17:02:07,204:INFO:certbot.log:Saving debug log to /config/log/letsencrypt/letsencrypt.log
2020-03-18 17:02:07,206:DEBUG:certbot.plugins.selection:Requested authenticator webroot and installer None
2020-03-18 17:02:07,220:DEBUG:certbot.plugins.selection:Single candidate plugin: * webroot
Description: Place files in webroot directory
Interfaces: IAuthenticator, IPlugin
Entry point: webroot = certbot.plugins.webroot:Authenticator
Initialized: <certbot.plugins.webroot.Authenticator object at 0x7fded3100a90>
Prep: True
2020-03-18 17:02:07,223:DEBUG:certbot.plugins.selection:Selected authenticator <certbot.plugins.webroot.Authenticator object at 0x7fded3100a90> and installer None
2020-03-18 17:02:07,224:INFO:certbot.plugins.selection:Plugins selected: Authenticator webroot, Installer None
2020-03-18 17:02:07,320:DEBUG:acme.client:Sending GET request to https://acme-v02.api.letsencrypt.org/directory.
2020-03-18 17:02:07,326:DEBUG:urllib3.connectionpool:Starting new HTTPS connection (1): acme-v02.api.letsencrypt.org
2020-03-18 17:02:12,333:DEBUG:certbot.log:Exiting abnormally:
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 141, in _new_conn
(self.host, self.port), self.timeout, **extra_kw)
File "/usr/lib/python3.6/site-packages/urllib3/util/connection.py", line 60, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "/usr/lib/python3.6/socket.py", line 745, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -3] Try again
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 601, in urlopen
chunked=chunked)
File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 346, in _make_request
self._validate_conn(conn)
File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 850, in _validate_conn
conn.connect()
File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 284, in connect
conn = self._new_conn()
File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 150, in _new_conn
self, "Failed to establish a new connection: %s" % e)
urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/requests/adapters.py", line 445, in send
timeout=timeout
File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 639, in urlopen
_stacktrace=sys.exc_info()[2])
File "/usr/lib/python3.6/site-packages/urllib3/util/retry.py", line 388, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='acme-v02.api.letsencrypt.org', port=443): Max retries exceeded with url: /directory (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/bin/certbot", line 11, in <module>
load_entry_point('certbot==0.30.2', 'console_scripts', 'certbot')()
File "/usr/lib/python3.6/site-packages/certbot/main.py", line 1364, in main
return config.func(config, plugins)
File "/usr/lib/python3.6/site-packages/certbot/main.py", line 1233, in certonly
le_client = _init_le_client(config, auth, installer)
File "/usr/lib/python3.6/site-packages/certbot/main.py", line 604, in _init_le_client
acc, acme = _determine_account(config)
File "/usr/lib/python3.6/site-packages/certbot/main.py", line 521, in _determine_account
config, account_storage, tos_cb=_tos_cb)
File "/usr/lib/python3.6/site-packages/certbot/client.py", line 181, in register
acme = acme_from_config_key(config, key)
File "/usr/lib/python3.6/site-packages/certbot/client.py", line 51, in acme_from_config_key
return acme_client.BackwardsCompatibleClientV2(net, key, config.server)
File "/usr/lib/python3.6/site-packages/acme/client.py", line 814, in __init__
directory = messages.Directory.from_json(net.get(server).json())
File "/usr/lib/python3.6/site-packages/acme/client.py", line 1152, in get
self._send_request('GET', url, **kwargs), content_type=content_type)
File "/usr/lib/python3.6/site-packages/acme/client.py", line 1101, in _send_request
response = self.session.request(method, url, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/requests/sessions.py", line 512, in request
resp = self.send(prep, **send_kwargs)
File "/usr/lib/python3.6/site-packages/requests/sessions.py", line 622, in send
r = adapter.send(request, **kwargs)
File "/usr/lib/python3.6/site-packages/requests/adapters.py", line 513, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='acme-v02.api.letsencrypt.org', port=443): Max retries exceeded with url: /directory (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again',))
2020-03-18 17:02:12,340:ERROR:certbot.log:An unexpected error occurred:
Is this an issue with this image or with the base image or is my configuration wrong?
Is this issue related?
https://github.com/jc21/nginx-proxy-manager/issues/180#issuecomment-599278001
Thank you!
Edit:
A little bit more information: I also tried version 1.6.0, but it has the same issue.
For both version: Connecting to my devices trough the manager using port 80 works fine. Connecting to them via 443 leads to a empty response.
Looks like the container is not able to reach the LetEncrypt server.
Can you try:
docker exec <container name> ping acme-v02.api.letsencrypt.org
Hmm yes, something does not seem right:
When I try the same with my AdGuard container it works:
AdGuard is running in host mode, while ProxyManager is running in bridge mode. But bridge mode should work as well, right?
Do you have any idea, how I can fix that?
Are you using AdGuard as your DNS server ? If yes. this may be the cause of the issue.
Are you able to ping acme-v02.api.letsencrypt.org from your Synology ?
Yes I'm using it as my DNS server.
Yes the ping from synology itself works:
Is this DNS server also used by you Synology? You can try to run the following command on your Synology: nslookup acme-v02.api.letsencrypt.org <DNS server IP>.
Also, can you confirm that this DNS server used by the container:
docker exec <container name> cat /etc/resolv.conf
I am having the same internal error message. I use dnsimple and I do have a certificate on my account for my domain. Pinging the address you said above is successful.
If your container accessible from the internet on port 80 ? You can check with https://www.yougetsignal.com/tools/open-ports/
Same issue here. I am able to ping with:
docker exec ping acme-v02.api.letsencrypt.org
No internal DNS beyond a redirect to an external PiHole server. I tried both requesting a new cert (internal error) and creating my own via a LetsEncrypt container. Both no dice.
Make sure the container is reachable from the internet on port 80. You can check with https://www.yougetsignal.com/tools/open-ports/
Morning. Port shows as not reachable but both my router settings and the use of the proxy manager is present over port 80. If I use the IP:80 I go directly to where I set redirect to.
443 shows open and listening. Could I just direct through that? 80 is still working fine, It appears my isp blocks port 80, though I can go it via both directly through the IP and via my duckdns dyndns.
If your ISP blocks port 80, then I guess your cannot use Nginx Proxy Manager with automatic certificate generation from LetsEncrypt. Port 80 is required. This is how LetsEncrypt validates that your are the owner of the DNS name.
Odd given I can use port 80 for the hosts from Nginx Proxy Manager. I'll do some more research and report back on a fix for future reference.
|
gharchive/issue
| 2020-03-18T17:06:22
|
2025-04-01T06:39:12.966448
|
{
"authors": [
"TheZoker",
"bookandrelease",
"jlesage",
"mdisieno"
],
"repo": "jlesage/docker-nginx-proxy-manager",
"url": "https://github.com/jlesage/docker-nginx-proxy-manager/issues/57",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
199320982
|
[WIP] A Proposal For Testing Things
As we talked about in https://github.com/jlord/git-it-electron/pull/178#issuecomment-270530180, it'd be nice to be able to run the verify tests automagically. This PR is my attempt to tackle the problem, which required some refactoring of how the verify scripts work.
Current Limitations
A couple of issues that I've had to work through:
verify scripts touch the DOM as part of rendering the results - I've attempted to move this code back up to the challenge script, but it's only done for one verify script so far (the repository tests).
some verify scripts are written as if they are blocking but due to using exec they need to be asynchronous - we need a way to inspect the results, so I've introduced a callback that takes an array of results. Maybe you have other ideas in mind.
How This Works
We're using tape to run scripts (I'm more familiar with mocha, so I'm open to other ways to get the best out of the tape API), but here's a simple test:
test('verifies the hello world repository', function (t) {
var folder = helper.extractFixture('hello-world')
verify(folder, function (result) {
var expected = 'This is a Git repository!'
var first = result[0]
t.ok(first.result)
t.equal(expected, first.message)
t.end()
})
})
For scenarios where we need the repository to be in a certain state, we can splat the folders onto disk under the tests/fixtures folder - and then extract them to a temporary folder when running the tests. To get this working, you can just rename the .git folder to _git and the repository state is then version-controllable.
The other notable change here is that we have a callback to poke at the results from the verify script (because exec is asynchronous), to ensure we're getting the expected results. I'm using the same shape of "message is a string" and "result is a boolean" that the existing verify scripts use.
To get this working in a test harness, I needed to move the code for addToList and challengeIncomplete/markChallengeCompleted out from the verify script to where the verify method is invoked. A partial implementation of this in challenge.js works for the repository script, but it definitely needs more work and testing as I might not be understanding it correctly.
What's Left To Do?
A rough list:
[ ] get feedback on this approach
[ ] sketch out the changes to test a path-less script, update challenge.js to handle this
[ ] propose a way to test request invocations (nock? something else?)
[ ] ???
verify scripts touch the DOM as part of rendering the results
Yeah, this is something I've wanted to improve; to separate these concerns. Which is probably a big enough project in itself to be a different PR and a thing to figure out before this?
I haven't gone through all this with a fine tooth comb but in general I feel like it's what I've been thinking! 👍
I have a couple things left to clean up, cause I don't think master is quite stable right now but then I'll pick up from what you've started here in moving out the DOM touching stuff and I'll get a PR going for that.
✨
:cool:
Which is probably a big enough project in itself to be a different PR and a thing to figure out before this?
Up to you - if you feel confident enough to do that refactoring before introducing tests then :thumbsup:
|
gharchive/pull-request
| 2017-01-07T00:08:40
|
2025-04-01T06:39:13.001096
|
{
"authors": [
"jlord",
"shiftkey"
],
"repo": "jlord/git-it-electron",
"url": "https://github.com/jlord/git-it-electron/pull/180",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
79449040
|
View not being update with angular-cache with localStorage
I'm building an Ionic app the uses angular-cache to leverage data persistence. This data is displayed in a ng-repeat list. Either by user interaction (in another view!) or new data being loaded in the background, I need the view to update to reflect this.
When developing my app everything works perfectly, however, as soon as I start using angular-cache with localStorage, views stop being updated. When I turn off localStorage (but still use in memory caching from angular-cache) everything works as expected: the data updates and so is the DOM.
I was able to reproduce the issue with this CodePen http://codepen.io/anon/pen/qdayMW. To demonstrate the issue you switch tabs (from home to about). If you look in the console you can see that new items are being added to the localStorage cache. However the homescreen doesn't update. If you then refresh the page (within the 10 second cache magAge) the page shows the correct list of items...
However, if we comment out line 81, the caching is no longer persistent after a refresh, however the new items are added to the DOM every time the view changes.
Note: I also posted this issue to the Ionic forum as I'm not sure which exactly is the root of the problem: http://forum.ionicframework.com/t/view-not-being-update-with-angular-cache-with-localstorage/25236
This must be something weird with ionic. I've been playing with your codepen, and as I click between the home and about tabs, the home controller code is only executed once at the very beginning, so while more items are being added to the cache, the resolve code that would get the latest items isn't run again.
The latest ionic implemented view caching (which is causing your "bug"), which you have to disable: http://ionicframework.com/docs/nightly/api/directive/ionNavView/
|
gharchive/issue
| 2015-05-22T13:08:57
|
2025-04-01T06:39:13.013024
|
{
"authors": [
"danielcha",
"jmdobry"
],
"repo": "jmdobry/angular-cache",
"url": "https://github.com/jmdobry/angular-cache/issues/181",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
909577463
|
[Bug] Kitsu anime tracking not working
Device information
Tachiyomi version: Latest stable and preview version
Android version: 10
Device: Oneplus 5T
The kitsu anime tracking only shows the manga's and not the anime.
Checked it with myanimelist but that one does work, but it's not the site I use to keep my library updated.
Would be awesome if it can be fixed.
Keep up the great work!
forgot to close. was fixed with https://github.com/jmir1/aniyomi/commit/9e2db50b77fdd99578eddc55d07cc343bacb049b.
|
gharchive/issue
| 2021-06-02T14:41:20
|
2025-04-01T06:39:13.017310
|
{
"authors": [
"Reacien",
"jmir1"
],
"repo": "jmir1/aniyomi",
"url": "https://github.com/jmir1/aniyomi/issues/39",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1814461043
|
error loading model: unexpectedly reached end of file
On a couple of models I am receiving this error:
llama.cpp: loading model from /Users/REDACTED/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998
error loading model: unexpectedly reached end of file
llama_load_model_from_file: failed to load model
This happens with a couple of the larger models:
nous-hermes:latest
llama2:13b
If I do ollama pull against them, the manifests match up and it doesn't re-pull anything. Since this looks like docker under the hood, are the models corrupt? or?
Any thoughts? FWIW, llama2:latest and wizard-vicuna:latest work fine.
M2 Macbook Pro 32 Gigs of ram.
Thanks @bkruger99, will check out why this is happening
Let me know if you need any additional debugging data from my side. You'll have to tell me how to enable other than running server via cli :)
Great! @bkruger99 is this on Mac? Thanks!
Yes!
Hardware:
Model Name: MacBook Pro
Model Identifier: Mac14,10
Model Number: Z174000EBLL/A
Chip: Apple M2 Pro
Total Number of Cores: 12 (8 performance and 4 efficiency)
Memory: 32 GB
OS: Ventura 13.4.1 (c)
@bkruger99 can you run:
sha2 -256 ~/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998
Check to see that sha sum matches, and if it doesn't you can
rm ~/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998
and then re-pull the image. There's a fix that I think was just merged that will make certain the sha sum is verified correctly when you're pulling the layers.
Yeah. there's something w/ manifest not verifying the sha256 when pulling. These two models did have a network interruption as the laptop went to sleep.
❯ shasum -a 256 sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998
f2a1788633ddf3edef0ee4d9d4e93c399bfeeeb7363015d7c1b630ff268cdcf5 sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998
I re-pulled llama2:12b and it's happy, I'll do the same with the rest of 'em.
The next version will check the SHAs; the re-pull is pretty tolerant of network interruptions, but wondering if the buffer wrote garbage onto the end of the partial file somehow. I haven't (yet) tested with sleeping the machine though, so that could have been the reason.
I'm going to go ahead and close the issue. Feel free to re-open it though.
|
gharchive/issue
| 2023-07-20T17:15:59
|
2025-04-01T06:39:13.040101
|
{
"authors": [
"bkruger99",
"jmorganca",
"pdevine"
],
"repo": "jmorganca/ollama",
"url": "https://github.com/jmorganca/ollama/issues/143",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2768101
|
Raise ArgumentError when passing invalid keys to has_selector?
It's a relatively common misconception that you can use :content as a key - e.g.
has_selector?('h1', :content=>'wooo!')
which would silently pass, even when 'wooo!' is nowhere to be found. This change should make that error much more obvious.
Related issues : https://github.com/jnicklas/capybara/issues/525
It would be nice to raise an error on passing a block in - eg
@session.should have_selector("//p") do |p_tag|
p_tag.should have_selector("//a")
end
which seems to be a common error on people switching to capybara via webrat. However, that's a bit more tricky and would probably need fixing at the rspec-matcher level, since the block doesn't make it as far as has_selector? and gets silently discarded.
ValidKeys on Selector could be replaced with something like this :
diff --git a/lib/capybara/selector.rb b/lib/capybara/selector.rb
index 1bff91d..0b85a04 100644
--- a/lib/capybara/selector.rb
+++ b/lib/capybara/selector.rb
@@ -15,6 +15,10 @@ module Capybara
def remove(name)
all.delete(name.to_sym)
end
+
+ def valid_keys
+ all.map{|selector| selector.custom_filters.keys}.flatten.uniq
+ end
end
def initialize(name, &block)
?
I'm not sure what you mean about the implementation for checking query keys, though. There's 3 places that keys passed to has_selector?() are used - node/matchers.rb, query.rb, and selector.rb. Whichever one is responsible for checking the keys will have to also check the keys of the other 2. I could move assert_valid_options to a class method to Query if you like?
Closing this pull request, we already have an open issue for this in #525, and I don't like this implementation
|
gharchive/issue
| 2012-01-09T12:20:18
|
2025-04-01T06:39:13.102597
|
{
"authors": [
"jdelStrother",
"jnicklas"
],
"repo": "jnicklas/capybara",
"url": "https://github.com/jnicklas/capybara/issues/602",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1512494687
|
🛑 MegaConstruct is down
In 8d9ab08, MegaConstruct (https://www.megaconstruct.ro) was down:
HTTP code: 0
Response time: 0 ms
Resolved: MegaConstruct is back up in edc67a4.
|
gharchive/issue
| 2022-12-28T07:27:55
|
2025-04-01T06:39:13.109536
|
{
"authors": [
"joahn3"
],
"repo": "joahn3/earthlink-uptime",
"url": "https://github.com/joahn3/earthlink-uptime/issues/1865",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2025303728
|
🛑 CEI Sales Wiki is down
In 8230d21, CEI Sales Wiki (https://wiki.ceisales.com/login) was down:
HTTP code: 0
Response time: 0 ms
Resolved: CEI Sales Wiki is back up in 2b0ac7d after 14 minutes.
|
gharchive/issue
| 2023-12-05T05:13:31
|
2025-04-01T06:39:13.111958
|
{
"authors": [
"joahn3"
],
"repo": "joahn3/earthlink-uptime",
"url": "https://github.com/joahn3/earthlink-uptime/issues/2883",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2038683456
|
🛑 Laravel Enso is down
In c645c42, Laravel Enso (https://laravel-enso.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Laravel Enso is back up in fa7077b after 22 minutes.
|
gharchive/issue
| 2023-12-12T23:09:57
|
2025-04-01T06:39:13.114576
|
{
"authors": [
"joahn3"
],
"repo": "joahn3/earthlink-uptime",
"url": "https://github.com/joahn3/earthlink-uptime/issues/2892",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2471341990
|
🛑 CEI Sales App is down
In 83575f8, CEI Sales App (https://app.ceisales.com/login) was down:
HTTP code: 0
Response time: 0 ms
Resolved: CEI Sales App is back up in 2091cef after 20 minutes.
|
gharchive/issue
| 2024-08-17T05:05:41
|
2025-04-01T06:39:13.116945
|
{
"authors": [
"joahn3"
],
"repo": "joahn3/earthlink-uptime",
"url": "https://github.com/joahn3/earthlink-uptime/issues/3630",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1253103493
|
🛑 Laravel Enso is down
In 8a5476a, Laravel Enso (https://laravel-enso.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Laravel Enso is back up in 654d5f1.
|
gharchive/issue
| 2022-05-30T20:38:24
|
2025-04-01T06:39:13.119241
|
{
"authors": [
"joahn3"
],
"repo": "joahn3/earthlink-uptime",
"url": "https://github.com/joahn3/earthlink-uptime/issues/719",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
860806476
|
[Feature Request] Change blue loading and badge accent colors in Midnight Dusk theme to something similar to J2K Midnight Dusk?
Why/User Benefit/User Problem
Improves the theme by keeping the color scheme.
What/Requirements
The blue accent (and possibly black background on loading) would be changed to better fit the Midnight Dusk theme.
SY Midnight Dusk
J2K Midnight Dusk
Why/User Benefit/User Problem
Improves the theme by keeping the color scheme.
What/Requirements
The blue accent would be changed to better fit the Midnight Dusk theme.
Someone can do that for the other themes too and why the text in black-red are red instead of white ? I think its better to just replace everything that blue to red and change the colour of the download badge to one that look good for each theme
Already addressed by main tachi
|
gharchive/issue
| 2021-04-19T00:31:09
|
2025-04-01T06:39:13.151373
|
{
"authors": [
"CaptainHalflife",
"ImRyley",
"jobobby04"
],
"repo": "jobobby04/TachiyomiSY",
"url": "https://github.com/jobobby04/TachiyomiSY/issues/277",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
157378264
|
async requests
Is any chance an option to use async requests? like asp.net core rc2 standard template
public async Task<IActionResult> Index()
{
return View(await _context.Reason.ToListAsync());
}
_context.Reason.ToPagedList(0, DefaultPageSize) does not have ToListAsync method
honestly PagedList is a completely separate thing from the pagertaghelper, pagertaghelper does not depend on PagedList and really the only reason pagedlist exists in this library is because it was based on Martijn Boland's MVCPaging project and it had a pagedlist which I copied. PagedList is kind of a hack for when you are retrieving more than a page of data but want to do paging against the result set.
I don't use PagedList at all myself and don't really care to try to support it or update it. There are other implementations of PagedList out there that anyone could also use, so it is completely independent from the pagertaghelper. From a brief search I found this pagedlist which has async support, but not sure it has a compatible nuget for aspnetcore rc2. But you can search and maybe find a better compatible one, I did not spend much time looking. Or if anyone is interested in implementing that and submitting a pull request I'm open to that as well.
I do async methods to get data in my own use of pagertaghelper, and I just return one page of data at a time from the database so I don't need any pagedlist.
maybe could do it like this:
public async Task<IActionResult> Index()
{
var data = await _context.Reason.ToListAsync();
return View(data.ToPagedList(0, DefaultPageSize));
}
but again that looks like it returns more than a page of data and then uses pagedlist to page over it which is what PagedList is for but also why PagedList is a hack compared to just return a page of data from the database using .Skip .Take etc
only thing I've ever used PagedList for is the demo pages because it was convenient to make a demo without any database, and much of the demo code was also copied from Martijn's project and that is how PagedList ended up in this project. I never use that PagedList in my own projects, I only use the taghelpers
hm, sorry I a bit confused here. if you dont use PagedList what you use in your Pagination project? I mean then you retrieve data from DB.
you can see an example here a controller which has methods for user admin and paging of users with search, and the corresponding partial view here
as I said the use of PagedList in the pager demo pages was mainly because I copied existing demo pages and that makes it easy to have demo without any db, you just make a big list of data and use pagedlist to page over it. but again PagedList is only for if you already retrieved (or created as in the demo code) more than a page of data and now you want to page over that data. It is much better to not use PagedList and only retrieve data one page at a time from the db. PageTagHelper doesn't know about PagedList at all.
not that in the example I linked, you will see model.Paging.* properties being set, in that model Paging is an instance of PaginationSettings it is simply a model to pass settings on to the pagertaghelper
this makes me think I should get rid of PagedList and not use it in the demos because it makes people think they need to use that when in fact it is best not to
fyi, I've reworked the demo app and removed all use of PagedList from the demo so as not to encourage use of that
Many tnks. checking
|
gharchive/issue
| 2016-05-29T09:46:12
|
2025-04-01T06:39:13.165262
|
{
"authors": [
"dvdobrovolskiy",
"joeaudette"
],
"repo": "joeaudette/cloudscribe.Web.Pagination",
"url": "https://github.com/joeaudette/cloudscribe.Web.Pagination/issues/12",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
301313799
|
Multiple connections to the same instance with different DBs not working
Hello,
I'm try to run the docker container with the REDIS_HOSTS env variable witch looks like this:
REDIS_HOSTS='dev:redis:6379:0,test:redis:6379:10'
but it seems that the given database is ignored, since the UI shows me for both connections the database 0. This leads to an instable system, since after you click on one of the two connections the the view stucks on "loading ...." instead of showing the connection parameters.
I also tried to configure multiple connections via REDIS_HOSTS environment variable. No luck for me too.
The undefined entries at the bottom are also a bit irritating.
@hirnschmalz You are right, i've seen this too but i forgot to mentioned the undifined fields
@tburschka The latest image solved this issue
@joeferner This issue is back in the latest docker image. 👎
Seems to be an issue with the with the change of the namespace @awearsolutions/redis-commander. It's a bit annoying to create a PR to fix a problem which is back one version later.
Should be working with latest github/docker versions. At least we use it without any problem...
Can issue be closed?
This problem should be fixed in current version.
If you think its not feel free to reopen issue with updated informations based on latest version.
|
gharchive/issue
| 2018-03-01T08:17:05
|
2025-04-01T06:39:13.176382
|
{
"authors": [
"hirnschmalz",
"sseide",
"tburschka"
],
"repo": "joeferner/redis-commander",
"url": "https://github.com/joeferner/redis-commander/issues/236",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
573023714
|
Docker Container image for PytaskIO
Make a PyTaskIO Docker image available to download from the Docker registry.
Also update the README with the docker registry install details.
|
gharchive/issue
| 2020-02-28T21:49:43
|
2025-04-01T06:39:13.182106
|
{
"authors": [
"joegasewicz"
],
"repo": "joegasewicz/pytask-io",
"url": "https://github.com/joegasewicz/pytask-io/issues/28",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1090171830
|
v0.30.1 won't publish - issue with dependency
When running cargo publish, the pin-project fails to build, because of some issue with a macro:
error[E0659]: `parse_quote_spanned` is ambiguous (`macro_rules` vs non-`macro_rules` from other module)
--> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pin_project/derive.rs:859:67
|
859 | proj_generics.make_where_clause().predicates.push(parse_quote_spanned! { span =>
| ^^^^^^^^^^^^^^^^^^^ ambiguous name
|
note: `parse_quote_spanned` could refer to the macro defined here
--> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/utils.rs:22:1
|
22 | / macro_rules! parse_quote_spanned {
23 | | ($span:expr => $($tt:tt)*) => {
24 | | syn::parse2(quote::quote_spanned!($span => $($tt)*)).unwrap_or_else(|e| panic!("{}", e))
25 | | };
26 | | }
| |_^
note: `parse_quote_spanned` could also refer to the macro imported here
--> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pin_project/derive.rs:7:5
|
7 | *,
| ^
= help: use `self::parse_quote_spanned` to refer to this macro unambiguously
error[E0659]: `parse_quote_spanned` is ambiguous (`macro_rules` vs non-`macro_rules` from other module)
--> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pinned_drop.rs:108:21
|
108 | *path = parse_quote_spanned! { path.span() =>
| ^^^^^^^^^^^^^^^^^^^ ambiguous name
|
note: `parse_quote_spanned` could refer to the macro defined here
--> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/utils.rs:22:1
|
22 | / macro_rules! parse_quote_spanned {
23 | | ($span:expr => $($tt:tt)*) => {
24 | | syn::parse2(quote::quote_spanned!($span => $($tt)*)).unwrap_or_else(|e| panic!("{}", e))
25 | | };
26 | | }
| |_^
note: `parse_quote_spanned` could also refer to the macro imported here
--> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pinned_drop.rs:3:50
|
3 | use syn::{spanned::Spanned, visit_mut::VisitMut, *};
| ^
= help: use `self::parse_quote_spanned` to refer to this macro unambiguously
For more information about this error, try `rustc --explain E0659`.
error: could not compile `pin-project-internal` due to 2 previous errors
warning: build failed, waiting for other jobs to finish...
error: failed to verify package tarball
This seems to be caused by this: https://github.com/taiki-e/pin-project/blob/eedd0885afc4bebad728f0adb1acffaaaed0d31a/pin-project-internal/src/pin_project/derive.rs
Which seems to have been changed just days ago: https://github.com/taiki-e/pin-project/commit/ba7d5fd13f81092e64c04a30fd0dd7dd70cb87a8
And was mentioned in this issue: https://github.com/taiki-e/pin-project/issues/337
cargo update did not work.
cargo publish --lockfile dit not work.
I don't know what I can try next.
Removing Cargo.lockl and running publish again worked.
|
gharchive/issue
| 2021-12-28T23:15:57
|
2025-04-01T06:39:13.190726
|
{
"authors": [
"joepio"
],
"repo": "joepio/atomic-data-rust",
"url": "https://github.com/joepio/atomic-data-rust/issues/264",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
181796481
|
Getting this error with my rogue...
Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value)
Ok. Will look into it (will be later this evening as I'm out and about at
present :))
--
Joe Vaughan
On 8 October 2016 at 03:41:10, lkj61 (notifications@github.com) wrote:
Frames [Load]: @(10/07/16 20:39:20) -
...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to
index field '?' (a nil value)
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/joev/SVUI-Temp/issues/80, or mute the thread
https://github.com/notifications/unsubscribe-auth/AATrWRguVwSt-gouFgw6WMK71PE4WqrFks5qxvS2gaJpZM4KRlUg
.
Thanks! Much appreciate all the work you are doing for us fans of SVUI. 😀
On Saturday, October 8, 2016, Joe Vaughan notifications@github.com wrote:
Ok. Will look into it (will be later this evening as I'm out and about at
present :))
--
Joe Vaughan
On 8 October 2016 at 03:41:10, lkj61 (notifications@github.com
<javascript:_e(%7B%7D,'cvml','notifications@github.com');>) wrote:
Frames [Load]: @(10/07/16 20:39:20) -
...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to
index field '?' (a nil value)
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/joev/SVUI-Temp/issues/80, or mute the thread
<https://github.com/notifications/unsubscribe-auth/AATrWRguVwSt-
gouFgw6WMK71PE4WqrFks5qxvS2gaJpZM4KRlUg>
.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/joev/SVUI-Temp/issues/80#issuecomment-252416152, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AJEkVZg5YQOMoY-j_UDeUkY3O-EwXyhGks5qx2q4gaJpZM4KRlUg
.
|
gharchive/issue
| 2016-10-08T01:41:10
|
2025-04-01T06:39:13.202385
|
{
"authors": [
"joev",
"lkj61"
],
"repo": "joev/SVUI-Temp",
"url": "https://github.com/joev/SVUI-Temp/issues/80",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1795017583
|
🛑 dr.veoj.me (my gaming persona) is down
In 777b0d1, dr.veoj.me (my gaming persona) (https://dr.veoj.me) was down:
HTTP code: 0
Response time: 0 ms
Resolved: dr.veoj.me (my gaming persona) is back up in 71fff75.
|
gharchive/issue
| 2023-07-08T17:01:27
|
2025-04-01T06:39:13.205727
|
{
"authors": [
"joevaughan"
],
"repo": "joevaughan/upptime",
"url": "https://github.com/joevaughan/upptime/issues/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
747964988
|
[BUG] No up-to-date light and temperature values in Homebridge anymore
Describe the bug
Temperature and light don't get updated in Homebridge and Home app anymore, although http://<PIR-IP>/api/v1/sensors shows up-to-date temperature and light values.
To Reproduce
Steps to reproduce the behavior:
In the Homebridge GUI navigate to Accessories
Light will have an old value
Open http://<PIR-IP>/api/v1/sensors in browser
Light value will be different, but correct
Expected behavior
The light and temperature values in Homebridge should have the same values as the ones provided by the API.
Plugin environment:
myStrom PIR Firmware Version: 3.8.2
Homebridge Server: Raspberry Pi 4
Plugin Version: 1.8.0
Node.js Version: 14.15.1
Homebridge Version: 1.1.6
Additional context
The issue appears only after some time. Upon re-starting Homebridge it works again, but will stop working after a couple of hours.
The change related to reachability (73298a58f2cfbf3dff2a12e146ab1d137b2aac64) introduced in v1.8.0 seems to break quite a few things, but only after some time running. 🤦 A downgrade to v1.7.1 should work around this while I work on a fix.
I've released homebridge-dingz@1.8.3-nightly.0 (homebridge-dingz@nightly) and am running it now. If you want to help with testing, feel free to install it and report back.
the nightly reverts some of the changes that might have caused some of the problems encountered here and elsewhere 🤕🤒
I've installed 1.8.3 nightly and will report back 👍
In the meantime I've worked on a v2.0.0-nightly of the plug-in. If you reinstall the nightly w/ homebridge-dingz@nightly you should be upgraded to this version.
The functionality remains largely the same but the code is less cluttered and better structured so should generally run better. There might be bugs though but you can always downgrade to v1.8.3-nightly (although I think the newest nightlies are much more robust).
I've had it running over longer periods of time now on my production dingz. So far I could not observe any problems with temperature/brightness values not updating.
(The easiest way to test if it still works after a few hours is to use your phone's flashlight to illuminate or your hand to darken the illumination: the LUX value should change significantly 😄)
It's apparently still not fixed though. 🤦 😞 But I now have a clue where it happened (unreleased mutex locks on the request queue) in the code. :octocat:
https://github.com/johannrichard/homebridge-dingz/blob/dee30b26ab348adb4db62c7f90f47c4ae107b684/src/dingzAccessory.ts#L1585-L1597
So far v2.0.0-nightly is working for me, but it has been running less than 24 hours.
|
gharchive/issue
| 2020-11-21T07:10:21
|
2025-04-01T06:39:13.255104
|
{
"authors": [
"johannrichard",
"qx54"
],
"repo": "johannrichard/homebridge-dingz",
"url": "https://github.com/johannrichard/homebridge-dingz/issues/123",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
84032138
|
Warning in Google Chrome console when generating pdf/excel document
There is a warning in Google Chrome console when generating document: Resource interpreted as Document but transferred with MIME type application/pdf.
Request Headers:
Accept:text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,/;q=0.8
Accept-Encoding:gzip, deflate
Accept-Language:en-US,en;q=0.8,ru;q=0.6
Cache-Control:max-age=0
Connection:keep-alive
Content-Length:305
Content-Type:application/x-www-form-urlencoded
User-Agent:Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/43.0.2357.81 Safari/537.36
Response Headers:
Content-Disposition:attachment; filename=report.pdf
Content-Type:application/pdf;charset=windows-1251
Date:Tue, 02 Jun 2015 13:43:47 GMT
Expires:Thu, 02 Jun 2016 06:50:44 GMT
Server:Apache-Coyote/1.1
Transfer-Encoding:chunked
Yeah I get that too, it still works though so I think its okay IMO
I get this error also for PDF's. But it will not even try to download. I change the file to download to something else (a non-pdf) but get the error and it still downloads.
Also do you know of a way to suppress this error. I would like to not put the download URL out in the open.
This isn't a warning in chrome its logged at info. If you are worried about your download URL being logged that's not good because anyone can view it in the chrome debugger or by monitoring http traffic. Would suggest using single use url/token/cookie if you are concerned about URL re-use. The web is not inherently secure: http://johnculviner.com/dont-trust-the-client/
understandable that there will always be a way to view the url, but I just dont want to hand it to them, would rather make them work just a little. If you have any thoughts as to suppress it that would be great.
But there was still the issue of pdf's not downloading. do you happen to get this in your last push of the script?
On your dont trust the client. Believe me I do not really. I treat the client as if they have some knowledge on web development so I close those holes when I can. But then I am not a security expert and cannot get everything. Though that would be nice.
BTW I am in chrome.
|
gharchive/issue
| 2015-06-02T13:49:44
|
2025-04-01T06:39:13.277595
|
{
"authors": [
"johnculviner",
"mike192020",
"shustinho"
],
"repo": "johnculviner/jquery.fileDownload",
"url": "https://github.com/johnculviner/jquery.fileDownload/issues/98",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
135531609
|
Errors Installing
I have Python2.7.11 for Windows x86/x64 installed on Windows 7 x64. When I run npm install lite-server --save-dev, I get the following errors:
TypeError: Request path contains unescaped characters.
And a few 404 status codes downloading 32-bit node.libs.
anyone else on windows able to test this?
moving to https://github.com/johnpapa/lite-server/issues/54
|
gharchive/issue
| 2016-02-22T20:16:07
|
2025-04-01T06:39:13.283145
|
{
"authors": [
"billfranklin",
"johnpapa"
],
"repo": "johnpapa/lite-server",
"url": "https://github.com/johnpapa/lite-server/issues/39",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
132752012
|
feat(bin): added run time option to config file path and name
Hey, It would be really good to have an option for custom path to configuration
especially if you hate dozen of files in your project root folder (like me)
or have to run different configuration for different needs
Thanks in advance
Regards, Dima
@johnpapa please review
i like the idea. @cgmartin is this how you would like to do it? i prefer yargs myself.
This looks good to me. I like the fact that it matches names with the browsersync --config option.
Yargs is what I'm most comfortable with, but minimist is fine. There are several popular args modules out there that all score pretty high. What we're doing here is pretty simple/straightforward in terms of arg options. Easy to replace if need be.
@johnpapa good to merge? Thanks for the PR @valorkin !
i havent had a chance to test it ... have you Chris? If so, Im cool with it as long as it ups the version number by a minor.
Tested it, works fine. Merging and bumping minor version.
Since it changes the API (not breaking), Shouldn't this be a minor bump to 2.1.0 ? (not a patch)
i bumped it to 2.1.0 and published
yes, minor, I fubar. Thanks for fix
no worries. thanks for testing
|
gharchive/pull-request
| 2016-02-10T16:38:38
|
2025-04-01T06:39:13.287994
|
{
"authors": [
"cgmartin",
"johnpapa",
"valorkin"
],
"repo": "johnpapa/lite-server",
"url": "https://github.com/johnpapa/lite-server/pull/32",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1190736386
|
Disable automatic quotation marks for attrs
I would like to an option in settings to disable automatic quotation marks - we use ' but it keeps putting " and it just grinds our gears :)
Duplicate of #840 and #917.
Try it with HTML: Auto Create Quotes disabled.
|
gharchive/issue
| 2022-04-02T20:50:42
|
2025-04-01T06:39:13.310288
|
{
"authors": [
"JTinkers",
"yaegassy"
],
"repo": "johnsoncodehk/volar",
"url": "https://github.com/johnsoncodehk/volar/issues/1139",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1355008263
|
Type narrowing not working in templates with union types
v-if is not narrowing union types.
<template>
<span v-if="user.id">({{ user.id }})</span>
</template>
<script setup lang="ts">
type User = {
id: number
email: string
}
let user: string | User
</script>
Property 'id' does not exist on type 'string | User'
See demo on stackblitz here.
That's just how TS works here, here's the same code on TS playground, raising the same errors:
https://www.typescriptlang.org/play?#code/C4TwDgpgBAqgzhATlAvFA3gKClAlgEwC4oA7AVwFsAjJbKCCgQ1wBti5hFcSBzTAX0yYWEYFDIJE7Ttx5QAPrEmoMeIlADkBDVEGZcAMygAKCUgB0BAJQY6AYwD2JOA5HmWDnqcmX8VgUA
This, however, would work:
type User = {
id: number
email: string
}
let user: string | User = { id: 'id' }
if (isObj(user)) {
console.log(user.id)
}
function isObj(obj: any): obj is Record<string, any> {
return typeof obj === 'object'
}
https://www.typescriptlang.org/play?#code/C4TwDgpgBAqgzhATlAvFA3gKClAlgEwC4oA7AVwFsAjJbKCCgQ1wBti5hFcSBzTAX0yYWEYFDIJE7Ttx5QAPrEmoMeIlADkBDVEGZcAMygAKXHADyVAFbGJSAJT2MdAMYB7EnDciAdCzc8tpI+BPYCQgZkJC7AuB54FtbGbtbEjCQg9sQpVglQAEoQ7oj4ADwcXLwANFDpIAB8zjiIomSIJFCgkG5GOagoaBo5RcAaAkA
Try typeof user === 'object' or 'id' in user.
|
gharchive/issue
| 2022-08-30T00:19:10
|
2025-04-01T06:39:13.314089
|
{
"authors": [
"LinusBorg",
"austin-agronick",
"johnsoncodehk"
],
"repo": "johnsoncodehk/volar",
"url": "https://github.com/johnsoncodehk/volar/issues/1788",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1081165871
|
Absolute import
i've seen that there has been an issue marked as resolved but it might have broke again because it doesn't seems to work for me.
when volar auto-import a component, it's always using the relative path
my config :
tsconfig.json
{
"compilerOptions": {
"baseUrl": ".",
"esModuleInterop": true,
"experimentalDecorators": true,
"jsx": "preserve",
"lib": ["esnext", "dom"],
"module": "esnext",
"moduleResolution": "node",
"paths": {
"@/*": ["src/*"],
"~/*": ["src/components/*"]
},
"resolveJsonModule": true,
"sourceMap": true,
"strict": true,
"target": "esnext",
"useDefineForClassFields": false,
"types": ["quasar"]
},
"exclude": ["node_modules"],
"include": ["src/**/*.ts", "src/**/*.d.ts", "src/**/*.tsx", "src/**/*.vue"]
}
vite.config.ts
import { quasar, transformAssetUrls } from '@quasar/vite-plugin';
import vue from '@vitejs/plugin-vue';
import { resolve } from 'path';
import { defineConfig } from 'vite';
// https://vitejs.dev/config/
export default defineConfig({
server: {
fs: {
allow: ['.'],
},
},
resolve: {
alias: [
{
find: '@',
replacement: resolve(__dirname, './src'),
},
{
find: '~',
replacement: resolve(__dirname, './src/components'),
},
],
},
plugins: [
vue({
template: { transformAssetUrls },
}),
quasar({
autoImportComponentCase: 'pascal',
sassVariables: 'src/assets/scss/variable.scss',
}),
],
});
Versions :
"typescript": "^4.3.2",
"vite": "^2.5.4",
"vue-tsc": "^0.29.0",
"Volar extension": "v0.29.8"
You can config "typescript.preferences.importModuleSpecifier": "non-relative" in vscode.
that already what i did, both in user and workspace settings, and it's still not importing using non-relative path, with or without the built-in typescript extension enabled
I have the same problem, also type hinting is not working in VS Code when using an absolute import.
|
gharchive/issue
| 2021-12-15T15:19:11
|
2025-04-01T06:39:13.317865
|
{
"authors": [
"Zenthae",
"johnsoncodehk",
"wendyprogramic"
],
"repo": "johnsoncodehk/volar",
"url": "https://github.com/johnsoncodehk/volar/issues/793",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1112931625
|
Update balance UI
Updates the balance text a bit to more closely resemble the screens in Figma and the Bitcoin Design guidelines on units and symbols and on protecting a wallet.
Basically, I turned on some font features according to the Figma designs and added a better symbol for when the balance is hidden.
It's not perfect yet, I also want to add:
The possibility to cycle through states (BTC, sats, hidden) by clicking on a balance. For that we need some global state and I didn't want to touch that quite yet.
Formatting based on the user’s locale. Currently I hardcoded the BTC symbol to be prefixed to the number (as it is done in the U.S. with USD amounts. However, I didn't know what to do with the sats symbol as it looked strange when prefixed to a number so I kept it as a postfix. Open for suggestions on this one!
For a first version it should be fine though. Let me know what you think!
Not sure how we want to organize custom CSS going forward. I used a CSS Module but am open to do it differently or just add everything to index.css for now.
📸
See also https://github.com/joinmarket-webui/joinmarket-webui/issues/19.
Wonderful!
Regarding the sat symbol: so far most interfaces still use "sat" or "sats," i.e. no symbol at all. Thunderhub—which looks quite neat in general I have to say—uses the sat symbol, and simply puts it at the end using a slightly dimmer font color:
Most other interfaces, including bitcoin and lightning wallets, simply use "sat" or "sats" either before or after the number (usually after).
1ML:
LightningNetwork.plus:
Updated to use the sat symbol (resolves #19).
The bitcoin design guidelines don't really use a specific sat symbol so maybe we should let the user pick if they want a symbol or just "sat(s)" in the upcoming settings page. Maybe @GBKS has some input on that? For now, I'd say let's keep it simple and just go with the symbol.
I also added digit groups using Intl.NumberFormat as suggested by the design guides. For now I hardcoded the formatting to the en-US locale but we could do more location specific things in the future.
Regarding browser compatibility of Intl.NumberFormat: It should be available on all somewhat modern browsers. Does someone have any experience with that? And a more broad question: What browser compatibility are we targeting in general?
Great! Let's keep it as is for now. We can do some polishing for v4 later on, I'm sure we can improve things here and there. Different font color for the symbol would also be easier on the eyes, for example.
|
gharchive/pull-request
| 2022-01-24T17:22:05
|
2025-04-01T06:39:13.346183
|
{
"authors": [
"dergigi",
"dnlggr"
],
"repo": "joinmarket-webui/joinmarket-webui",
"url": "https://github.com/joinmarket-webui/joinmarket-webui/pull/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
134876078
|
Add support for nullable property
Other way for #21
WDYT @sagikazarmark ?
How do you know if a multiple type is nullable?
It's done in the ObjectGuesser
I don't see any change in denormalization. You use isset in denormalization, but that won't work with null values.
Right, should be good now
Looks good to me. Although it lacks tests, doesn't it?
Yes will add a better test system following this
Cool.
|
gharchive/pull-request
| 2016-02-19T14:16:48
|
2025-04-01T06:39:13.375714
|
{
"authors": [
"joelwurtz",
"sagikazarmark"
],
"repo": "jolicode/jane",
"url": "https://github.com/jolicode/jane/pull/23",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1464378660
|
[MCU Tray] MKS Skipr
Made an MKS Skipr tray for my personal use, but wanted to share with the community. I hope this can make it into the project.
MCU Tray - MKS Skipr - Heatserts.zip
I made a total hash of merging this PR to the right branch, so I just cherry-picked the commits to 0.9.9-unstable instead.
Let me know how the fit goes. Thanks!
|
gharchive/pull-request
| 2022-11-25T10:59:35
|
2025-04-01T06:39:13.392741
|
{
"authors": [
"Killajoedotcom",
"jon-harper"
],
"repo": "jon-harper/OmniBox",
"url": "https://github.com/jon-harper/OmniBox/pull/71",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
299690929
|
TLS v1.2 for KitKat (and below?)
With GitHub having removed support for TLS v1.0 and v1.1, ForkHub no longer works, it gives an error toast (Loading account & organizations failed).
https://stackoverflow.com/questions/28943660/how-to-enable-tls-1-2-support-in-an-android-application-running-on-android-4-1 shows how to enable TLS v1.2 support for pre-Lollipop. There must be something working, since I can access GitHub on both Chrome Mobile and Lightning Browser.
In the meantime, if anyone can direct me to how to get the "News" look for an arbitrary repo via web browser, I'd appreciate it. The closest I can find is Insights, but that doesn't list individual comments.
square/okhttp#2372 may be instructive here.
Noting also, this affects the version of Forkhub found in the Blackberry World App store, which is an Android app, I think maybe v1.2.0 (tagged release of apk), and the Android runtime provided in BB10 is 4.x era. It was a sad day when I opened the app and found that I couldn't get in. I will give a compliment, and say this is THE best performing Android app in terms of speed that I've found on the Blackberry BB10 platform.
I'm willing to dig in here and help find a work around, maybe submit a pull request, since I don't have the option to upgrade my Android runtime, and find Forkhub to be essential to my daily work and my ability to be away from my desk but deal with the number of projects and companies I work with that have their stuff on GH.
|
gharchive/issue
| 2018-02-23T12:07:02
|
2025-04-01T06:39:13.397455
|
{
"authors": [
"Matrixcoffee",
"goldfndr",
"truedat101"
],
"repo": "jonan/ForkHub",
"url": "https://github.com/jonan/ForkHub/issues/345",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1297756534
|
same scoreboard for all players....
sry but a new problem.
I have made a scoreboard with placeholders and for me it works.
BUT my friend joined and has MY scoreboard, my name in scoreboard, my ranks, my money....not his own.
sry this is a discord screenshot
Ok, I have released a new version. Could you check out if the issue is still present there?
no the bug isn fixed.
NOW a new player joined the game, he has his own scoreboard but NOW I see his scoreboard
iam xinaru he is Zeus
or must i delete the old config?
okay the bug isn fixed
i have delete the old config and add the lines new.
but the scoreboard dont show the own scoreboard always.
Ok, I believe it has to do with the order the players join. I will have another look at it
Alright, I have updated some logic concerning the buffering of scoreboards. In addition, this version will log if placeholder usage is detected. You should not need to delete the config file. If the issue persists, please check in the server logs if it is says "Detected use of placeholders" or "Detected no use of placeholders"
EasyScoreboard-2.5.3.jar.zip
sry but the problem ist there with the new version too.
08.07 17:27:59 [Server] Server thread/ERROR [Sponge]: The Scheduler tried to run the task de_yottaflops_easyscoreboard-S-3 owned by Plugin{id=de_yottaflops_easyscoreboard, name=EasyScoreboards, version=2.5.3, description=A plugin to easily create scoreboards for lobbys, etc., authors=[YottaFLOPS], source=./mods/EasyScoreboard-2.5.3.jar}, but an error occured.
now the plugin is crashed
Ok, this was probably due to another change I made. I apologize. Try with this version, if you could.
EasyScoreboard-2.5.3.jar.zip
ähm...i use 2.5.3?
you have uplöoad this above? this version i use and the error comes
It should be a new build with the same version number
ah okay i try it
okay tried.
scoreboard there but my fried has MY scoreboard
08.07 19:52:04 [Server] Server thread/INFO [EasyScoreboard: Config]: Detected use of placeholders
I believe that I have finally fixed the problem and in my testing it worked. Let me know if the problem persists for you.
EasyScoreboard-2.5.3-2.jar.zip
This time it looks like it's finally going to work.
Just a friend and I on the server, everyone had their own sb. we are both separately leased and jointed it was still each his own sb.
the only thing i forgot to test is the /esb reload command.
Amazing. Thanks for your help debugging this issue and your patience. I have created a new release here but it should be identical (apart from the version number) to the .jar posted above. Thanks again
|
gharchive/issue
| 2022-07-07T16:45:14
|
2025-04-01T06:39:13.409051
|
{
"authors": [
"XiNaru",
"jonas-peeters"
],
"repo": "jonas-peeters/EasyScoreboards",
"url": "https://github.com/jonas-peeters/EasyScoreboards/issues/47",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
635703781
|
Serialize Unit variant to/form string
Closes #115
Todo
[ ] Add test case
bors r+
|
gharchive/pull-request
| 2020-06-09T19:32:23
|
2025-04-01T06:39:13.410451
|
{
"authors": [
"jonasbb"
],
"repo": "jonasbb/serde_with",
"url": "https://github.com/jonasbb/serde_with/pull/116",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2408403753
|
Navigating to New Product Details Page with Different Product ID Fails to Load New Data
Question
I am developing an e-commerce app using Flutter with GetX for state management. I have a ProductDetailsView that displays product details using a named route and arguments. In the ProductDetailsController, I load product details based on the product ID passed through arguments.
Here is my controller code:
import 'package:e_commerce/api/send_request.dart';
import 'package:e_commerce/app/data/color_collections.dart';
import 'package:e_commerce/app/widgets/theme_mode.dart';
import 'package:get/get.dart';
import 'package:webview_flutter/webview_flutter.dart';
class ProductDetailsController extends GetxController {
WebViewController? webViewController;
var webViewHeight = 100.0.obs;
var productDetailsLoading = true.obs;
var productDetails = <String, dynamic>{}.obs;
var productId = 0.obs;
@override
void onInit() {
super.onInit();
productId.value = Get.arguments['product_id'];
loadPage();
}
@override
void onClose() {
super.onClose();
productId.close();
}
void loadPage() async {
productDetailsLoading.value = true;
var resp = await SendRequest("/Products/Manager/CRUD/GetProductDetails", data: {
"product_id": productId.value,
});
productDetails.value = resp["product_details"];
productDetailsLoading.value = false;
if (webViewController == null) {
webViewController = WebViewController()
..setJavaScriptMode(JavaScriptMode.unrestricted)
..setBackgroundColor(isDarkMode()
? DynamicColorsCollection.backgroundColorDark
: DynamicColorsCollection.backgroundColorLight)
..setNavigationDelegate(
NavigationDelegate(
onProgress: (int progress) {},
onPageStarted: (String url) {},
onPageFinished: (String url) async {
final result = await webViewController?.runJavaScriptReturningResult(
"document.documentElement.scrollHeight.toString();");
String heightStr = result as String;
double height = double.parse(heightStr.replaceAll('"', ''));
updateHeight(height);
},
onHttpError: (HttpResponseError error) {},
onWebResourceError: (WebResourceError error) {},
onNavigationRequest: (NavigationRequest request) {
if (request.url.startsWith('https://www.youtube.com/')) {
return NavigationDecision.navigate;
}
return NavigationDecision.prevent;
},
),
)
..loadRequest(Uri.parse(
"http://some-domain.com/products/description/${resp["product_details"]["product_url"]}"));
} else {
webViewController?.loadRequest(Uri.parse(
"http://some-domain.com/products/description/${resp["product_details"]["product_url"]}"));
}
}
void updateHeight(double height) {
webViewHeight.value = height;
}
}
Here is the start of my view code:
class ProductDetailsView extends GetView<ProductDetailsController> {
const ProductDetailsView({Key? key}) : super(key: key);
@override
Widget build(BuildContext context) {
Get.create<ProductDetailsController>(() => ProductDetailsController());
return Obx(() {
if (controller.productDetailsLoading.value) {
return Material(
color: isDarkMode()
? DynamicColorsCollection.backgroundColorSecondaryDark
: DynamicColorsCollection.backgroundColorSecondaryLight,
child: LoadingShimmer(),
);
}
return CustomScaffold(
backgroundColor: isDarkMode()
? DynamicColorsCollection.backgroundColorDark
: DynamicColorsCollection.backgroundColorLight,
The issue arises when I navigate to a new product details page from the current product details page using a button. The button's code is as follows:
Get.toNamed(
Routes.PRODUCT_DETAILS,
arguments: {
"product_id": id,
},
preventDuplicates: false,
);
When I navigate like this, the new product details page does not load the new product data. Instead, it shows the data of the previous product.
How can I ensure that navigating to a new product details page with a different product ID loads the new product's data correctly?
I tried navigating to the new product details page from the current product details page by using the Get.toNamed method with the new product ID as an argument. I expected the new product details page to load and display the details of the new product corresponding to the new product ID. However, instead of loading the new product's data, the page continued to display the data of the previously viewed product. I expected the ProductDetailsController to fetch and show the details of the new product based on the updated product ID passed in the arguments, but this did not happen.
Because onInit olny called once, Mayby you can use GetBuilder and create new controller instance each the page is called
try add global: false, to the GetBuilder
Thanks, That worked 👍🏻👍🏻👍🏻👍🏻👍🏻🔥🔥🔥🔥🔥🔥🔥🔥
|
gharchive/issue
| 2024-07-15T10:34:30
|
2025-04-01T06:39:13.416900
|
{
"authors": [
"inyong1",
"zenzzenpl"
],
"repo": "jonataslaw/getx",
"url": "https://github.com/jonataslaw/getx/issues/3146",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
328860832
|
Abort search when pressing escape, but accept search/input when escap…
…e+enter has been pressed.
Codecov Report
Merging #629 into master will increase coverage by <.01%.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #629 +/- ##
=========================================
+ Coverage 71.3% 71.3% +<.01%
=========================================
Files 135 135
Lines 12709 12711 +2
=========================================
+ Hits 9062 9064 +2
Misses 3647 3647
Impacted Files
Coverage Δ
prompt_toolkit/key_binding/bindings/vi.py
57.73% <100%> (+0.05%)
:arrow_up:
prompt_toolkit/key_binding/bindings/emacs.py
82.46% <100%> (+0.11%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 677190f...eb90798. Read the comment docs.
|
gharchive/pull-request
| 2018-06-03T20:48:30
|
2025-04-01T06:39:13.436419
|
{
"authors": [
"codecov-io",
"jonathanslenders"
],
"repo": "jonathanslenders/python-prompt-toolkit",
"url": "https://github.com/jonathanslenders/python-prompt-toolkit/pull/629",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
694099912
|
No way to remove handles or update their color position in code
This is a request to add support for removing handles and/or updating their position in code with a color value
Added PR #57
|
gharchive/issue
| 2020-09-05T14:44:34
|
2025-04-01T06:39:13.469255
|
{
"authors": [
"chrisCGL",
"tache"
],
"repo": "joncardasis/ChromaColorPicker",
"url": "https://github.com/joncardasis/ChromaColorPicker/issues/54",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1111565891
|
Added state to the model type and updated the language docs
Closes #189 by added a state item to the model type in the core AaC specification. Also started to fill in information about the AaC language in the docs by describing the model type and the data items within it. Specifically there is not a brief description of the model state and the intended utilization of state in modeling a system.
Just as a side note, I was able to successfully exercise this change through the CLI in a GitPod environment but I got all sorts of errors when I tried to do it locally. I put this here for awareness that it is apparently not difficult to break or misconfigure the local development environment. I've not been able to figure out what's gone wrong yet but I'm sure it is something I've done to pollute my local environment. All in all, this has highlighted the value of strict control of the developer space for a project as it is through our GitPod configuration.
|
gharchive/pull-request
| 2022-01-22T13:37:37
|
2025-04-01T06:39:13.470798
|
{
"authors": [
"jondavid-black"
],
"repo": "jondavid-black/AaC",
"url": "https://github.com/jondavid-black/AaC/pull/190",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2466973465
|
prerelease 0.4
This will be a prerelease to help troubleshoot any GUI installation issues. The plan will be to make the full release right before the workshop at Biomag.
I guess the linkcheck is going to fail for
https://jonescompneurolab.github.io/hnn-core/stable/governance.html
Since this is a new file that doesn't exist on the stable docs yet...
What about including #860 ?
@kmilo9999 I'll plan to review it so it can be included in the release before the biomag conference, but for right now I want to get this pip installable so we can test the current changes.
This isn't on the release steps doc but I think @jasmainak mentioned we need to also add to doc/_templates/navbar.html. This is for the dropdown that lets you switch between docs by release version.
note to self, we need to update the picture on the front page!
Woohoo! v0.4! Thanks to all the contributors!
@gtdang there's a slight issue with #865
The image shows up in the readme but not in the built docs, do you mind taking a look?
If it calls for a different PR we can just rename this one...
Are you pushing to pypi also? I see that it shows 0.3 ...
I was going to until I caught the docs error,...
It seemed like the image work with a local make html build but when I viewed it through circleCI the image was missing
|
gharchive/pull-request
| 2024-08-14T22:47:36
|
2025-04-01T06:39:13.475384
|
{
"authors": [
"gtdang",
"jasmainak",
"kmilo9999",
"ntolley"
],
"repo": "jonescompneurolab/hnn-core",
"url": "https://github.com/jonescompneurolab/hnn-core/pull/863",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2598973776
|
🛑 HR - Croatia is down
In ad2ae6d, HR - Croatia (https://www.skyshowtime.com/hr/help/) was down:
HTTP code: 404
Response time: 774 ms
Resolved: HR - Croatia is back up in 2acbcd2 after 10 minutes.
|
gharchive/issue
| 2024-10-19T09:53:31
|
2025-04-01T06:39:13.478034
|
{
"authors": [
"jonesyriffic"
],
"repo": "jonesyriffic/gsp-sst",
"url": "https://github.com/jonesyriffic/gsp-sst/issues/896",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2756972756
|
이정안님 개인 프로젝트 리뷰 1, offline 9 책전달+저녁식사, 2025-01-06
용산역 근처, 오후 6시 30분
가족 여행 일정이 길어져서 그 다음주인 13일로 변경 요청
|
gharchive/issue
| 2024-12-24T01:15:12
|
2025-04-01T06:39:13.478891
|
{
"authors": [
"jongfeel"
],
"repo": "jongfeel/Activity",
"url": "https://github.com/jongfeel/Activity/issues/500",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
222599420
|
How I can show the side menu without segue?
New Issue Checklist
I have read the guidelines for contributing and I understand:
[x] My issue is happening in the latest version of SideMenu.
[x] My issue was not solved in the README.
[x] My issue can not be answered on stackoverflow.com.
[x] My issue is not a request for new functionality that I am unwilling to build and contribute with a pull request.
[ ] My issue is reproducible in the demo project.
Issue Description
I do something like this
let VC1: SideMenuTableViewController = SideMenuTableViewController() let menuLeftNavigationController = UISideMenuNavigationController(rootViewController: VC1) SideMenuManager.menuLeftNavigationController = menuLeftNavigationController) self.viewController.present(SideMenuManager.menuLeftNavigationController!, animated: true, completion: nil)
but doesn't work, I follow the instructions in the description of side menu but does't work, any idea?
You're not following the README correctly.
|
gharchive/issue
| 2017-04-19T02:35:17
|
2025-04-01T06:39:13.483612
|
{
"authors": [
"MervinFlores",
"jonkykong"
],
"repo": "jonkykong/SideMenu",
"url": "https://github.com/jonkykong/SideMenu/issues/180",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
33507740
|
Events not unbound when re-initialising tree
I'm using the treeview control in a popup dialog representing a gallery selector, the treeview being the basis for the folder listing. Now, upon changing the scope of my dialog, I effectively get a new set of folders, so I need to rebuild the tree. Looking through the source, if I call .treeview({...}) on a control that has previously been initialised, it chains to Tree::_init -> Tree::_destroy().
The problem I am finding, is I rebind the onNodeSelected event when I rebuild the tree, but, the original event function is fired, and then the new event function:
http://jsfiddle.net/ML5WX/1/
It seems the original events are not being removed, so the number of events that get fired per node increases with each rebind.
Good spot @Antaris, thanks @Feoni4 for sharing!
|
gharchive/issue
| 2014-05-14T16:16:34
|
2025-04-01T06:39:13.486346
|
{
"authors": [
"Antaris",
"jonmiles"
],
"repo": "jonmiles/bootstrap-treeview",
"url": "https://github.com/jonmiles/bootstrap-treeview/issues/4",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
90012622
|
Q: Updating via Web Sockets
If I define my data statically in a page and call $('#treeview1').treeview({ data: dataToUse }); such as in your examples, the tree works fine, but when I try to update the tree's data in an "onmessage" function in web sockets by invoking $('#treeview1').treeview({ data: dataToUse }); again with the new data, the tree does not update. Any suggestions?
Sounds like an issue with your implementation, not the treeview.
I would suggest debugging your code, perhaps take a look at the context it's running in and/or verify the data being returned is the correct format.
Check the console for errors, ask a question on stackoverflow.
Yes, my fault. The data I sent to the page was very subtly malformed. Thanks.
hiee i also have same issue...when i use the data being returned (that i printed on console and i copied it) it works fine...but when i return data from any function the tree does not update. Any suggestions?
|
gharchive/issue
| 2015-06-22T05:02:45
|
2025-04-01T06:39:13.488566
|
{
"authors": [
"BioCompDevSvcs",
"jonmiles",
"vivek144"
],
"repo": "jonmiles/bootstrap-treeview",
"url": "https://github.com/jonmiles/bootstrap-treeview/issues/94",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
267206253
|
Swap @if burguers statement into a @mixins
Hello, nice burguers 👍 🍔 !
Usually each project has just one burguer.
And of course we can go into the files and comment the @import, but this means touching files in node_modules ( that problably will rely on a problem if we try a package version update ) 😢 .
Wouldn't be a better approach just have @mixins for each burguer ( and have them listed in the DOC) so i can @include just the one that i want ?
Keeping your files clean 😃.
@HectorLS Hey Hector, thanks!
This was designed in a way that you won’t have to mess with anything in node_modules and also include just the ones you want.
In the Customization section of the README.md there are instructions on how to do so.
In short, you basically take the $hamburger-types map, remove all the ones you don’t want and just leave the ones that you want included. For example, if you only want the squeeze variant, you would put the following in your settings file or one of your .scss files for your project:
$hamburgers-types: (
squeeze
);
When the CSS is compiled, only that type will be included.
Hope that helps.
|
gharchive/issue
| 2017-10-20T15:05:22
|
2025-04-01T06:39:13.500852
|
{
"authors": [
"HectorLS",
"jonsuh"
],
"repo": "jonsuh/hamburgers",
"url": "https://github.com/jonsuh/hamburgers/issues/47",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
186053760
|
fixed relative url to http url
Github markdown treated the url without the http prefix as a relative link. Added the http prefix.
Thanks very much for this, it was fixed on the develop branch a couple of weeks ago (236f082 Documentation Tweaks) and is now in master.
|
gharchive/pull-request
| 2016-10-29T02:12:05
|
2025-04-01T06:39:13.743783
|
{
"authors": [
"Nzen",
"jorabin"
],
"repo": "jorabin/KeePassJava2",
"url": "https://github.com/jorabin/KeePassJava2/pull/8",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
579541254
|
'Command failed' in Electron app
Hello and thanks for a great library!
I´ve got a problem with starting and stopping my local mysql process from my electron app.
( I am on macOS Catalina and mysql.server is set in my $PATH)
I can do it with both child_process spawn and child_process exec like this:
exec("sudo mysql.server start", (error, stdout, stderr) => {
if (error) {
console.log("error: " + error);
console.log("stderr: " + stderr);
}
console.log("stdout: " + stdout);
});
and:
const child = spawn('sudo', ['mysql.server', 'start']);
child.on("exit", code => {
console.log(`Child process exited with code ${code}`);
});
child.stdout.on("data", data => {
console.log(`stdout: ${data}`);
});
child.stderr.on("data", data => {
console.log(`stderr: ${data}`);
});
They both give me the sudo password prompt in my terminal though, so its not working when building the app.
So I try to do it with sudo-prompt like this:
sudo.exec("mysql.server start", { name: "My App" }, (error, stdout, stderr) => {
if (error) {
console.log("error: " + error);
console.log("stderr: " + stderr);
}
console.log("stdout: " + stdout);
});
I get the prompt dialog like expected, but then the command fails and gives me Error: Command failed: mysql.server start in the terminal.
The strange thing is that when I run apachectl start with sudo-prompt, it works like a charm!
Any thoughts on what could cause this issue?
Thanks in advance!
@DavidForss thanks for opening this.
Did you get this working?
If not, the issue might be this: mysql.server is set in my $PATH) since sudo-prompt needs you to be explicit with environment variables, because of the difficulties of making this work cross-platform.
|
gharchive/issue
| 2020-03-11T20:57:18
|
2025-04-01T06:39:13.747812
|
{
"authors": [
"DavidForss",
"jorangreef"
],
"repo": "jorangreef/sudo-prompt",
"url": "https://github.com/jorangreef/sudo-prompt/issues/120",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1429872919
|
Add SHA commit to page footer
use the drone sha commit to view on the webpage footer
https://docs.drone.io/pipeline/environment/reference/
Good starting point
|
gharchive/issue
| 2022-10-31T13:42:31
|
2025-04-01T06:39:13.787264
|
{
"authors": [
"thamudi"
],
"repo": "jordanopensource/community-web",
"url": "https://github.com/jordanopensource/community-web/issues/80",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1272120514
|
Document keyboard shortcuts
Is your feature request related to a problem? Please describe.
Just lost my position in a lecture recording as a result of using the cmd+1 keyboard shortcut (i'm on a mac, so i guess the linux/windows equivalents would be different).
The shortcuts cmd+n (with $n \in \{1, 2, ..., 8\}$) in Chrome will usually switch to the n-th tab from the left, but it seems like the video player overrides these commands to jump to different sections of a video. However, this is not documented anywhere.
Describe the solution you'd like
Document the keyboard commands.
Many web-based applications (such as e.g. Gmail or GitHub) are using the shift+? keyboard shortcut to present an overlay/menu listing all possible keyboard shortcuts, so this might be a good thing to do as well.
Documenting the commands would:
make it clearer what they do exactly (i'm still not sure what exactly the overwritten cmd+n commands do. do they always jump in 14-minute intervals, or do they jump to 1/10-of-the-whole-video timestamps?)
make the commands more discoverable. (after realising that the cmd+n commands exist i continued to press some random buttons, and it seems like j, k, l, m, z, and x also are mapped to certain actions, but then again it is impossible to easily know what they do exactly, without delving into the source code of the website)
Describe alternatives you've considered
Just not do anything and leave it as-is
Add a section listing all available keyboard shortcuts somewhere else (e.g. in the settings page)
Have some kind of questionmark-button at the bottom right of the video player, next to the camera mode and feedback buttons
Yeah, good point with both documentation and the player stealing the alt+n/cmd+n shortcuts!
About the documentation: basically, these are 1:1 YouTube's shortcuts where those made sense in TUM-Live (see #476 and YouTube docs). I think I could add something to document them, probably building on #489. I wasn't aware how many sites had keyboard shortcuts until I tried shift+? on them :)
The issue with stolen shortcuts should be easy to fix by checking that the alt modifier is not pressed. Probably checking the control key makes sense too, but not shift, otherwise shortcuts that require shift won't work anymore :( I'm not aware how to solve this issue in full generality so that shortcuts are never stolen.
@just-max Is this issue solved by #587?
If not why did you not mark this as a resolving PR?
To anyone reading this issue: the bug of stealing shortcuts is resolved in #587 but the enhancement of documenting shortcuts is not. Therefore the issue remains open for now.
|
gharchive/issue
| 2022-06-15T11:53:16
|
2025-04-01T06:39:13.826184
|
{
"authors": [
"CommanderStorm",
"just-max",
"lukaskollmer"
],
"repo": "joschahenningsen/TUM-Live",
"url": "https://github.com/joschahenningsen/TUM-Live/issues/530",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1394752001
|
mediapipe fails to install on Apple Silicon
When I do:
python -m pip install .
I get:
Processing /Users/chanana/code/investigations/webcam2
Preparing metadata (setup.py) ... done
Collecting numpy~=1.20
Using cached numpy-1.23.3-cp39-cp39-macosx_11_0_arm64.whl (13.4 MB)
Collecting opencv-contrib-python~=4.5.4
Using cached opencv_contrib_python-4.5.5.64-cp37-abi3-macosx_11_0_arm64.whl (38.7 MB)
ERROR: Could not find a version that satisfies the requirement mediapipe~=0.8 (from matrix-webcam) (from versions: none)
ERROR: No matching distribution found for mediapipe~=0.8
My machine is an M1 MacBookPro with macOS 12.6
I tried the following and it worked for me:
Change line 36 of setup.py to:
'mediapipe~=0.8; platform_system != "Darwin" and platform.machine != "arm64"',
It's directly from this issue.
After this, doing python -m pip install . as given in instructions works!
actually that was fixed this morning in version 0.4.1 :) Try pip install matrix-webcam --upgrade
just guessing here, but this would break on a let's say a Raspberry/Linux arm64 machine
check v0.4.2
I'm getting this crash after running matrix-webcam v0.4.2 on M1 Pro:
objc[25534]: Class CaptureDelegate is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee490) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_videoio.3.4.16.dylib (0x10901c860). One of the two will be used. Which one is undefined.
objc[25534]: Class CVWindow is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee4e0) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_highgui.3.4.16.dylib (0x108048a68). One of the two will be used. Which one is undefined.
objc[25534]: Class CVView is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee508) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_highgui.3.4.16.dylib (0x108048a90). One of the two will be used. Which one is undefined.
objc[25534]: Class CVSlider is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee530) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_highgui.3.4.16.dylib (0x108048ab8). One of the two will be used. Which one is undefined.
Add a python dependency on "@com_google_protobuf//:protobuf_python"
Traceback (most recent call last):
File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/descriptor.py", line 47, in <module>
from google.protobuf.pyext import _message
ImportError: dlopen(/opt/homebrew/lib/python3.10/site-packages/google/protobuf/pyext/_message.cpython-310-darwin.so, 0x0002): symbol not found in flat namespace (__ZN6google8protobuf15FieldDescriptor12TypeOnceInitEPKS1_)
Traceback (most recent call last):
File "/opt/homebrew/bin/matrix-webcam", line 5, in <module>
from matrix_webcam.__main__ import main
File "/opt/homebrew/lib/python3.10/site-packages/matrix_webcam/__main__.py", line 13, in <module>
from mediapipe.python.solutions.selfie_segmentation import SelfieSegmentation
File "/opt/homebrew/lib/python3.10/site-packages/mediapipe/__init__.py", line 16, in <module>
from mediapipe.python import *
File "/opt/homebrew/lib/python3.10/site-packages/mediapipe/python/__init__.py", line 28, in <module>
import mediapipe.python.packet_getter
File "/opt/homebrew/lib/python3.10/site-packages/mediapipe/python/packet_getter.py", line 20, in <module>
from google.protobuf import symbol_database
File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/symbol_database.py", line 62, in <module>
from google.protobuf import descriptor_pool
File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/descriptor_pool.py", line 63, in <module>
from google.protobuf import descriptor
File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/descriptor.py", line 47, in <module>
from google.protobuf.pyext import _message
ImportError: dlopen(/opt/homebrew/lib/python3.10/site-packages/google/protobuf/pyext/_message.cpython-310-darwin.so, 0x0002): symbol not found in flat namespace (__ZN6google8protobuf15FieldDescriptor12TypeOnceInitEPKS1_)
Could you help me with it?
What were your steps to set it up? It seems like you're using python3.10, perhaps you could try 3.9.13 instead and see if it still fails? I think using an older version of the protobuf library (by using py3.9) might help but that's just a guess
@chanana Thanks! python 3.9.13 helps me.
|
gharchive/issue
| 2022-10-03T13:37:58
|
2025-04-01T06:39:13.833156
|
{
"authors": [
"anikitin-intermedia",
"chanana",
"joschuck"
],
"repo": "joschuck/matrix-webcam",
"url": "https://github.com/joschuck/matrix-webcam/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
65886422
|
Bower support
Please add this cool jQuery plugin to the Bower.
My Fork is available as package jquery-colpick. All versions 2.* are identical to those here.
@mrgrain Thank you very much!
|
gharchive/issue
| 2015-04-02T09:19:08
|
2025-04-01T06:39:13.837617
|
{
"authors": [
"mrgrain",
"radekdostal"
],
"repo": "josedvq/colpick-jQuery-Color-Picker",
"url": "https://github.com/josedvq/colpick-jQuery-Color-Picker/issues/46",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.