id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
179613632
Triangle subdivide Also contains several fixes in the kdtree. Therefore lets just merge it onto master and continue from there with radiosity and subdivision. Coverage decreased (-1.8%) to 46.033% when pulling 336b657c9688c05d9955cf9858efea78b0012551 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage increased (+1.5%) to 49.365% when pulling 3c692fc1febc132889896250b508f498c82fbc50 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.8%) to 46.033% when pulling 21452b4ed6144a3a83c14069051445b5dc3fb054 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.05%) to 46.779% when pulling f8ed7d1cda90710966a28282d8b55675233e7bb7 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.8%) to 46.033% when pulling 71915703b578fd40ccbb0a29587aac297e58c4ff on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Keine Ahnung wie man die nicht abgedeckten Zeilen mit Test zu covern sind. KDTree::build kannst du ja mit einer kleinen Box aufrufen https://coveralls.io/builds/8119862/source?filename=lib%2Fkdtree.cpp#L101 So coverst du box.surface_area() == 0. Das hier würde ich eine Hilfsmethode packen, da die Methode schon sehr lang ist. Coverage decreased (-1.7%) to 46.082% when pulling df22071f2f522f1af3433c326cc8590b0a5b6ff6 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-1.1%) to 46.708% when pulling e0ff81e96e487819891ff520c4c524cd7b741839 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Coverage decreased (-0.7%) to 47.126% when pulling 01d71371778cdff9edf123721bb478ffa13aaa09 on triangle-subdivide into 295d6d1adba77f90dd8a7a9d6104c6234f76fee3 on master. Tests done. Coverage decreased weil wir radiohit erweitert haben und da nichts testen. Awesome work!
gharchive/pull-request
2016-09-27T21:05:24
2025-04-01T06:39:12.190102
{ "authors": [ "boxdot", "coveralls", "jeschkies" ], "repo": "jeschkies/renderer", "url": "https://github.com/jeschkies/renderer/pull/60", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2486632383
Project group and some optimization for faster retrieval of workitems. I do this PR to add my personal use case (project group) as it can be useful. It's only here as a workitem retriever for now but can be largely modified. It's implemented so it's not more complicated to use than a project. I believe that my code is clean, but any review is very welcomed as I'm not an expert. Looks good to me, never used it but seems like a nice addition. Also thanks for the cleanup. I'll merge it if you could resolve the merge conflict please
gharchive/pull-request
2024-08-26T11:29:41
2025-04-01T06:39:12.194048
{ "authors": [ "WH-Yoshi", "jesper-raemaekers" ], "repo": "jesper-raemaekers/python-polarion", "url": "https://github.com/jesper-raemaekers/python-polarion/pull/181", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
101102747
Allow a notation that's just "This belongs to a certain class". Partial types, I suppose. So the original idea here was to take generics, and impose a constraint on them. It wouldn't be a complete constraint, only something that would say "Anything of this particular kind of a class". It was something like A > function specification for generics. There are two problems with this approach: How Lily distinguish between someone that wants a partial type constraint, and someone who goofed. I'd like to avoid a symbol soup here, so A > ~function doesn't set well. How does one create a list of partially typed values, since the constraint is against a generic? Uhh...You kinda don't. Because you can't refer to generics outside of a generic function. So, on to the new idea. Let's say you're trying to make a list of those partially a function things. It's gonna look like this: var v: list[~function] = [list::select, integer::to_s, string::split] So now you've got a list of things that are a function. That's what the tilde is for: It says "these things are all just a function. I don't know any more about them." Since there's an annotation for ~function supplied, Lily will create a list of kinda-functions. Otherwise, it will retain the same behavior. The reason for this, is that I don't like that sort of behavior (unifying in a way that reduces away all subtype information) without explicit intent. Lily understands variance, and will also allow functions to specify that they want a partial function. Also, with that in mind, Lily should have no problem determining that any particular kind of a function with arguments or a return type of any sort is a function. Lily having a better understanding of variance was also a motivation for this change. Initially, only functions will be able to be specified as being partial (no subtype information). However, in the future, I'll extend this if it turns out there's a use for it on other types. I feel like an idiot when reading this back to myself now. What I'm trying to solve can easily be solved by having a callable type which would act as a parent type to what function is. There would be no need for any of the above to be done, as Lily understands when doing variance to follow the parent's generic count. So just adding a callable type and maybe refining a couple areas is all that's necessary. I'll get around to it when I add the doc thing specified by #127. No need for a whole bug and a new syntax. Besides, I can only see adding what this asks for as being a problem later (Valid questions like "Why not partial lists, partial hashes, partial higher kinds begin showing up). So there's no need for a special bug for this, I think.
gharchive/issue
2015-08-14T21:01:25
2025-04-01T06:39:12.205623
{ "authors": [ "jesserayadkins" ], "repo": "jesserayadkins/lily", "url": "https://github.com/jesserayadkins/lily/issues/126", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
175508657
How to set Image and text in same Bubble cell New issue checklist [x] I have read all of the README, documentation, and FAQ. [x] I have reviewed the contributing guidelines. Confirmation: ____ [x] I have searched existing issues and this is not a duplicate. General information Library version(s): iOS version(s): Devices/Simulators affected: Reproducible in the demo project? (Yes/No): Related issues: Bug report Expected behavior Actual behavior Steps to reproduce Crash log? Screenshots? Videos? Sample project? Thanks! :smile: Please provide all the information in the issue template so we can help! And don't forget to read our contribution guidelines. :blush:
gharchive/issue
2016-09-07T13:54:24
2025-04-01T06:39:12.211502
{ "authors": [ "Komal2905", "jessesquires" ], "repo": "jessesquires/JSQMessagesViewController", "url": "https://github.com/jessesquires/JSQMessagesViewController/issues/1820", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2101635068
🛑 Boucherie La Lienne is down In 48f4091, Boucherie La Lienne (https://boucherielalienne.ch) was down: HTTP code: 0 Response time: 0 ms Resolved: Boucherie La Lienne is back up in dc653b3 after 5 minutes.
gharchive/issue
2024-01-26T05:53:24
2025-04-01T06:39:12.216821
{ "authors": [ "jessica-habaneros" ], "repo": "jessica-habaneros/uptime", "url": "https://github.com/jessica-habaneros/uptime/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1868560569
[Bug]: node18 memory consumption growth up to25% Version 29.6.4 Steps to reproduce This happens on a NX powered private repo - on demand I can produce a reproduction repo if needed Expected behavior with node16 our test suite takes about 2GB memory in run in band mode Actual behavior with node18 our test suite takes about 2.6GB memory and has hard time to finish on CircleCI Additional context During memory leak hunt I can see that string allocation loads the same modules several times leading to significant growth, shouldn't this in --runInBand mode never happen with the given cache/resolver? Environment System: OS: macOS 13.3.1 CPU: (8) arm64 Apple M1 Pro Binaries: Node: 18.17.0 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/node Yarn: 3.6.2 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/yarn npm: 9.6.7 - ~/Library/Caches/fnm_multishells/63202_1692873754942/bin/npm Update: I've tried --no-cache and this had no effect on loaded modules Probably related to https://github.com/jestjs/jest/issues/11956 Yeah, dupe of that one
gharchive/issue
2023-08-27T16:53:05
2025-04-01T06:39:12.220662
{ "authors": [ "Delapouite", "SimenB", "vire" ], "repo": "jestjs/jest", "url": "https://github.com/jestjs/jest/issues/14454", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
341296543
failing beforeAll() causes even passing tests in the scope to fail 🐛 Bug Report I'm looking for the best practice for aborting a describe block if the test setup fails. For example, when testing an API, if authenticating fails, it's pointless to run any other tests. As suggested in the issue template, I searched StackOverflow first, where I found this incomplete answer, advising to place the initialization code in a beforeAll block. Hence, raising the issue here because I haven't seen anything documenting the practice of returning failure from beforeAll If a beforeAll block fails, tests in that describe block are still run, and they fail, even if otherwise they would pass! To Reproduce describe('test that a 3rd party API remains consistent', () => { beforeAll(() => expect('login').toBe('successful')); // this will fail test('API function 1', () => expect(1).toBe(1)); // each... test('API function 2', () => expect(2).toBe(2)); // ...of these... test('API function 3', () => expect(3).toBe(3)); // ...will fail too }); Expected behavior Jest should report that the beforeAll() failed, and bail the describe scope without executing further tests in it. If it's somehow intended behavior that all tests should still be executed, and still marked as failed (which I find odd), this should be documented under beforeAll and in the Setup and Teardown guide. Other tests in the file, outside of the failing block, should still be executed. Link to repl or repo (highly encouraged) https://repl.it/@DanDascalescu/beforeAll-failure-should-bail-the-test Run npx envinfo --preset jest System: OS: Linux 4.15 Ubuntu 16.04.4 LTS (Xenial Xerus) CPU: x64 Intel(R) Core(TM) i7-7500U CPU @ 2.70GHz Binaries: Node: 10.5.0 - /usr/local/bin/node Yarn: 1.7.0 - /usr/bin/yarn npm: 6.2.0 - /usr/local/bin/npm npmPackages: jest: ^23.4.1 => 23.4.1 See also #6527 @aaronabramov have we changed the behaviour here for circus? Bailing after a failing lifecycle hook makes sense to me yes! this was a bug in jasmine (or a feature). We changed this behavior in jest-circus. @dandv jest-circus isn't shipped with jest by default yet, but you can test it by yarn add jest-circus and adding "testRunner": "jest-circus/runner" to your jest config I would like to add to this a request for named hooks. The test is marked as a failure and not run, but named hooks (especially when the hooks reside outside of the test file) would aid in debugging. @alycda feel free to open up a separate issue for that 🙂 Closing this as the behavior is fixed in jest-circus Closing this as the behavior is fixed in jest-circus It doesn't seem to be fixed in latest Jest (v23.6.0). So it's fixed for what users exactly? Facebook internal? See the comment above: https://github.com/facebook/jest/issues/6695#issuecomment-405326998 Install circus (which will become the default in Jest at some point) and you should be good. Also see #7198 Install circus (which will become the default in Jest at some point) Thanks I'll follow. Still why then it's not set as default? Are there any other issues we should be aware of? Is jest-circus considered stable? You can follow #6295 for state of making it the default. jest-circus seems to not fix this. All 3 tests still run and fail. Here's the command I ran: $ yarn run jest --testRunner=jest-circus/runner --bail Am I missing a configuration value somewhere? This is still as broken as ~1yr ago, even with circus... Please consider reopening. To my eyes this just looks like an issue with the reporting. beforeAll is only run once and none of the other tests actually execute, but they all fail with the same reason, and they all print the same error. We should probably avoid printing the individual tests if beforeAll fails. Right now the implementation just checks if we've had an error and if yes, print it and fail the test I've seen a similar issue with reporting. It's hard to disentangle from proprietary code, but what happens is that a test.todo('description', callback); reports the MongoDB error thrown by the previously executed test (a bulk write failure) instead of the correct error that only a description should be specified for .todo tests. I tried throwing an error directly in that failing test but that doesn't reproduce the behavior. Happy to see if the .todo issue reproduces once this is fixed. @SimenB the issue seems to be only partially fixed in circus. since nested hooks are still being executed. To my eyes this just looks like an issue with the reporting. beforeAll is only run once and none of the other tests actually execute, but they all fail with the same reason, and they all print the same error. We should probably avoid printing the individual tests if beforeAll fails. Right now the implementation just checks if we've had an error and if yes, print it and fail the test without actually executing it @SimenB Should we create a separate issue for this? Or was the intention to re-open this to continue the work in this one? Sure, a new issue about the state today and what we want the end state to be makes sense to me Is it a duplicate of #577? @pkuczynski I think so, yes. Still not working as expected... Not via throwing an error or calling done.fail(err) Definitely not stale, still see this issue in latest jest. Undigging topic - imo. test suite should interrupt tests execution with appropriate message when one of hooks will fail or timeout (beforeEach/All, afterEach/All). Unstale As a present for this issue's 5th birthday, I would like to ask a maintainer to reopen #2713, because throwing from beforeAll should really stop executing any further tests. @SimenB? It seems this framework is totally abandonned... RIP!
gharchive/issue
2018-07-15T06:36:23
2025-04-01T06:39:12.238095
{ "authors": [ "JaneJeon", "SimenB", "aaronabramov", "alycda", "chrismwendt", "dandv", "denissabramovs", "fider", "medikoo", "pkuczynski", "svyandun", "thernstig", "vtgn" ], "repo": "jestjs/jest", "url": "https://github.com/jestjs/jest/issues/6695", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2475719495
compile frontend in oldSchool, update task info In #69 there was uncompiled frontend, also task-info.yaml files didn't reflect the new changes. This is fixed @Daoortor could you please fix Qodana?
gharchive/pull-request
2024-08-20T13:53:07
2025-04-01T06:39:12.247011
{ "authors": [ "Daoortor", "nbirillo" ], "repo": "jetbrains-academy/kotlin-onboarding-collections", "url": "https://github.com/jetbrains-academy/kotlin-onboarding-collections/pull/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
256651489
Supporting auto-generation of CA issuer keypair #79 adds a basic CA issuer that reads a signing keypair from a Secret in the Kubernetes API server in order to issue certificates. For convenience, it may be desirable to support an 'automatically generate a signing keypair' mode. ref #85 #84 Instead of adding a 'generate CA' field to the Issuer, I think we should look at adding a selfsigned certificate issuer and adding an IsCA field to the Certificate resource type (as described in #85 #84) Closing in favour of #84
gharchive/issue
2017-09-11T10:15:59
2025-04-01T06:39:12.262339
{ "authors": [ "munnerz" ], "repo": "jetstack-experimental/cert-manager", "url": "https://github.com/jetstack-experimental/cert-manager/issues/83", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
443287523
Cert manager stuck at attempting to acquire leader lease Describe the bug: Have been running with cert manager for a year now, but suddenly, it is being stuck at "attempting to acquire leader lease", with no error messages. It is working fine on a similar cluster, with the same settings. Environment details:: Kubernetes version (e.g. v1.10.2): 1.11.8-gke.6 Cloud-provider/provisioner (e.g. GKE, kops AWS, etc): GKE cert-manager version (e.g. v0.4.0): v0.6.2 Install method (e.g. helm or static manifests): helm /kind bug After reinstall of helm chart, and some patience, then it worked.
gharchive/issue
2019-05-13T09:11:49
2025-04-01T06:39:12.265136
{ "authors": [ "jonasdkhansen" ], "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/issues/1669", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
305999446
Issue certificate using dns01 via route53 stuck on SelfCheck status Is this a BUG REPORT or FEATURE REQUEST?: Uncomment only one, leave it on its own line: /kind bug /kind feature What happened: Trying to configure dns01 route53 provider and it works using staging letsencrypt ClusterIssuer. When changing ClusterIssuer to live, it stuck on status Normal PrepareCertificate 15m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 15m cert-manager-controller Presenting dns-01 challenge for domain auth-service-trunk.gel.net Normal PresentChallenge 15m cert-manager-controller Presenting dns-01 challenge for domain auth.test.gel.tech Normal SelfCheck 14m cert-manager-controller Performing self-check for domain auth-service-trunk.gel.net Normal SelfCheck 14m cert-manager-controller Performing self-check for domain auth.test.geo.tech I've checked route 53 and I see there _acme-challenge. TXT records for both domains. Same live ClusterIssuer works as expected with http01 provider. In log no errors. How to understand what wrong? Can I enable debug mode in some way? What you expected to happen: Successfully issued certificate How to reproduce it (as minimally and precisely as possible): Anything else we need to know?: Environment: Kubernetes version (use kubectl version): Server Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.8", GitCommit:"2f73858c9e6ede659d6828fe5a1862a48034a0fd", GitTreeState:"clean", BuildDate:"2018-02-09T21:23:25Z", GoVersion:"go1.8.3", Compiler:"gc", Platform:"linux/amd64"} Cloud provider or hardware configuration**: AWS Install tools: Installed via helm from https://github.com/kubernetes/charts/tree/master/stable/cert-manager Others: Hi @argusua is your cluster DNS able to resolve external domain names? cert-manager with be trying to fetch the _acme-challenge' using the DNS provided by the cluster to the cert-manager` pod. If that DNS does not allow Internet DNS look-ups, that will fail. If you have this problem, then from k8s 1.9 you can specify custom DNS servers for each Pod, so you could add an Internet DNS server like Google 8.8.8.8 to the cert-manager pod. https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pods-dns-config I've noticed the similar issue. Everything was working 4 days ago but when I applied a new certificate resource I got the following error. Error preparing issuer for certificate: error waiting for key to be available for domain cert-manager pod logs I0316 14:00:52.147870 1 server.go:68] Listening on http://0.0.0.0:9402 I0316 14:00:52.153768 1 leaderelection.go:174] attempting to acquire leader lease... I0316 14:00:52.230488 1 leaderelection.go:184] successfully acquired lease kube-system/cert-manager-controller I0316 14:03:07.518985 1 controller.go:138] clusterissuers controller: syncing item 'letsencrypt-production' I0316 14:03:10.273963 1 helpers.go:122] Setting lastTransitionTime for ClusterIssuer "letsencrypt-production" condition "Ready" to 2018-03-16 14:03:10.273947464 +0000 UTC m=+138.299805182 I0316 14:03:10.279701 1 controller.go:152] clusterissuers controller: Finished processing work item "letsencrypt-production" I0316 14:03:10.279766 1 controller.go:138] clusterissuers controller: syncing item 'letsencrypt-production' I0316 14:03:10.991498 1 controller.go:152] clusterissuers controller: Finished processing work item "letsencrypt-production" I0316 14:04:19.614244 1 controller.go:187] certificates controller: syncing item 'test/test-certificate' I0316 14:04:19.615050 1 sync.go:107] Error checking existing TLS certificate: secret "test-certificate" not found I0316 14:04:19.615205 1 sync.go:209] Preparing certificate with issuer I0316 14:04:19.616105 1 prepare.go:230] Compare "" with "https://acme-v01.api.letsencrypt.org/acme/reg/31265273" I0316 14:08:14.093642 1 helpers.go:165] Setting lastTransitionTime for Certificate "test-certificate" condition "Ready" to 2018-03-16 14:08:14.093626634 +0000 UTC m=+442.119484352 I0316 14:08:14.093733 1 sync.go:213] Error preparing issuer for certificate: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded E0316 14:08:14.110479 1 sync.go:190] [test/test-certificate] Error getting certificate 'test-certificate': secret "test-certificate" not found Certificate apiVersion: certmanager.k8s.io/v1alpha1 kind: Certificate metadata: name: test-certificate namespace: test spec: secretName: test-certificate issuerRef: name: letsencrypt-production kind: ClusterIssuer dnsNames: - test.mydomain.com acme: config: - dns01: provider: route53 domains: - test.mydomain.com Cluster Issuer apiVersion: certmanager.k8s.io/v1alpha1 kind: ClusterIssuer metadata: name: letsencrypt-production spec: acme: server: https://acme-v01.api.letsencrypt.org/directory email: myemail@mydomain.com privateKeySecretRef: name: letsencrypt-production dns01: providers: - name: route53 route53: accessKeyID: AKIAXXXXXXXXXXXXX secretAccessKeySecretRef: name: route53-config key: secret-access-key region: us-east-1 Certificate detailed description Name: test-certificate Namespace: test Labels: <none> Annotations: <none> API Version: certmanager.k8s.io/v1alpha1 Kind: Certificate Metadata: Cluster Name: Creation Timestamp: 2018-03-16T14:08:31Z Generation: 0 Resource Version: 33688876 Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/test/certificates/test-certificate UID: 9rf2b603-98uj-11e8-8521-1239c7046d88 Spec: Acme: Config: Dns 01: Provider: route53 Domains: test.mydomain.com Common Name: Dns Names: test.mydomain.com Issuer Ref: Kind: ClusterIssuer Name: letsencrypt-production Secret Name: test-certificate Status: Acme: Authorizations: <nil> Conditions: Last Transition Time: 2018-03-16T14:19:33Z Message: Error checking ACME domain validation: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded Reason: ErrCheckAuthorization Status: False Type: Ready Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning ErrorPrepareCertificate 26m (x29 over 5h) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "test.mydomain.com": context deadline exceeded Warning ErrorCheckCertificate 10m (x30 over 5h) cert-manager-controller Error checking existing TLS certificate: secret "test-certificate" not found Normal PrepareCertificate 10m (x30 over 5h) cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 10m (x30 over 5h) cert-manager-controller Presenting dns-01 challenge for domain test.mydomain.com Normal SelfCheck 9m (x30 over 5h) cert-manager-controller Performing self-check for domain test.mydomain.com TXT record was created in Route53 hosted zone and it was publicly accessible cert-manager: v0.2.3 kubernetes: v1.7.9 cloud provider: AWS Hi @whereisaaron Thanks for fast response. Looks like yes, dns resolution works inside cert-manager container: Also, today status changed for certificate to: Status: Acme: Authorizations: <nil> Conditions: Last Transition Time: 2018-03-16T15:12:40Z Message: Error checking ACME domain validation: [error waiting for authorization for domain "auth.test.gel.tech": context canceled, error waiting for authorization for domain "auth-service-trunk.gel.net": context canceled] Reason: ErrCheckAuthorization Status: False Type: Ready Same here: $ kubectl get clusterissuer -oyaml apiVersion: v1 items: - apiVersion: certmanager.k8s.io/v1alpha1 kind: ClusterIssuer metadata: clusterName: "" creationTimestamp: 2018-03-18T11:00:27Z generation: 0 name: letsencrypt-prod namespace: "" resourceVersion: "3499055" selfLink: /apis/certmanager.k8s.io/v1alpha1/letsencrypt-prod uid: 90d32342-2a9b-11e8-ae12-4ccc6ad2499b spec: acme: dns01: providers: - name: aws route53: accessKeyID: AK232AHSKGRLPPABQ hostedZoneID: "" region: eu-central-1 secretAccessKeySecretRef: key: secret-access-key name: route53-credentials-secret email: info@example.com privateKeySecretRef: key: "" name: ingress-key-pair server: https://acme-staging.api.letsencrypt.org/directory status: acme: uri: https://acme-staging.api.letsencrypt.org/acme/reg/232323 conditions: - lastTransitionTime: 2018-03-18T11:00:28Z message: The ACME account was registered with the ACME server reason: ACMEAccountRegistered status: "True" type: Ready kind: List metadata: resourceVersion: "" selfLink: "" kubectl describe certificate Name: rig2-tls Namespace: default Labels: <none> Annotations: <none> API Version: certmanager.k8s.io/v1alpha1 Kind: Certificate Metadata: Cluster Name: Creation Timestamp: 2018-03-18T11:02:58Z Generation: 0 Owner References: API Version: extensions/v1beta1 Block Owner Deletion: true Controller: true Kind: Ingress Name: rig2-zcash-miner UID: 95c1ed1c-2a2a-11e8-ae12-4ccc6ad2499b Resource Version: 3499961 Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/default/certificates/rig2-tls UID: eb1acec6-2a9b-11e8-ae12-4ccc6ad2499b Spec: Acme: Config: Dns 01: Provider: aws Domains: rigs.umine.farm Common Name: Dns Names: rigs.umine.farm Issuer Ref: Kind: ClusterIssuer Name: letsencrypt-prod Secret Name: rig2-tls Status: Acme: Authorizations: <nil> Conditions: Last Transition Time: 2018-03-18T11:06:38Z Message: Error checking ACME domain validation: error waiting for key to be available for domain "rigs.umine.farm": context deadline exceeded Reason: ErrCheckAuthorization Status: False Type: Ready Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal SelfCheck 7m (x5 over 21m) cert-manager-controller Performing self-check for domain rigs.umine.farm Warning ErrorCheckCertificate 4m (x6 over 22m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-tls" not found Normal PrepareCertificate 4m (x6 over 22m) cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 4m (x6 over 22m) cert-manager-controller Presenting dns-01 challenge for domain rigs.umine.farm Warning ErrorPrepareCertificate 39s (x6 over 18m) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "rigs.umine.farm": context deadline exceeded and route53 record is created and working well My ingress: ingress: enabled: true annotations: #{} kubernetes.io/ingress.class: nginx nginx.ingress.kubernetes.io/rewrite-target: '/' kubernetes.io/tls-acme: "true" certmanager.k8s.io/acme-challenge-type: 'dns01' certmanager.k8s.io/acme-dns01-provider: 'aws' # nginx.ingress.kubernetes.io/ssl-redirect: "false" path: /rig2 hosts: - rigs.umine.farm tls: #[] - secretName: rig2-tls hosts: - rigs.umine.farm @whereisaaron thanks for the tip regarding dns. I can confirm where I exec into cert manager and try pining that address it fails. External addresses are ok, but I configured my cluster woith the same name as dns address, so I guess that's why the query doesn't go to upstream server which is 8.8.8.8 Not fully fixed. Got it working for rigs.umine.farm and got fake certificate from staging acme server. Changed server to prod and was hoping to get a new certificate (proper one) but that didn't work. Changed DNS to another name and started getting errors during self-check again. Warning ErrorCheckCertificate 17m (x2 over 20m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-cert" not found Normal PrepareCertificate 17m (x2 over 20m) cert-manager-controller Preparing certificate with issuer Warning ErrorPrepareCertificate 17m cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "monitor.umine.farm": context deadline exceeded Normal PresentChallenge 17m (x2 over 20m) cert-manager-controller Presenting dns-01 challenge for domain monitor.umine.farm Normal SelfCheck 16m (x2 over 19m) cert-manager-controller Performing self-check for domain monitor.umine.farm Warning ErrorCheckCertificate 1m (x5 over 15m) cert-manager-controller Error checking existing TLS certificate: secret "rig2-cert" not found Normal PrepareCertificate 1m (x5 over 15m) cert-manager-controller Preparing certificate with issuer Warning ErrorPrepareCertificate 1m (x4 over 11m) cert-manager-controller Error preparing issuer for certificate: error waiting for key to be available for domain "monitor.umine.farm": context deadline exceeded Normal PresentChallenge 1m (x5 over 15m) cert-manager-controller Presenting dns-01 challenge for domain monitor.umine.farm Normal SelfCheck 1m (x5 over 14m) cert-manager-controller Performing self-check for domain monitor.umine.farm 'context deadline exceeded' means the HTTP request timed out. With the range of stuff you have reported my guess would be you've having some connectivity issues with Internet requests from your cluster. It looks like your cert-manager is fine, but it has, perhaps intermittent, trouble making DNS requests and making HTTP requests to the Internet to check the challenges are in place. Looks like DNS and HTTP request are correct Request performed from the cert-manager container: dig acme-v01.api.letsencrypt.org ; <<>> DiG 9.11.2-P1 <<>> acme-v01.api.letsencrypt.org ;; global options: +cmd ;; Got answer: ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 48485 ;; flags: qr rd ra; QUERY: 1, ANSWER: 3, AUTHORITY: 0, ADDITIONAL: 1 ;; OPT PSEUDOSECTION: ; EDNS: version: 0, flags:; udp: 4096 ;; QUESTION SECTION: ;acme-v01.api.letsencrypt.org. IN A ;; ANSWER SECTION: acme-v01.api.letsencrypt.org. 60 IN CNAME api.letsencrypt.org-ng.edgekey.net. api.letsencrypt.org-ng.edgekey.net. 60 IN CNAME e14990.dscx.akamaiedge.net. e14990.dscx.akamaiedge.net. 20 IN A 104.95.238.173 ;; Query time: 6 msec ;; SERVER: 10.3.0.10#53(10.3.0.10) ;; WHEN: Mon Mar 19 10:15:00 UTC 2018 ;; MSG SIZE rcvd: 158 curl https://acme-v01.api.letsencrypt.org/directory { "key-change": "https://acme-v01.api.letsencrypt.org/acme/key-change", "meta": { "terms-of-service": "https://letsencrypt.org/documents/LE-SA-v1.2-November-15-2017.pdf" }, "new-authz": "https://acme-v01.api.letsencrypt.org/acme/new-authz", "new-cert": "https://acme-v01.api.letsencrypt.org/acme/new-cert", "new-reg": "https://acme-v01.api.letsencrypt.org/acme/new-reg", "rYI1xLOBxqk": "https://community.letsencrypt.org/t/adding-random-entries-to-the-directory/33417", "revoke-cert": "https://acme-v01.api.letsencrypt.org/acme/revoke-cert" } As I said before, nothing has been changed in our k8s cluster or cert-manager configuration and everything was working a week ago. @simonkey007 cert-manager isn't trying to access letsencrypt.org, it is trying to access the domain for the certificate being issued, to check that the challenge record is in place. e.g. 'umine.farm'. Still can't get it work for my live ingress. I've tried today create test ingress and it works as expected. apiVersion: extensions/v1beta1 kind: Ingress metadata: name: lets-test annotations: kubernetes.io/ingress.class: "nginx" kubernetes.io/tls-acme: "true" certmanager.k8s.io/cluster-issuer: "letsencrypt-live" certmanager.k8s.io/acme-challenge-type: "dns01" certmanager.k8s.io/acme-dns01-provider: "route53" spec: tls: - secretName: test-crt-tls hosts: - lets.test.gelato.tech - letstest.gelatogroup.net rules: - host: lets.test.gelato.tech http: paths: - path: / backend: serviceName: jenkins servicePort: 80 - host: letstest.gelatogroup.net http: paths: - path: / backend: serviceName: jenkins servicePort: 80 For live ingress apiVersion: extensions/v1beta1 kind: Ingress metadata: annotations: certmanager.k8s.io/acme-challenge-type: dns01 certmanager.k8s.io/acme-dns01-provider: route53 certmanager.k8s.io/cluster-issuer: letsencrypt-live kubernetes.io/ingress.class: nginx kubernetes.io/tls-acme: "true" creationTimestamp: 2018-03-16T07:19:07Z generation: 269 name: auth-service-kubernetes-live-nginx namespace: default resourceVersion: "119221337" selfLink: /apis/extensions/v1beta1/namespaces/default/ingresses/auth-service-kubernetes-live-nginx uid: 5068e65a-28ea-11e8-94fa-0ae1ba4584f4 spec: rules: - host: auth.live.gelato.tech http: paths: - backend: serviceName: auth-service-kubernetes-live-nginx servicePort: 80 path: / - host: auth-service.gelatogroup.com http: paths: - backend: serviceName: auth-service-kubernetes-live-nginx servicePort: 80 path: / tls: - hosts: - auth.live.gelato.tech - auth-service.gelatogroup.com secretName: auth-service-kubernetes-live-nginx-tls I get next certificate like this: Name: auth-service-kubernetes-live-nginx-tls Namespace: default Labels: <none> Annotations: <none> API Version: certmanager.k8s.io/v1alpha1 Kind: Certificate Metadata: Cluster Name: Creation Timestamp: 2018-03-20T14:23:59Z Deletion Grace Period Seconds: <nil> Deletion Timestamp: <nil> Owner References: API Version: extensions/v1beta1 Block Owner Deletion: true Controller: true Kind: Ingress Name: auth-service-kubernetes-live-nginx UID: 5068e65a-28ea-11e8-94fa-0ae1ba4584f4 Resource Version: 119219236 Self Link: /apis/certmanager.k8s.io/v1alpha1/namespaces/default/certificates/auth-service-kubernetes-live-nginx-tls UID: 54dba49f-2c4a-11e8-94fa-0ae1ba4584f4 Spec: Acme: Config: Dns 01: Provider: route53 Domains: auth.live.gelato.tech auth-service.gelatogroup.com Common Name: Dns Names: auth.live.gelato.tech auth-service.gelatogroup.com Issuer Ref: Kind: ClusterIssuer Name: letsencrypt-live Secret Name: auth-service-kubernetes-live-nginx-tls Status: Conditions: <nil> Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning ErrorCheckCertificate 21m cert-manager-controller Error checking existing TLS certificate: secret "auth-service-kubernetes-live-nginx-tls" not found Normal PrepareCertificate 21m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 21m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 20m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Warning ErrorCheckCertificate 9m cert-manager-controller Error checking existing TLS certificate: secret "auth-service-kubernetes-live-nginx-tls" not found Normal PrepareCertificate 9m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 9m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Normal PresentChallenge 9m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 9m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Additionally, I found today in logs next error: E0320 14:43:49.216953 1 prepare.go:167] Error cleaning up solver: Failed to change Route 53 record set: InvalidChangeBatch: Tried to delete resource record set [name='_acme-challenge.auth-service.gelatogroup.com.', type='TXT'] but it was not found status code: 400, request id: 19c8e4a0-2c4d-11e8-99b7-430c9c1f3614 I've tried create this record manually, but error again apeared. @argusua I don't think you need to create it manually as it provides a response with a long non-readable string when you query it (run test record in route53). And I am just curious - can single secret work with multiple hosts? I had an idea of one secret/certificate per host but maybe it works with several as well (I am not talking about wildcard certificates as your domain names are different.) Get some progress on my issue. I've created new clusterIssuer (previous I get by migrating from kube-lego as described here https://github.com/jetstack/cert-manager/blob/master/docs/user-guides/migrating-from-kube-lego.md). Now I get a certificate but only for first domain: Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal PrepareCertificate 4m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 4m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Normal PresentChallenge 4m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 3m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Normal PrepareCertificate 2m cert-manager-controller Preparing certificate with issuer Normal PresentChallenge 2m cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Normal PresentChallenge 2m cert-manager-controller Presenting dns-01 challenge for domain auth.live.gelato.tech Normal SelfCheck 1m cert-manager-controller Performing self-check for domain auth.live.gelato.tech Normal ObtainAuthorization 49s cert-manager-controller Obtained authorization for domain auth.live.gelato.tech Warning ErrorPrepareCertificate 3s cert-manager-controller Error preparing issuer for certificate: error presenting acme authorization for domain "auth-service.gelatogroup.com": Failed to change Route 53 record set: InvalidChangeBatch: RRSet with DNS name _acme-challenge.auth-service.gelatogroup.com. is not permitted in zone sourcelogistic.com. status code: 400, request id: 3ae889c0-2eb0-11e8-952e-ad6827c36e4f Normal RenewalScheduled 3s cert-manager-controller Certificate scheduled for renewal in 1438 hours Normal PrepareCertificate 2s (x2 over 3s) cert-manager-controller Preparing certificate with issuer Warning ErrorPrepareCertificate 2s cert-manager-controller Error preparing issuer for certificate: error presenting acme authorization for domain "auth-service.gelatogroup.com": Failed to change Route 53 record set: InvalidChangeBatch: RRSet with DNS name _acme-challenge.auth-service.gelatogroup.com. is not permitted in zone sourcelogistic.com. status code: 400, request id: 985b9304-2eb0-11e8-8ec5-9feeb1e51a10 Normal RenewalScheduled 2s cert-manager-controller Certificate scheduled for renewal in 1438 hours Normal PresentChallenge 1s (x2 over 3s) cert-manager-controller Presenting dns-01 challenge for domain auth-service.gelatogroup.com Looks like cert-manager uses incrorect zoneId for set dns record in route53. Finally, I found why the wrong zoneId had used. It's because of CNAME record. My current issue resolved. I had/have the same issue with http01 it seams, where the cert-manager cant curl the well-known url from inside the cluster, but externaly it is possible to access. Just the edge router missing some loopback ... spend 15h for this knowledge. The error description "error waiting for key to be available for domain" is very missleading. Should be "error waiting for collange-key resource to be available for domain ..." @Zetanova often referred to as 'hairpin NAT', needed if your egress and ingress have the same external IP address. It can be confusing to configure. @whereisaaron Yes, i disovered it too. My problem was that i just missundestud the error message. I put active/passive vm iptabales+keepalived in front of the k8s and it seams that i cant get the hairpin NAT to work. I got a similar issue on dns01 config. Log was like follows. (As @whereisaaron says, this should be shown only http01 config...) $ kubectl describe certificate ... Events: ... Warning ErrorPrepareCertificate 2m cert-manager-controller Error preparing issuer for certificate: [error waiting for key to be available for domain "{censored}": context canceled] ... In my case, it was fixed by just deleting (and restarting) the cert-manager pod. I'm going to close this now as it appears to be resolved - thanks everyone for providing support and help debugging 😄 Just tried to generate letsencrypt's wildcard cert via dns01/route53 validation and getting: E0621 14:13:18.291341 1 controller.go:186] certificates controller: Re-queuing item "namespace/xxx-net" due to error processing: Failed to change Route 53 record set: InvalidChangeBatch: Tried to delete resource record set [name='_acme-challenge.xxx.net.', type='TXT'] but it was not found Looks like it should just silently ignore that error or don't delete entries which don't exists in route53? Did you manage to resolve that issue? Using the latest cert manager v0.3.0. same problem with 0.5.0, was working fine in 0.4.1, rolling back we had same issue, it got resolved when I created the TXT record manually, then acme cert manager changed that record because it was expired and deleted successfully. I got the txt record key from cert manager logs and the value from aws cloudtrail logs. Next time I will try creating the txt record with some random value to see if that helps
gharchive/issue
2018-03-16T16:43:48
2025-04-01T06:39:12.295897
{ "authors": [ "Zetanova", "argusua", "cryptcoin-junkey", "grebois", "infa-ddeore", "mlushpenko", "munnerz", "simonkey007", "urbaniak", "whereisaaron" ], "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/issues/398", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1407905390
Prompt user to check supported versions if unknown Before: $ go run main.go operator deploy --version=foo operator version foo does not exist exit status 1 After: $ jsctl operator versions v0.0.1-alpha.17 $ go run main.go operator deploy --version=foo operator version foo is unknown or not supported by this version of jsctl. Run 'jsctl operator versions' to see the supported operator versions. exit status 1 Signed-off-by: Charlie Egan charlieegan3@users.noreply.github.com Thanks Irbe 🙏
gharchive/pull-request
2022-10-13T14:27:24
2025-04-01T06:39:12.300245
{ "authors": [ "charlieegan3" ], "repo": "jetstack/jsctl", "url": "https://github.com/jetstack/jsctl/pull/38", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
306959545
Use new test fixture in Cassandra controls I'm trying to get rid of the Cassandra specific test fixture in ./pkg/controllers/cassandra/testing In this branch I've switched: ServiceAccounts Roles RoleBindings I'll deal with Pilot and NodePool controls in separate branches. Release note: NONE /test e2e /test e2e v1.10 /retest /lgtm /approve
gharchive/pull-request
2018-03-20T17:13:48
2025-04-01T06:39:12.306997
{ "authors": [ "munnerz", "wallrj" ], "repo": "jetstack/navigator", "url": "https://github.com/jetstack/navigator/pull/291", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
376957886
Document how to get the root key after init It would be good to have a few lines in the documentation about how to retrieve the root key after init, to make it possible to login to vault. if you use AWS KMS you can do it with this script: export REGION=us-west-2 export AWS_PROFILE=dev echo "Fetching Vault unseal keys and root token from AWS..." aws ssm get-parameters --names kubernetes-vault-root --region us-west-2 | jq -r '.Parameters[].Value' | base64 -D> /tmp/root-token ROOT_TOKEN=$(aws kms decrypt --region $REGION --ciphertext-blob fileb:///tmp/root-token --encryption-context Tool=vault-unsealer | jq -r '.Plaintext' | base64 -D) for i in {0..4}; do aws ssm get-parameters --names kubernetes-vault-unseal-${i} --region us-west-2 | jq -r '.Parameters[].Value' | base64 -D > /tmp/unseal-${i} echo -n "UNSEAL_KEY_$((i+1)): $(aws kms decrypt --region $REGION --ciphertext-blob fileb:///tmp/unseal-${i} --encryption-context Tool=vault-unsealer | jq -r '.Plaintext' | base64 -D) " done echo "Initial Root token: $ROOT_TOKEN" #31 THanks for typing that out! Yes, my question was related to AWS KMS, sorry for not clarifying
gharchive/issue
2018-11-02T20:31:29
2025-04-01T06:39:12.308811
{ "authors": [ "innovia", "trondhindenes" ], "repo": "jetstack/vault-unsealer", "url": "https://github.com/jetstack/vault-unsealer/issues/29", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
67272239
Celery doesn't account for running and reserved tasks This is a similar problem to #3 . I'm using the Redis backend for Celery. I tried my hand at a solution in https://github.com/aspiredu/hirefire/commit/3ae625cc527330357a66df8d8c5e2701213d5af7, but it only works for small numbers of dynos, workers, and tasks. Even when it does work, it's painfully slow. The issue is that active, reserved, and scheduled tasks don't show up in the queue. Missing the scheduled ones is fine, but missing active and reserved means that my queues, which are running a lot of long-running tasks, get shut down well before stuff is done processing, which causes the dynos to flap as they are put back in the queue, then taken out, then when the process repeats. Unfortunately I can confirm that #9 has not solved this issue, even when I'm using py-ampq. I haven't tested with librabbitmq or iron-mq, but I find it likely that I'll get the same results. An IRC conversation with @ask gave me two ideas to try. Both are somewhat involved for something as light as HireFire, but we've gotta do what we've gotta do. Use Flower to get data on the running worker queues. Use heartbeats from worker to get up-to-date data. (from me) use inspect() (that's how flower gets it's info) for all queues/procs/workers at once and somehow re-use that data for the life of the request to the HireFire info URL. 2 requires some upstream code-changes to Celery (the heartbeats don't currently include scheduled and reserved counts), and also requires a persistent monitor to receive those events. I'm honestly not sure which path to take first. I'll just have to take 1 and run with it. @ryanhiebert did you find any solution for this trouble ? No solution yet. It's been somewhat back-burnered in my current workload, but trying to use Flower to get a better picture of the running queues seems like the route I'm likely to try first. Any and all help would be welcome! Any updates on this issue? No, sorry. Haven't been able to work on it any more yet. I wrote a basic implementation that seems to work. import itertools from celery import Celery from django.conf import settings from gevent.pool import Group from hirefire.procs.celery import CeleryProc def _inspect(method): app = Celery('app', broker=settings.BROKER_URL) inspect_result = getattr(app.control.inspect(), method)() app.close() return inspect_result class OFCeleryProc(CeleryProc): def quantity(self): """ Returns the aggregated number of tasks of the proc queues. """ if hasattr(self.channel, '_size'): # Redis return sum(self.channel._size(queue) for queue in self.queues) # AMQP (Include both libs, since we have both installed) from librabbitmq import ChannelError as LCE from amqp.exceptions import ChannelError as ACE count = 0 for queue in self.queues: try: queue = self.channel.queue_declare(queue, passive=True) except (LCE, ACE): # The requested queue has not been created yet pass else: count += queue.message_count active, reserved = Group().map(_inspect, ('active', 'reserved')) num_active = sum(1 for v in itertools.chain.from_iterable(active.itervalues()) if not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues) if active else 0 num_reserved = sum(1 for v in itertools.chain.from_iterable(reserved.itervalues()) if not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues) if reserved else 0 return count + num_active + num_reserved @joshowen : That can work OK, for a small number of queues. Unfortunately, I have around 20 queues, and the combined time of running this made for timeouts in Heroku. I've got to find something faster. https://github.com/celery/celery/issues/2839 This issue with Celery also makes this issue worse. Tasks that are running are immediately shut down, instead of being given the time that Heroku usually gives to let the dynos shut down gracefully. Having this fixed wouldn't be sufficient by itself, but it would make the problem less of a deal for short tasks. @ryanhiebert Even using gevent? @joshowen: Actually, that's not what I had done before. I'm going to have to dig more into it, and see if it works for me. I'm not currently using gevent. With gevent I was able to run the _inspect calls in parallel and got around the heroku timeouts Let me try it. I still do have quite a few queues (not just workers), with different procs, so even with gevent it means that I'll have to run ~20 of these to get all the stats for HireFire. Worth a try, though. @joshowen thanks for your example code, we just ran into the same issue. I build a similar solution, still slow, but less calls: the app.control.inspect() calls always return the data for every queue, not only one. We have just extended HireFireMiddleware to do these calls once (not only for active and reserved, but also for scheduled) and then the subsequent CeleryProcs will get the data from their queues out of it. So queue size does not have any effect on the calls, perhaps only on the time celery needs to call all the workers. (cc @ryanhiebert ) will try to bake this into a PR, I think this is a common problem if you want to scale to 0 :) @syphar: that sounds awesome. I'm looking forward to it. @syphar: I'm working on this now for work, and if possible I'd like to see your approach. Any chance you could let me see your work (via very raw, perhaps broken PR, or perhaps a gist of the interesting code), so that I can take inspiration at least? @ryanhiebert yes, sure. Sorry for the delay many things happening :) middleware: class HireFireMiddleware(OriginalHireFireMiddleware): """ An extended HireFireMiddleware which includes counts worker-stats. we query for running, reserved or scheduled tasks and add them to the queued tasks per queue. The related procs have to have a method called ``quantity_with_running`` which will get the inspection data, so it's queried only once per call from hirefire. """ def dump_procs(self, procs): celery_inspect = { method: getattr(app.control.inspect(), method)() for method in ('active', 'reserved', 'scheduled') } data = [ { 'name': name, 'quantity': proc.quantity_with_running(celery_inspect) or 'null', } for name, proc in procs.items() ] return json.dumps(data, cls=TimeAwareJSONEncoder, ensure_ascii=False) def info(self, request): payload = self.dump_procs(self.loaded_procs) return HttpResponse(payload, content_type='application/json') procs.py class UpdatedCeleryProc(CeleryProc): def quantity_with_running(self, celery_inspect): count = super(UpdatedCeleryProc, self).quantity() for func, tasks in celery_inspect.items(): if not tasks: continue queue_tasks = [ v for v in itertools.chain.from_iterable(tasks.values()) if ( not v.get('acknowledged', False) and v.get('delivery_info', {}).get('exchange') in self.queues ) ] count += len(queue_tasks) return count class LowPrioWorkerProc(UpdatedCeleryProc): name = 'celeryworkerlowprio' queues = ['lowprio'] this code as expected: only works with celery not perfect: increases response time (it's around ~3s on our production system) good: the amount of procs/queues doesn't matter, since the inspect result contains everything. runs in production for 4 weeks here only tested on python 3.5 What was blocking me (when I had some time) was the question how to design this the best way to make it optional make it backend-agnostic Awesome. I'll work on figuring out those when I'm implementing it for myself. This is very, very, helpful. Thank you. @ryanhiebert we can get this into a working PR together, if you have some nice ideas on the design :) Would you be up for a pairing session, perhaps? You can hit me up on irc.freenode.net, I'm ryanhiebert, or you can email me ryan@ryanhiebert.com If you're interested in pairing, I've set up a floobits workspace at https://floobits.com/ryanhiebert/hirefire. I do have a good idea, and that's to send a cache dictionary into the procs, so that they can share state between themselves. I'm working up an implementation. Here's my work-in-progress branch on the approach I came up with. Feel free to give me feedback. I'll open up a PR after I've done some more testing in my environment. https://github.com/aspiredu/hirefire/tree/global-cache
gharchive/issue
2015-04-09T03:28:10
2025-04-01T06:39:12.347868
{ "authors": [ "joshowen", "ryanhiebert", "syphar", "touilleMan" ], "repo": "jezdez/hirefire", "url": "https://github.com/jezdez/hirefire/issues/8", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1251746594
Support gstreamer Need to write an example using the macro with an gstreamer element, and possibly create another gst_element macro to make it easier. I implemented a webrtcsrc element based on this crate, I think we have mostly everything we need tbh. I also started adding macros to ease GstElement implementation here, at this point it implements element metadata and pad templates declaration in the macros like: #[gobject::gst_element( class(final), long_name = "TheTestElement", classification = "Test/Filter", description = "Just a test", author = "Thibault Saunier <tsaunier@igalia.com>", pad_templates( src(presence="always"), // `__` is transformed to `_%` as "%" is not a valid character sink__u(direction="sink", presence="sometimes", caps="video/x-raw"), ) )] My next step will be to improve the way we interact with Pads creation and integration into the Elements, maybe adding a macros like: #[gobject::gst_pad(sink__u)] impl SrcPad { fn chain() -> Result<gst::FlowSuccess, gst::FlowError> { ... } } but I am still thinking.
gharchive/issue
2022-05-28T23:44:27
2025-04-01T06:39:12.351914
{ "authors": [ "jf2048", "thiblahute" ], "repo": "jf2048/gobject", "url": "https://github.com/jf2048/gobject/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2301986153
Creation of a roadmap section Idea Contribution [X] I have read all the feature request issues. [X] I'm interested in working on this issue [X] I'm part of GSSOC organization Explain feature request Creation of a roadmap section,which will contact roadmap and their respected study material and links for various fields, Android development,web development ,ml , DSA etc Explain your solution No response Any alternative approaches/features No response Additional Context No response @jfmartinz kindly assign me this issue @jfmartinz Maybe this issue is not solved yet! I would like to resolve this issue ASAP, creating a roadmap section for almost every tech field- Data Science, Web Development, App Development, UI/UX Design, and Machine Learning... Kindly assign me this issue, and I'll resolve it soon. This is actually completed but the author forgot to reference the PR. but you can still add some resources that youd like and create a PR for that, just make sure that those resources are not duplicate, thanks @jfmartinz ok then I will do so as a separate pr
gharchive/issue
2024-05-17T07:08:50
2025-04-01T06:39:12.374180
{ "authors": [ "Soumya6Tiwari", "jfmartinz", "renunegii" ], "repo": "jfmartinz/ResourceHub", "url": "https://github.com/jfmartinz/ResourceHub/issues/276", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
903925621
make sure statuses are included with the buildinfo. [ ] All tests passed. If this feature is not already covered by the tests, I added new tests. [x] This pull request is on the dev branch. [x] I used gofmt for formatting the code before submitting the pull request. fixes #356 @AnalogJ, Since we're currently considering to make major changes and improvements to the entire build promotion process, I suggest that for now, we wait with changes to the build-info schema to include promotion statuses. Hi @eyalbe4 I'd really like to get this merged if possible. Any chance you could provide an update about the future of Promotion status's in this library? I've also rebased my branch on the latest from master. Hey @eyalbe4 @yahavi Any additional feedback for this PR? @eyalbe4 @yahavi any updates here? Sorry to keep bugging you @AnalogJ, Since this change may conflict with other planned changes to the build promotion process, will you be able to share more details about the need? How are you planning to use and consume this data? With this information, maybe we'll be able to come up with an alternative solution. Hey, This is a critical feature for us. Basically we use the Jfrog Go SDK to create "Builds". The builds contain a list of versioned artifacts that will be used in a deployment. These artifacts are then scanned using XRAY, and deployed to various non-production environments automatically. The "PromotionStatus" field is used as a manual flag to ensure that Production Deployments can only use vetted/production-ready artifacts. I see @AnalogJ. Would you be willing perhaps to tag the build-info artifact in Artifactory, with a property that will indicate the status of the build? Every time you publish a build-info, it is also saved as a file under the artifactory-build-info repository. You can set and get the value of this custom property on/from the build-info artifact, to support your logic. This functionality is supported by this library as well. See the Setting Properties on Files in Artifactory and Deleting Properties from Files in Artifactory sections in the README. This PR doesnt seem relevant since https://github.com/jfrog/jfrog-client-go/blob/master/artifactory/services/promote.go is merged. Closing.
gharchive/pull-request
2021-05-27T15:42:43
2025-04-01T06:39:12.384337
{ "authors": [ "AnalogJ", "eyalbe4" ], "repo": "jfrog/jfrog-client-go", "url": "https://github.com/jfrog/jfrog-client-go/pull/357", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1320061483
Some defaults in the user HCL dont match the description Describe the bug profile_updatable and disable_ui_access default to True when the docs say they should be False Requirements for and issue [x] A description of the bug [ ] A fully functioning terraform snippet that can be copy&pasted (no outside files or ENV vars unless that's part of the issue). If this is not supplied, this issue will likely be closed without any effort expended. [ ] Your version of artifactory (you can curl it at $host/artifactory/api/system/version [ ] Your version of terraform [ ] Your version of terraform provider Expected behavior Docs match actuals Additional context Add any other context about the problem here. https://github.com/jfrog/terraform-provider-artifactory/pull/517 Actually docs here say it should be true https://github.com/jfrog/terraform-provider-artifactory/blob/master/docs/resources/user.md not sure what is the correct default value and which bit is wrong :) @apr-1985 Per my comment, it's the Description that is incorrect.
gharchive/issue
2022-07-27T19:57:56
2025-04-01T06:39:12.389830
{ "authors": [ "alexhung", "apr-1985" ], "repo": "jfrog/terraform-provider-artifactory", "url": "https://github.com/jfrog/terraform-provider-artifactory/issues/518", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
45118362
issue 470 allow creation of location with only try_files This commit will fix #470. @buckstr the proposed PR won't allow you to create a location with only try_files. As is, if you only set try_files and don't set proxy, location_alias, stub_status, fastcgi or www_root then the template vhost/locations/empty.erb will be rendered. This doesn't include try_files. Hi -- There's been no further activity on this since the 'needs Rebase' and 'needs tests' labels were set. If we don't have any further comments in the next month, this PR will be closed. Addressed in #834 already.
gharchive/pull-request
2014-10-07T15:05:19
2025-04-01T06:39:12.392383
{ "authors": [ "3flex", "buckstr", "wyardley" ], "repo": "jfryman/puppet-nginx", "url": "https://github.com/jfryman/puppet-nginx/pull/471", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
476590766
Conversion from FromSqlError to Error Error needs column index, but FromSqlError doesn't have it. I've fudged it by putting obviously-wrong index value, and then hiding that fact in Display. Ideally the Error should use Option<usize> for index (and type), but I assume it's better to avoid unnecessary backwards-compat break. Fixes #555 The test failure looks like a vcpkg installation issue, unlikely to be related to this change. Thank you.
gharchive/pull-request
2019-08-04T20:27:50
2025-04-01T06:39:12.406462
{ "authors": [ "gwenn", "kornelski" ], "repo": "jgallagher/rusqlite", "url": "https://github.com/jgallagher/rusqlite/pull/556", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
52574994
feature request: routie.back(), routie.next() it would be great to have back and next method for navigating through history What's wrong with using window.history.back() and window.history.forward()? These should work out of the box. I haven't dug into why, but in cordova window.history.back() and window.history.forward() don't seem to work -- at least not on the emulated Nexus 4.
gharchive/issue
2014-12-20T20:00:55
2025-04-01T06:39:12.407982
{ "authors": [ "GinkgoFJG", "akaRem", "benplum" ], "repo": "jgallen23/routie", "url": "https://github.com/jgallen23/routie/issues/35", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
130815186
Display an error message when file upload fails A common error that has been occurring is the failure of the /api/upload endpoint. While I'd like to get around to finding out why the error is happening so frequently, the least we can do is show an error message. Here is the console output of when it happened to me twice today: Looks good to me. See my one comment. It could be that disk space on the server is getting low. It might be worth looking into. I think that might be it. Disk space, I mean. Hiding the alert on upload is a good idea. I'll update the PR. On Feb 2, 2016 7:40 PM, "James Lamine" notifications@github.com wrote: Looks good to me. See my one comment. It could be that disk space on the server is getting low. It might be worth looking into. — Reply to this email directly or view it on GitHub https://github.com/jglamine/calvinwebprint/pull/49#issuecomment-178917875 .
gharchive/pull-request
2016-02-02T21:02:15
2025-04-01T06:39:12.457891
{ "authors": [ "jglamine", "tylervz" ], "repo": "jglamine/calvinwebprint", "url": "https://github.com/jglamine/calvinwebprint/pull/49", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1560036992
Bring "Convert labels to SVG" export option to Desktop version + conversion issue [ ] I agree to follow the Code of Conduct that this project adheres to. [ ] I have searched the issue tracker for a feature request that matches the one I want to file, without success. So, I've this big file I'm working on that has a lot (really a lot) of text labels and I'll have it to upload it to Wikimedia Commons. Since Commons doesn't like certain tags about text labels I have to convert that into vectorized text, feature that is only available on the web version. I'd really like to have it on Deskto p so that I wouldn't have to switch working environment every time I need to export my drawings. Second, since this file I'm working on is as I said pretty text intensive, the converter struggles to convert all the text and ends with an error. I'll upload the file for context so that you can replicate the issue. Carrara-Avenza rev. 3 (2023-01-23).zip Cheers! convert to SVG requires a back-end, it can't be ported to desktop.
gharchive/issue
2023-01-27T16:20:50
2025-04-01T06:39:12.496373
{ "authors": [ "OverFloyd", "davidjgraph" ], "repo": "jgraph/drawio", "url": "https://github.com/jgraph/drawio/issues/3319", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
305977313
The future is ember-animated I propose to either sunset this addon and give people a transition path to ember-animated or perhaps reimagine it as a thin wrapper atop ember-animated. Thoughts welcome! ember-sortable has accomplished exactly what I attempted to create in an app I'm working on in a reliable way. I'd hate to see it go, but if the plan is to sunset the addon then I'd be happy to replace what we have with a thin wrapper atop of ember-animated should that be created. I still think this is worth exploring, but it’s a long way off so closing for now.
gharchive/issue
2018-03-16T15:42:24
2025-04-01T06:39:12.504075
{ "authors": [ "jgwhite", "joe-wroten-q2" ], "repo": "jgwhite/ember-sortable", "url": "https://github.com/jgwhite/ember-sortable/issues/178", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
835750943
Mongock is not compatible with the native build Describe the bug Mongock is a library to help us to handle the data migration with a MongoDb database. JHipster already uses it with Mongo and we reused the same with Quarkus. Mongock uses org.reflections that cause an error when we build it with GraalVM to generate a native executable. To Reproduce Steps to reproduce the behavior: Generate a project with MongoDb as a database. Install GraalVM openjdk 11.0.10 2021-01-19 OpenJDK Runtime Environment GraalVM CE 21.0.0.2 (build 11.0.10+8-jvmci-21.0-b06) OpenJDK 64-Bit Server VM GraalVM CE 21.0.0.2 (build 11.0.10+8-jvmci-21.0-b06, mixed mode, sharing) Run the following command: ./mvnw clean verify -Pnative -DskipTests Caused by: com.oracle.graal.pointsto.constraints.UnsupportedFeatureException: Unsupported features in 2 methods Detailed message: Error: com.oracle.graal.pointsto.constraints.UnresolvedElementException: Discovered unresolved type during parsing: org.apache.commons.vfs2.VFS. To diagnose the issue you can use the --allow-incomplete-classpath option. The missing type is then reported at run time when it is accessed the first time. Trace: at parsing org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(Vfs.java:287) Call path from entry point to org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(URL): at org.reflections.vfs.Vfs$DefaultUrlTypes$7.createDir(Vfs.java:287) at org.reflections.vfs.Vfs.fromURL(Vfs.java:98) at org.reflections.vfs.Vfs.fromURL(Vfs.java:90) at org.reflections.Reflections.scan(Reflections.java:236) at org.reflections.Reflections$1.run(Reflections.java:199) at java.lang.Thread.run(Thread.java:834) at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:519) at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192) at com.oracle.svm.core.code.IsolateEnterStub.PosixJavaThreads_pthreadStartRoutine_e1f4a8c0039f8337338252cd8734f63a79b5e3df(generated:0) Error: com.oracle.graal.pointsto.constraints.UnresolvedElementException: Discovered unresolved type during parsing: org.apache.commons.vfs2.VFS. To diagnose the issue you can use the --allow-incomplete-classpath option. The missing type is then reported at run time when it is accessed the first time. Trace: at parsing org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(Vfs.java:281) Call path from entry point to org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(URL): at org.reflections.vfs.Vfs$DefaultUrlTypes$7.matches(Vfs.java:281) at org.reflections.vfs.Vfs.fromURL(Vfs.java:97) at org.reflections.vfs.Vfs.fromURL(Vfs.java:90) at org.reflections.Reflections.scan(Reflections.java:236) at org.reflections.Reflections$1.run(Reflections.java:199) at java.lang.Thread.run(Thread.java:834) at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:519) at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192) at com.oracle.svm.core.code.IsolateEnterStub.PosixJavaThreads_pthreadStartRoutine_e1f4a8c0039f8337338252cd8734f63a79b5e3df(generated:0) at com.oracle.graal.pointsto.constraints.UnsupportedFeatures.report(UnsupportedFeatures.java:129) at com.oracle.svm.hosted.NativeImageGenerator.runPointsToAnalysis(NativeImageGenerator.java:770) ... 8 more Expected behavior The native build should succeed without errors. Suggestions Move from Mongock to Liquibase with Mongo. Find a way to make Mongock compatible with the native build. Additional context Use ./mvnw dependency:tree -Dverbose to check which library is using VFS. Check this issue: https://github.com/quarkusio/quarkus/issues/9801 cc @danielpetisme closed by #180 I have just realised about this ticket and it seems that has been somehow sorted out. However is there anything that we can provide natively from Mongock to help on this or improve. Just let me know.
gharchive/issue
2021-03-19T09:28:00
2025-04-01T06:39:12.520004
{ "authors": [ "avdev4j", "dieppa" ], "repo": "jhipster/generator-jhipster-quarkus", "url": "https://github.com/jhipster/generator-jhipster-quarkus/issues/179", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
494845895
Add caffeine option for cacheProvider Overview of the feature request Caffeine is a new option available for cache since JHipster 6.3.0 Motivation for or Use Case It should be possible to select this option in JDL Related issues or PR https://github.com/jhipster/generator-jhipster/pull/10303 [X] Checking this box is mandatory (this is just to show you read everything) @murdos : Thanks for adding this; seems I've missed this. 😄
gharchive/issue
2019-09-17T20:51:33
2025-04-01T06:39:12.577523
{ "authors": [ "SudharakaP", "murdos" ], "repo": "jhipster/jhipster-core", "url": "https://github.com/jhipster/jhipster-core/issues/364", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2720110144
Gradle CI doesn't launch unit tests It seems that the gradle ci don't launch unit test only IT. I have tried to put a failing unit test and the ci was green Is this the behaviour or a misconfiguration on my side ? A possible working solution github-actions.yml jobs: tests: name: tests runs-on: ubuntu-latest timeout-minutes: 20 steps: - name: 'Setup: checkout project' uses: actions/checkout@v4 - name: 'Setup: environment' id: setup uses: ./.github/actions/setup - name: Setup Gradle uses: gradle/gradle-build-action@v2 - name: 'Test: run unit tests' run: ./gradlew clean test --no-daemon - name: 'Test: run integration tests' run: ./gradlew clean integrationTest --no-daemon @fabienpuissant : I looked at a random build action: feat(liquibase): async liquibase configuration should not be activated if liquibase is disabled. At least for this example, the tests were executed: > Task :test JwtReaderTest > should not authenticate user from empty token STANDARD_OUT 2024-12-05T23:23:09.354Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid empty JWT token JwtReaderTest > should not authenticate user from blank token STANDARD_OUT 2024-12-05T23:23:09.359Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid empty JWT token JwtReaderTest > should not authenticate user from invalid token STANDARD_OUT 2024-12-05T23:23:09.481Z INFO --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid JWT token 2024-12-05T23:23:09.482Z TRACE --- [ Test worker] t.j.f.s.a.i.primary.JwtReader : Invalid JWT token: Invalid compact JWT string: Compact JWSs must contain exactly 2 period characters, and compact JWEs must contain exactly 4. Found: 0 Here is the part of the github-actions.yml responsible for that: - name: 'Test: verify ${{ matrix.app }}' if: steps.tests-requirement-check.outputs.execute_tests == 'true' working-directory: /tmp/jhlite/${{ matrix.app }}/ run: | if [ -f 'mvnw' ]; then ./mvnw clean verify -Dsonar.qualitygate.wait=true sonar:sonar elif [ -f 'gradlew' ]; then ./gradlew clean build sonar --no-daemon --info else npm install npm test fi Please, could you provide an example that reproduces the not desired behavior? @renanfranca It's on a generated app See https://github.com/fabienpuissant/test-jh-lite-gradle-ci/actions/runs/12202622941 should fail because of : https://github.com/fabienpuissant/test-jh-lite-gradle-ci/blob/main/src/test/java/com/mycompany/myapp/FailingTest.java In the test-ci for jhlite it seems to be ok 😄 Not 100% sure just on my mobile, but with the soft ordering should run after for the integration test phase this does not define a dependency. So when just executing integrationtest this does not imply running unit tests. To work out of the box without Jacoco, it is going need to change the Gradle module to directly link unit and integration tests. Then, when adding Jacoco, remove that link to make Jacoco the glue. @fabienpuissant : I think your solution is better and more flexible 😉. Even if we add the Jacoco module, the CI configuration will need to be updated to execute Jacoco instead of the IntegrationTest task. IMHO, If you want, feel free to proceed with the implementation—if not, I can take care of it on Monday 👍. Thanks for reporting that improvement/bug! cc: @pascalgrimaud , @murdos and @atomfrede
gharchive/issue
2024-12-05T11:11:31
2025-04-01T06:39:12.587627
{ "authors": [ "atomfrede", "fabienpuissant", "renanfranca" ], "repo": "jhipster/jhipster-lite", "url": "https://github.com/jhipster/jhipster-lite/issues/11513", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1054632631
Typo in Equation 5 in paper? Equation 5 in the paper makes use of f_S, however, the corresponding portion in Figure 2 (orange) only makes use of f_T. Additionally, f_T makes more sense since the input is in the target domain. Should the f_S in equation 5 be f_T? ”Note that we can also consider a feature-level method Which distinguishes the features or semantics from two image sets as viewed under a task network. “ I think both f _ s and f _ t can do experiments.
gharchive/issue
2021-11-16T09:14:09
2025-04-01T06:39:12.602707
{ "authors": [ "arjung128", "zhe-juanz" ], "repo": "jhoffman/cycada_release", "url": "https://github.com/jhoffman/cycada_release/issues/41", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2447264123
"This is the solution of the issue" > "This is the solution of the issue" this is a quote Originally posted by @jhon-elen in https://github.com/jhon-elen/Prueba3/issues/7#issuecomment-2267641691 Anothe comment @jhonelentest Header Header Cell Cell Cell Cell import os os.chdir() Details This is a detail
gharchive/issue
2024-08-04T19:18:41
2025-04-01T06:39:12.608616
{ "authors": [ "jhon-elen" ], "repo": "jhon-elen/Prueba3", "url": "https://github.com/jhon-elen/Prueba3/issues/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1503778256
Abrocoma bennetti is mentioned in MSW1/2 but Abrocoma bennettii was mentioned in original publication via @myrmoteras - [...] For example, Abrocoma bennetti does not match COL because it is spelled with out a ..ii in MSW2 (and also in MSW1, and 3). The original in has a ...ii though. https://doi.org/10.5281/zenodo.7457193 or https://github.com/plazi/treatments-xml/blob/main/data/03/FA/64/03FA64167865FF8AFF34F96C4ECCA3D4.xml and with access to the original publication. now also in BLR. [...] also via @myrmoteras @myrmoteras in running preston ls | preston grep -o --no-line "(Abrocoma bennetti)([^i])" after cloning Plazi Community. (2022). Plazi Treatments XML Archive hash://sha256/3cfd60b8b19e76d208377537835de92efdb5b945a6a71765b74ed2fe22298b42 hash://md/594923284e3eb9965b8cbad149c76cd0f (hash://sha256/3cfd60b8b19e76d208377537835de92efdb5b945a6a71765b74ed2fe22298b42) [Data set]. Zenodo. https://doi.org/10.5281/zenodo.7443343 I was able to cite all locations of Abrocoma bennetti without the ...ii suffix in a recent plazi/treatments-xml corpus and appended them to annotations.csv via https://github.com/jhpoelen/msw-plazi/commit/ea24ff964ea1373302ac01d406a0ac9226f65c68 . The associated treatment metadata headers in which the name (without ii suffix) occurs are: <document ID-DOI="http://doi.org/10.5281/zenodo.6228755" ID-GBIF-Dataset="aa7a2f2b-ea55-4ae2-863a-93bf3bdde535" ID-GBIF-Taxon="119379023" ID-ISSN="1175-5326" ID-Zenodo-Dep="6228755" checkinTime="1459984091607" checkinUser="plazi" docAuthor="Valim, Michel P. &amp; Linardi, Pedro Marcos" docDate="2008" docId="03948456FFB1FFED26E95090DEA41A59" docLanguage="en" docName="zt01899p024.pdf" docOrigin="Zootaxa 1899" docStyle="DocumentStylede.uka.ipd.idaho.easyIO.settings.Settings@1effd53c" docStyleName="zootaxa.2007.monograph" docTitle="Gyropus scalaris Werneck 1942" docType="treatment" docVersion="6" lastPageId="16" lastPageNumber="17" masterDocId="FFADFC2EFFBFFFFD267E5649DD2D1865" masterDocTitle="A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae)" masterLastPageNumber="24" masterPageNumber="1" pageId="14" pageNumber="15" updateTime="1645556127074" updateUser="ExternalLinkService"> <document ID-DOI="http://doi.org/10.5281/zenodo.4592306" ID-GBIF-Dataset="e61aa52f-2e17-4afc-b30a-6cd468eea36b" ID-GBIF-Taxon="180302034" ID-ISSN="1545-1410" ID-Zenodo-Dep="4592306" ID-ZooBank="B8876E05-78A8-45FA-A34A-9FFB56741A20" checkinTime="1614695199489" checkinUser="felipe" docAuthor="Pérez, M. Julieta &amp; Díaz, M. Mónica" docDate="2018" docId="03970124E858FFE6FF5BB08A2EBC056C" docLanguage="en" docName="MammalSpecies.50.963.74-83.pdf" docOrigin="Mammalian Species 50 (963)" docSource="http://dx.doi.org/10.1093/mspecies/sey010" docStyle="DocumentStyle:E260F2FEBB038053A7447521347BED36.2:MammalSpecies.2016-.journal_article" docStyleId="E260F2FEBB038053A7447521347BED36" docStyleName="MammalSpecies.2016-.journal_article" docStyleVersion="2" docTitle="Octodontomys gliroides" docType="treatment" docVersion="4" lastPageId="6" lastPageNumber="80" masterDocId="FFAE795CE859FFE0FF8AB0412A6B0312" masterDocTitle="Octodontomys gliroides (Rodentia: Octodontidae)" masterLastPageNumber="83" masterPageNumber="74" pageId="1" pageNumber="75" updateTime="1643466530690" updateUser="ExternalLinkService" zenodo-license-document="CLOSED"> <document ID-DOI="http://doi.org/10.5281/zenodo.6228735" ID-GBIF-Dataset="aa7a2f2b-ea55-4ae2-863a-93bf3bdde535" ID-GBIF-Taxon="119379012" ID-ISSN="1175-5326" ID-Zenodo-Dep="6228735" checkinTime="1459984091607" checkinUser="plazi" docAuthor="Valim, Michel P. &amp; Linardi, Pedro Marcos" docDate="2008" docId="03948456FFB8FFFA26E952E1DE221F7B" docLanguage="en" docName="zt01899p024.pdf" docOrigin="Zootaxa 1899" docStyle="DocumentStylede.uka.ipd.idaho.easyIO.settings.Settings@1effd53c" docStyleName="zootaxa.2007.monograph" docTitle="Gyropus longus Neumann 1912" docType="treatment" docVersion="6" lastPageNumber="8" masterDocId="FFADFC2EFFBFFFFD267E5649DD2D1865" masterDocTitle="A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae)" masterLastPageNumber="24" masterPageNumber="1" pageId="7" pageNumber="8" updateTime="1645556127074" updateUser="ExternalLinkService"> <document ID-DOI="http://doi.org/10.5281/zenodo.7353039" ID-ISBN="0-89327-235-3" ID-Zenodo-Dep="7353039" approvalRequired="41" approvalRequired_for_document="2" approvalRequired_for_matCits="35" approvalRequired_for_originalDoi="1" approvalRequired_for_treatments="3" checkinTime="1665792792562" checkinUser="GgServerImporter" docAuthor="James H. Honacki, Kenneth E. Kinman &amp; James W. Koeppl" docDate="1982" docId="03BC87A94624FFB823F2F6CB530FF624" docLanguage="en" docName="MammalSpeciesofTheWorld.1982.Rodentia.560-594.pdf.imd" docOrigin="Mammal Species of the World (1 st Edition), Lawrence, Kansas, USA: Alien Press, Inc. &amp; The Association of Systematics Collections" docTitle="Abrocoma bennetti Waterhouse 1837" docType="treatment" docVersion="3" lastPageNumber="585" masterDocId="FF85FFD1463DFFA12306FFF55074FFFA" masterDocTitle="Order Rodentia (Part 6)" masterLastPageNumber="594" masterPageNumber="560" pageNumber="585" updateTime="1669255589438" updateUser="ExternalLinkService"> <document ID-DOI="http://doi.org/10.5281/zenodo.7353083" ID-ISBN="1-56098-217-9" ID-Zenodo-Dep="7353083" approvalRequired="3" approvalRequired_for_document="2" approvalRequired_for_originalDoi="1" checkinTime="1667534037376" checkinUser="GgServerImporter" docAuthor="Charles A. Woods" docDate="1993" docId="E11F878EFFCAFFC5FF3DF9DBFC54D984" docLanguage="en" docName="MammalSpeciesofTheWorld.1993.Hystricognathi.771-806.pdf.imd" docOrigin="Mammal Species of the World (2 nd Edition), Washington and London: Smithsonian Institution Press" docTitle="Abrocoma bennetti Waterhouse 1837" docType="treatment" docVersion="3" lastPageNumber="789" masterDocId="1D26FFF6FFD8FFD7FFE1FC69FFB0DF05" masterDocTitle="Order Rodentia - Suborder Hystricognathi" masterLastPageNumber="806" masterPageNumber="771" pageNumber="789" updateTime="1669256602727" updateUser="ExternalLinkService"> <document ID-DOI="http://doi.org/10.5281/zenodo.7353083" ID-ISBN="1-56098-217-9" ID-Zenodo-Dep="7353083" approvalRequired="3" approvalRequired_for_document="2" approvalRequired_for_originalDoi="1" checkinTime="1667534037376" checkinUser="GgServerImporter" docAuthor="Charles A. Woods" docDate="1993" docId="E11F878EFFCAFFC5FF50F949FDC9DA9C" docLanguage="en" docName="MammalSpeciesofTheWorld.1993.Hystricognathi.771-806.pdf.imd" docOrigin="Mammal Species of the World (2 nd Edition), Washington and London: Smithsonian Institution Press" docTitle="Abrocoma Waterhouse 1837" docType="treatment" docVersion="3" lastPageNumber="789" masterDocId="1D26FFF6FFD8FFD7FFE1FC69FFB0DF05" masterDocTitle="Order Rodentia - Suborder Hystricognathi" masterLastPageNumber="806" masterPageNumber="771" pageNumber="789" updateTime="1669256602727" updateUser="ExternalLinkService"> generated via cat ../msw-plazi2/annotations.csv\ | grep Donat\ | cut -d ',' -f6\ | sed 's/cut://g'\ | sed 's+!/b.*$++g'\ | sort\ | uniq\ | grep hash\ | parallel "preston cat {1} | head -n1" So,according to Plazi's recent treatments-xml corpus, the assumed invalid name (i.e., Abrocoma bennetti without ii prefix) appears in: Valim, Michel P., & Linardi, Pedro Marcos. (2008). Gyropus scalaris Werneck 1942. In A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae), pp. 1-24 in Zootaxa (Vol. 1899, pp. 15–17). Zenodo. https://doi.org/10.5281/zenodo.6228755 Pérez, M. Julieta, & Díaz, M. Mónica. (2018). Octodontomys gliroides. In Octodontomys gliroides (Rodentia: Octodontidae), pp. 74-83 in Mammalian Species (Vol. 50, Number 963, pp. 75–80). Zenodo. https://doi.org/10.5281/zenodo.4592306 Valim, Michel P., & Linardi, Pedro Marcos. (2008). Gyropus longus Neumann 1912. In A taxonomic catalog, including host and geographic distribution, of the species of the genus Gyropus Nitzsch (Phthiraptera: Amblycera: Gyropidae), pp. 1-24 in Zootaxa (Vol. 1899, p. 8). Zenodo. https://doi.org/10.5281/zenodo.6228735 James H. Honacki, Kenneth E. Kinman, & James W. Koeppl. (1982). Order Rodentia (Part 6). In Mammal Species of the World (1st Edition) (pp. 560–594). Alien Press, Inc. & The Association of Systematics Collections. https://doi.org/10.5281/zenodo.7353039 Charles A. Woods. (1993). Order Rodentia - Suborder Hystricognathi. In Mammal Species of the World (2nd Edition) (pp. 771–806). Smithsonian Institution Press. https://doi.org/10.5281/zenodo.7353083 and . . . annotations now exist to point to exactly which position of the texts they occur. You should actually be able to generate a little web widget that places the annotation exact on top of the location the suspected text occurs. . .
gharchive/issue
2022-12-19T23:39:55
2025-04-01T06:39:12.621723
{ "authors": [ "jhpoelen" ], "repo": "jhpoelen/msw-plazi", "url": "https://github.com/jhpoelen/msw-plazi/issues/2", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
1820088029
The document object obtained from parsing with jsoup.parse is missing one layer of elements compared to the original HTML. https://test.em.bonahl.com/#/user/login original : jsoup: https://github.com/jhy/jsoup/blob/1762412a28fa7b08ccf71d93fc4c98dc73086e03/src/main/java/org/jsoup/parser/HtmlTreeBuilderState.java#L417 It seems that nesting forms is not allowed here. However, modern frontend frameworks have implemented nested form nesting. Input: <div> <form id=1> <form id=2> </form> </form> jsoup parse: <div> <form id="1"> </form> </div> Chrome parse (Inspect Document) Spec: https://html.spec.whatwg.org/multipage/parsing.html#parsing-main-inbody A start tag whose tag name is "form" If the form element pointer is not null, and there is no template element on the stack of open elements, then this is a parse error; ignore the token. ... I believe jsoup is working per spec and to the same output as current browsers by only allowing one form on the stack at a time. Shadow DOM enables you to attach a DOM tree to an element, and have the internals of this tree hidden from JavaScript and CSS running in the page. Modern frontend frameworks have implemented neste HTML forms through Shadow DOM, and it have been widely used. jsoup should adapt to this.
gharchive/issue
2023-07-25T11:05:08
2025-04-01T06:39:12.661119
{ "authors": [ "666asd", "jhy", "skylakeliu" ], "repo": "jhy/jsoup", "url": "https://github.com/jhy/jsoup/issues/1983", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
341462100
no tabs on linux Description I assume that tabs are supported, since an issue for tabs feature-request has already been closed long time ago #116 and from some issues it seems that tabs are supported, ex in OSX : #647. However on linux (ubuntu 18.04) this is not the case, tabs open in new window. Steps to reproduce issue middle-click or ctrl-click a internal link opens a new window instead of a new tab. links are not either considered as external URLs because they do not open in web browser but in nativefier. Details Are you nativefying a public website ? => tried few websites, some hosted locally and some public websites. Feature request? Have you looked at nativefier --help to see if an existing option could fit your needs? => yes and I trired to specify --internal-urls but it did not change the behaviour. Full nativefier command used to build your app: nativefier --name "Gitlab" --icon "/path/to/gitlab/icon.png" --internal-urls "https://mydomain.comgitlab/*" "https://mydomain.com/gitlab" Version of Nativefier (run nativefier --version): v7.6.4 Version of node.js (run node --version): v9.2.0 OS: Ubuntu 18.04 The nativefier support for tabs comes from electron; support for native tabs in electron for macOS was added here: https://github.com/electron/electron/pull/9052 If electron gets updated to support tabs on more platforms then it should be relatively simple to add support for them to nativefier. It would be possible to change how middle-click and ctrl-click behave in nativefier, to open external links in an external browser when native tabs are unavailable, but I'm not convinced such a change would provide value, since external links already open in an external browser when you left-click on them. Being able to use middle-click to override the default "open externally" behavior seems useful to me. @monkeydri 👍 to what @bacongravy said. See Release Notes for Nativefier v7.6.3: macOS: Add tabs, used automatically instead of windows (PR #579). Provided by Electron and (so far) no available for Windows/Linux, contributions welcome. alright, thanks fro your answer. It's a pity, as I don't think native tabs will be available on linux/windows anytime soon on Electron as there is no equivalent feature on them as macOS native tabs (ie. an option to display multiple windows from an app as tabs). @bacongravy Can the module electron-tabs help to support tabbing in Linux?
gharchive/issue
2018-07-16T10:19:36
2025-04-01T06:39:12.670798
{ "authors": [ "bacongravy", "mhgharieb", "monkeydri", "ronjouch" ], "repo": "jiahaog/nativefier", "url": "https://github.com/jiahaog/nativefier/issues/655", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1394210871
Guidance running script Hello, very interested in the project, I would like to run the script. Supposedly there are some parameters, but I don't know how to add them. Running the script without parameters gives blank output. I also tried running the blocks of code as files, the first downloads model files and then errors at line 7: list indices must be integers or slices, not str How should one go about running this? I'm using windows Make sure you can run Whisper without any problems first. Next: import whisper from stable_whisper import modify_model model = whisper.load_model('base', 'cuda') modify_model(model) Then just use the model how you normally how you would use it. It should behave just like how whisper normally behaves, but with some additional values in the results. I'm having the same problem. I can use Whisper with no issues, but I always get this error when executing the example script. word_timestamps = results['segments']['word_timestamps'] TypeError: list indices must be integers or slices, not str I'm also using Windows 10 and I'm a complete novice, so apologies if I'm making a silly mistake. you can run like this but all the data is visible, so not very clean: make "run.py" file and put import whisper from stable_whisper import modify_model model = whisper/load_model('base', 'cuda') modify_model(model) results = model.transcribe('out.wav') print(results) type python run.py in the terminal. Maybe @jianfch can write a packaged demo script later. This is not for command line tools Yes, but that does not have single word timings, thats just generating a normal (better than youtube though w/punctuation) .srt I'd like to make .ass timed text subtitles for all videos. so I need the timestamps and word data Not sure about .ass, but you can change results_to_sentence_srt to results_to_token_srt to get the word timings. FYI - The "word" timestamps are actually token timestamps. Since token:word is not always 1:1 (varies by language), you may need to do some additional processing to get individual word timings. Thanks, that's super helpful! word_timestamps = results['segments']['word_timestamps'] TypeError: list indices must be integers or slices, not str that was a mistake on my part for not updating the execution lines in readme word_timestamps are under each segment, instead of the segments which is a list so the word_timestamps for first segments would be: first_segment_token_timestamps = results['segments'][0]['word_timestamps'] if you want a list of just the word_timestamps word_timestamps = [seg['word_timestamps'] for seg in results['segments']]
gharchive/issue
2022-10-03T06:32:19
2025-04-01T06:39:12.679419
{ "authors": [ "ClashSAN", "jak3-taylor", "jianfch" ], "repo": "jianfch/stable-ts", "url": "https://github.com/jianfch/stable-ts/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
224355012
在uxcore中使用jsonp方式获取数据会报错 SyntaxError: Unexpected token : 具体什么错? 不好意思,忘了jsonp要求的返回数据格式不是json。已经通过修改header实现跨域访问。
gharchive/issue
2017-04-26T06:16:50
2025-04-01T06:39:12.687635
{ "authors": [ "isoloist", "jias" ], "repo": "jias/natty-fetch", "url": "https://github.com/jias/natty-fetch/issues/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2350364198
✨ チーム開発の資料を追加 概要 インターンで使用する資料を追加しました 変更点 README.mdの追加 docs.mdの追加 drawi.o integrationをrecommendationsに追加 スクラムマスターがlinterにかかってしまったのでfilterを追加した @dicenull 演習と目次を修正完了しました
gharchive/pull-request
2024-06-13T07:22:21
2025-04-01T06:39:12.693977
{ "authors": [ "kimflip" ], "repo": "jigintern/intern-dev-tutorial", "url": "https://github.com/jigintern/intern-dev-tutorial/pull/41", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
145102350
subregion_select dropdown has value numeric instead of alphabet code Hi there, As part of testing just observed that for country Denmark, the subregion select dropdown has value numeric. State Hovedstaden Midtjylland Nordjylland Sjælland Syddanmark On saving it, the same numeric value is being saved in table. Is it normal behaviour, any fix will be really appreciated. Those are the region codes as found by the iso-codes debian package so this is expected behavior unless that codes have changed since the last time the script was run in the carmen gem. Select tag looks like this: Hovedstaden Midtjylland Nordjylland Sjælland Syddanmark So on saving value, 84 saves in the database instead of state code..
gharchive/issue
2016-04-01T06:54:55
2025-04-01T06:39:12.706799
{ "authors": [ "arun-manthani", "ecbypi" ], "repo": "jim/carmen-rails", "url": "https://github.com/jim/carmen-rails/issues/58", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1319038763
cm-chessboard: Figure out the status of pawn promotion Pawn promotion seems an issue with all of these “just a board” packages. See closed issue “ Pawn promotion support? #19 at cm-chessboard. To the extent that I’m just replicating FEN strings, I don’t think there’s any problem. Where a need would arise is either (a) when I’m stepping through a main line (rather than using FENs) or (b) when the user is adding an line that has a promotion. Closed as out of scope for this repository.
gharchive/issue
2022-07-27T05:26:19
2025-04-01T06:39:12.722428
{ "authors": [ "jimratliff" ], "repo": "jimratliff/pgn4people-poc-demo", "url": "https://github.com/jimratliff/pgn4people-poc-demo/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
890704247
about the length of tokens Hello, I have read your paper and am quite interested in your work! There is a question about the tokens. I notice you truncat the passage tokens with 120 in MSMARCO Passage Retrieval, however, for ANCE, the original paper uses 512 tokens. So does the number of tokens have the impact on the accuracy? No We just don't have that many advanced GPUs to afford very long input :( BTW, the average passage length is about 70 tokens and truncating to 120 shouldn't be a problem. hhhhhhhhha! Thanks a lot for your reply! I have no multiple GPUs either! Lol! I agree with your settings!
gharchive/issue
2021-05-13T05:00:39
2025-04-01T06:39:12.792086
{ "authors": [ "KaishuaiXu", "jingtaozhan" ], "repo": "jingtaozhan/DRhard", "url": "https://github.com/jingtaozhan/DRhard/issues/6", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
39788993
Set default value while AutoMigrate If no default value, the query with gorm will get error. Hi @wb14123 I just pushed a commit, so even the data in database is NULL, gorm won't get any error now. So we don't need the default value for migrations Thank you.
gharchive/issue
2014-08-08T03:08:14
2025-04-01T06:39:12.799155
{ "authors": [ "jinzhu", "wb14123" ], "repo": "jinzhu/gorm", "url": "https://github.com/jinzhu/gorm/issues/190", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
97720510
Fixed issue https://github.com/jinzhu/gorm/issues/151 where postgresql connection problems fail silently For reference: https://github.com/jinzhu/gorm/issues/151 Errors connection to a postgresql fail silently, resulting in errors. This fix sends a ping right after the connection has been made. If the ping fails, we assume the connection is broken. Accepting better solutions. This quick fix did it for me :) Should we always Ping() the database after connection? It's not the most pretty solution. But it an extra safetycheck to verify that the connection is in fact alive. Should i rewrite the pr and make the Ping() generic? Hi @pariz Yes, I think it is a better solution than just check the postgres. I agree. Code has been updated. Epic oneliner! :) picked your code to the master, thank you.
gharchive/pull-request
2015-07-28T14:32:53
2025-04-01T06:39:12.802452
{ "authors": [ "jinzhu", "pariz" ], "repo": "jinzhu/gorm", "url": "https://github.com/jinzhu/gorm/pull/585", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
200065337
datas on ajax call not work correctly Problemes 1- The filter don't work. 2- After filtering with a range then the datas are correctly loaded. I don't have the probleme when in load datas on external files data (places.js) Please help me: Here is a part of my code: $(document).ready(function(){ var places = []; $.ajax({ type: 'GET', contentType: "json", url: 'http://www.mydomaine.com/client/places', async: false, success:function(data){ places = data; } }); renderCategories(); initInputs(); var place = places[0]; GoogleMap.init(place.coordinate.latitude, place.coordinate.longitude, places); var afterFilter = function(result){ $('#total_places').text(result.length); GoogleMap.updateMarkers(result); } afterFilter(places); //search: {ele: '#searchbox', fields: ['runtime']}, // With specific fields var FJS = FilterJS(places, '#places', { template: '#place-template', search: {ele: '#searchbox', fields: ['name', 'categories','disponible','regionId','departementId']}, callbacks: { afterFilter: afterFilter } }); FJS.addCriteria({field: 'rating', ele: '#rating_filter', type: 'range'}); FJS.addCriteria({field: 'disponible', ele: '#is_closed_criteria input:checkbox'}); FJS.addCriteria({field: 'categories', ele: '#categories_criteria input:checkbox', all: 'all'}); FJS.addCriteria({field: 'regionId', ele: '#region_criteria input:checkbox', all: 'all'}); FJS.addCriteria({field: 'departementId', ele: '#region_criteria input:checkbox', all: 'all'}); window.FJS = FJS; }); I'm having trouble understanding your problem. The Filter doesn't seem to give any results until you actually change the range? Are you initializing it with the min and max values from the start? Code looks correct, but can you send me your site url or HTML. It will help me to debug i.e How you initializing your range filters. I correct the problems : it's comes from the initiating of the range. Thanks a lot.
gharchive/issue
2017-01-11T11:11:26
2025-04-01T06:39:12.810632
{ "authors": [ "Acen", "diouf83", "jiren" ], "repo": "jiren/filter.js", "url": "https://github.com/jiren/filter.js/issues/143", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1308399302
Consider adding compute unit limit to bundles Problem Proposed Solution @buffalu What should the limit be?
gharchive/issue
2022-07-18T19:12:55
2025-04-01T06:39:12.822956
{ "authors": [ "anoushk1234", "buffalu" ], "repo": "jito-foundation/jito-solana", "url": "https://github.com/jito-foundation/jito-solana/issues/81", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
303241067
[iOS] Fix uncaught NSInvalidArgumentException in RTCPeerConnection's createAnswer WebRTC appears to fail RTCPeerConnection's createAnswer without an error message i.e. nil. But NSMutableDictionary does not accept nil as a value and causes an uncaught exception. Based on Daniel Ornelas' comment that the "error" key should actually be "NSLocalizedDescription" and looking at the error.localizedDescription, I suppose it's better to morph this one further. ok, let me know when it's ready to be merged
gharchive/pull-request
2018-03-07T20:05:38
2025-04-01T06:39:12.856594
{ "authors": [ "lyubomir", "paweldomas" ], "repo": "jitsi/react-native-webrtc", "url": "https://github.com/jitsi/react-native-webrtc/pull/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
138081139
Spaces vs Tabs Why are you using spaces now? Tabs makes the code look more clean and organized in my opinion. I'm not sure how its cleanliness/organization is adversely affected by this change, could you elaborate on that? I decided to switch to spaces mainly because of the inconsistency between different programs/websites/browsers/users. Through some of these lenses tabs come out as long as 8 spaces (rather than the standard 4, in VS), requiring more scrolling or resulting in more code flowing onto subsequent lines (much less readable), can be especially painful on phones/small resolution devices. Tabs also caused problems with 'pretty spaced' code (where params/array initializers are lined up nicely for easy reading) which could easily end up ugly. Admittedly the biggest issue here was that tabs were used for more than just the start-of-line to start-of-code portions; but making this change certainly helped identify/eliminate those that were in the middle and makes it very easy to do so in the future. It's also easier to find stray tabs than stray spaces, if keeping the project on a 100% consistent convention ever became relevant (using different forms I’ve heard, can cause lotsa merge issues). Productivity wise I find the two indentation options to be quite equivalent, Visual Studios can easily be set to turn tab key-presses into 4-spaces (Tools->Options->Text Editor->C/C++->Tabs); white space can be easily jumped with ctrl+arrows or deleted with ctrl+delete and inserting or re-inserting a ; or } automatically spaces chunks of code. For tabs the only advantages I read up on and considered legitimate was the ability for users to set their size (in some programs, not all give this option) and the relative ease for newer coders - which didn't, in my opinion, outweigh their disadvantages. Some discussion on the matter: http://programmers.stackexchange.com/questions/57/tabs-versus-spaces-what-is-the-proper-indentation-character-for-everything-in-e It appears I can't link SEN shoutbox convos, so http://www.staredit.net/shoutboxsearch/ + starter phrase "like my friend and I decide for our project that indentation is to be this many spaces" then hit context.
gharchive/issue
2016-03-03T05:38:56
2025-04-01T06:39:12.860669
{ "authors": [ "jjf28", "xboi209" ], "repo": "jjf28/Chkdraft", "url": "https://github.com/jjf28/Chkdraft/issues/92", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1270145971
Method for creating and printing board The user needs a way to put a sudoku board into the program. Boards should be printed in a way that is easy to read. A graphical user interface will be created using Swing. The GUI will allow for a user to create a sudoku board to be solved, as well as display solutions.
gharchive/issue
2022-06-14T01:45:35
2025-04-01T06:39:12.881072
{ "authors": [ "jjklemm" ], "repo": "jjklemm/sudoku-solver", "url": "https://github.com/jjklemm/sudoku-solver/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1868932522
Sub listed item in numbered list in 01-jupyter.ipynb Please check the numbered list in one of the first markdown cells of the notebook. For the sub listed item I think one more space in front of the "1." is necessary, otherwise the item just becomes part of the main list. fixed. thanks.
gharchive/issue
2023-08-28T04:23:24
2025-04-01T06:39:12.889890
{ "authors": [ "ccolomb2", "jkitchin" ], "repo": "jkitchin/f23-06623", "url": "https://github.com/jkitchin/f23-06623/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
148340174
Update code: issue #2 Updating model from controller does not update control Can you take this solution along in your code? Thanks scope.$watch(attrs['ngModel'], function (newTime) { scope.initTime = newTime; reinitTime(); }); Spot on! This fix was exactly what i was looking for. Can someone make a PR please? I had already made a PR: https://github.com/jkuri/ngTimepicker/pull/4 Sorry, I missed it. Thanks for your PR, it's merged.
gharchive/issue
2016-04-14T12:05:38
2025-04-01T06:39:12.910528
{ "authors": [ "elineopsommer", "jkuri", "leypoldt" ], "repo": "jkuri/ngTimepicker", "url": "https://github.com/jkuri/ngTimepicker/issues/3", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1478454439
Lack of support for valid endpoints Consider the following code from Usage with Express getUser: t.procedure.input(z.string()).query((req) => { req.input; // string return { id: req.input, name: 'Bilbo' }; }), After adding this line .meta({ openapi: { method: 'GET', path: '/getUser' } }) trpc-openapi fail to process it TRPCError: [query.getUser] - Input parser must be a ZodObject This is correct. - z.string() + z.object({ id: z.string() }) This is correct. You have linked to the tRPC documentation. trpc-openapi has a few additional requirements so you may need to make a few changes to your procedure. - z.string() + z.object({ id: z.string() }) Please make sure you have read this 👉 https://github.com/jlalmes/trpc-openapi/blob/master/README.md Well, it's string and not object (see the // string comment - it's there, in the documentation, not something I added), so it can't be correct. And thing is that we not only need to change the procedure, but change API requests as well, which is too high price to pay. Hi Sergey. Please read the README before commenting again. so it can't be correct trpc-openapi requires that your inputs are wrapped in an z.object. need to change the procedure Yes. change API requests I don't think you've understood REST. Consider the following procedure. getUser: t.procedure .meta({ openapi: { method: 'GET', path: '/user' } }) .input(z.object({ userId: z.string() })) .output(z.object({ id: z.string(), name: z.string() })) .query(({ input }) => { console.error(typeof req.input.userId); return { id: req.input, name: 'Bilbo' }; }), This is how it can be invoked. @trpc/server 👉 GET /getUser?input=${encodeURI(JSON.stringify({ userId: 'abc123' }))} trpc-openapi 👉 GET /user?userId=abc123 too high price to pay That's your decision to make.
gharchive/issue
2022-12-06T07:35:26
2025-04-01T06:39:12.921152
{ "authors": [ "elmeister", "jlalmes" ], "repo": "jlalmes/trpc-openapi", "url": "https://github.com/jlalmes/trpc-openapi/issues/196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
788965082
SSL support Is it possible to attach to a kafka instance with SSL enabled? For example, when I'm attaching to my cluster as a producer, I'll use the following: KafkaProducer(bootstrap_servers=kafkaServer, security_protocol="SSL", ssl_check_hostname=True) When attaching to the kafka cluster through this extension, I can specify the bootstrap server and port (443 in my case), but I can't seem to find a way to change the security protocol. Currently SSL is only enabled when using an authentication scheme, i.e. you have a username/password. So you need SSL but no auth? Currently SSL is only enabled when using an authentication scheme, i.e. you have a username/password. So you need SSL but no auth? Yes, currently we don't have any auth, just SSL. I tried to specify SASL/Plain as authentication but with a blank username password, but that doesn't seem to work. Yes, currently we don't have any auth, just SSL. I tried to specify SASL/Plain as authentication but with a blank username password, but that doesn't seem to work. Ok, so your use case seems quite simple but if we want to do SSL support right (and advertise we support SSL), we need to be able configure SSL certificate configuration, as per https://kafka.js.org/docs/configuration#ssl Currently, the ssl boolean is only set here: https://github.com/jlandersen/vscode-kafka/blob/a0257d0932263cc66bce73c5579a9d530f9655b0/src/client/client.ts#L325 Ok, so your use case seems quite simple but if we want to do SSL support right (and advertise we support SSL), we need to be able configure SSL certificate configuration, as per https://kafka.js.org/docs/configuration#ssl Currently, the ssl boolean is only set here: https://github.com/jlandersen/vscode-kafka/blob/a0257d0932263cc66bce73c5579a9d530f9655b0/src/client/client.ts#L325 Brilliant! That works well for my purposes, it connects fine and seems to be working well. Brilliant! That works well for my purposes, it connects fine and seems to be working well.
gharchive/issue
2021-01-19T12:02:03
2025-04-01T06:39:12.927371
{ "authors": [ "fbricon", "mike-england" ], "repo": "jlandersen/vscode-kafka", "url": "https://github.com/jlandersen/vscode-kafka/issues/84", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1861984934
Cannot see files in folders since Container Manager feature Current Behavior Hello Since the appearance of Container Manager instead of Docker package I noticed that the backup was constantly stuck at 0 files found (0mb) which is not normal. So I browsed through files and noticed that folders and files were not visible. The permissions are good (USER_ID=0 and GROUP_ID=0). The problem happened after a reboot following an image update. I noticed that by restarting the container from the command line, the problem was gone. Does anyone also have the problem? What to do to continue to benefit from the evolutions of Container Manager (simple update feature) without having to restart systematically on the command line. I specify that the bug is in no way related to the good work of our dear JLesage ;-) Expected Behavior No response Steps To Reproduce No response Environment OS: OS version: CPU: Docker version: Device model: Browser/OS: Container creation Container Manager package Container log no abnormal log Container inspect No response Anything else? No response I assume you are using a Synology NAS? Looks like the issue where /volume1 is mapped to /storage (see #310). Is it the case? Yes I use a Synology NAS but the issue not seems to be the same as #310 tab. I will try to make new tests later You can also share the output of docker inspect <container name> so we can have a complete picture of the container's configuration. I deleted the container and created a new one. Issue seems gone when container started from Container Manager Hello, Now I get 1,5 years to backup 4 TB. How to increase backup speed ? My ISP speed is about 700 mbps but crashplan backup is really slow. Yes, it is known that upload speed is slow. However, because of deduplication, a lot of data don't really have be uploaded. If you look at Tools->History, you can see the "Effective rate": the value can be very high and would typically improve over time. I had the same problem. In fact, when the docker container was created (a long time ago), the storage volume was not requested. Since the last update, not only is it requested, but if it hadn't been configured originally, it points to an empty virtual folder. I installed portainer to change the configuration and set storage bind to volume1. Since then, I've been able to browse storage, but for crashplan, it's not the same folder structure. As a result, the backup started from scratch. I hope that deduplication will play its part, otherwise I'm in for several months of backups.
gharchive/issue
2023-08-22T18:30:20
2025-04-01T06:39:12.950399
{ "authors": [ "Bagu", "jlesage", "walawa75" ], "repo": "jlesage/docker-crashplan-pro", "url": "https://github.com/jlesage/docker-crashplan-pro/issues/438", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
442215161
Use custom DB instance Hi, thanks a lot to share this ,it is really great. I'm wandering can we use our custom DB instance (like MySQL or Maria DB) instead of MySQL in container ,thanks. If you want to use your own DB instance, look at the original project: https://github.com/jc21/nginx-proxy-manager
gharchive/issue
2019-05-09T12:38:42
2025-04-01T06:39:12.952170
{ "authors": [ "KevinZhangt", "jlesage" ], "repo": "jlesage/docker-nginx-proxy-manager", "url": "https://github.com/jlesage/docker-nginx-proxy-manager/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
583876334
Let's Encrypt: Internal Error I used the hassio addon for quite some time now. Today I tried to migrate the proxy manager to my synology. For that I use the synology docker manager and this docker image. So I installed the latest version (1.7.0) and configured my port forwarding. The only issue I had was with the lets encrypt certificate. When I tried to request a certificate, I got an "Internal error": This is the content in the /config/log/letsencrypt/letsencrypt.log: 2020-03-18 17:02:07,145:DEBUG:certbot.main:certbot version: 0.30.2 2020-03-18 17:02:07,147:DEBUG:certbot.main:Arguments: ['--non-interactive', '--config', '/etc/letsencrypt.ini', '--cert-name', 'npm-15', '--agree-tos', '--email', 'mail@zoker.me', '--preferred-challenges', 'dns,http', '--webroot', '--domains', 'wg.zkr.io'] 2020-03-18 17:02:07,150:DEBUG:certbot.main:Discovered plugins: PluginsRegistry(PluginEntryPoint#manual,PluginEntryPoint#null,PluginEntryPoint#standalone,PluginEntryPoint#webroot) 2020-03-18 17:02:07,202:DEBUG:certbot.log:Root logging level set at 20 2020-03-18 17:02:07,204:INFO:certbot.log:Saving debug log to /config/log/letsencrypt/letsencrypt.log 2020-03-18 17:02:07,206:DEBUG:certbot.plugins.selection:Requested authenticator webroot and installer None 2020-03-18 17:02:07,220:DEBUG:certbot.plugins.selection:Single candidate plugin: * webroot Description: Place files in webroot directory Interfaces: IAuthenticator, IPlugin Entry point: webroot = certbot.plugins.webroot:Authenticator Initialized: <certbot.plugins.webroot.Authenticator object at 0x7fded3100a90> Prep: True 2020-03-18 17:02:07,223:DEBUG:certbot.plugins.selection:Selected authenticator <certbot.plugins.webroot.Authenticator object at 0x7fded3100a90> and installer None 2020-03-18 17:02:07,224:INFO:certbot.plugins.selection:Plugins selected: Authenticator webroot, Installer None 2020-03-18 17:02:07,320:DEBUG:acme.client:Sending GET request to https://acme-v02.api.letsencrypt.org/directory. 2020-03-18 17:02:07,326:DEBUG:urllib3.connectionpool:Starting new HTTPS connection (1): acme-v02.api.letsencrypt.org 2020-03-18 17:02:12,333:DEBUG:certbot.log:Exiting abnormally: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 141, in _new_conn (self.host, self.port), self.timeout, **extra_kw) File "/usr/lib/python3.6/site-packages/urllib3/util/connection.py", line 60, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "/usr/lib/python3.6/socket.py", line 745, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno -3] Try again During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 601, in urlopen chunked=chunked) File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 346, in _make_request self._validate_conn(conn) File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 850, in _validate_conn conn.connect() File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 284, in connect conn = self._new_conn() File "/usr/lib/python3.6/site-packages/urllib3/connection.py", line 150, in _new_conn self, "Failed to establish a new connection: %s" % e) urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/requests/adapters.py", line 445, in send timeout=timeout File "/usr/lib/python3.6/site-packages/urllib3/connectionpool.py", line 639, in urlopen _stacktrace=sys.exc_info()[2]) File "/usr/lib/python3.6/site-packages/urllib3/util/retry.py", line 388, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='acme-v02.api.letsencrypt.org', port=443): Max retries exceeded with url: /directory (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again',)) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/bin/certbot", line 11, in <module> load_entry_point('certbot==0.30.2', 'console_scripts', 'certbot')() File "/usr/lib/python3.6/site-packages/certbot/main.py", line 1364, in main return config.func(config, plugins) File "/usr/lib/python3.6/site-packages/certbot/main.py", line 1233, in certonly le_client = _init_le_client(config, auth, installer) File "/usr/lib/python3.6/site-packages/certbot/main.py", line 604, in _init_le_client acc, acme = _determine_account(config) File "/usr/lib/python3.6/site-packages/certbot/main.py", line 521, in _determine_account config, account_storage, tos_cb=_tos_cb) File "/usr/lib/python3.6/site-packages/certbot/client.py", line 181, in register acme = acme_from_config_key(config, key) File "/usr/lib/python3.6/site-packages/certbot/client.py", line 51, in acme_from_config_key return acme_client.BackwardsCompatibleClientV2(net, key, config.server) File "/usr/lib/python3.6/site-packages/acme/client.py", line 814, in __init__ directory = messages.Directory.from_json(net.get(server).json()) File "/usr/lib/python3.6/site-packages/acme/client.py", line 1152, in get self._send_request('GET', url, **kwargs), content_type=content_type) File "/usr/lib/python3.6/site-packages/acme/client.py", line 1101, in _send_request response = self.session.request(method, url, *args, **kwargs) File "/usr/lib/python3.6/site-packages/requests/sessions.py", line 512, in request resp = self.send(prep, **send_kwargs) File "/usr/lib/python3.6/site-packages/requests/sessions.py", line 622, in send r = adapter.send(request, **kwargs) File "/usr/lib/python3.6/site-packages/requests/adapters.py", line 513, in send raise ConnectionError(e, request=request) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='acme-v02.api.letsencrypt.org', port=443): Max retries exceeded with url: /directory (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7fded3046d30>: Failed to establish a new connection: [Errno -3] Try again',)) 2020-03-18 17:02:12,340:ERROR:certbot.log:An unexpected error occurred: Is this an issue with this image or with the base image or is my configuration wrong? Is this issue related? https://github.com/jc21/nginx-proxy-manager/issues/180#issuecomment-599278001 Thank you! Edit: A little bit more information: I also tried version 1.6.0, but it has the same issue. For both version: Connecting to my devices trough the manager using port 80 works fine. Connecting to them via 443 leads to a empty response. Looks like the container is not able to reach the LetEncrypt server. Can you try: docker exec <container name> ping acme-v02.api.letsencrypt.org Hmm yes, something does not seem right: When I try the same with my AdGuard container it works: AdGuard is running in host mode, while ProxyManager is running in bridge mode. But bridge mode should work as well, right? Do you have any idea, how I can fix that? Are you using AdGuard as your DNS server ? If yes. this may be the cause of the issue. Are you able to ping acme-v02.api.letsencrypt.org from your Synology ? Yes I'm using it as my DNS server. Yes the ping from synology itself works: Is this DNS server also used by you Synology? You can try to run the following command on your Synology: nslookup acme-v02.api.letsencrypt.org <DNS server IP>. Also, can you confirm that this DNS server used by the container: docker exec <container name> cat /etc/resolv.conf I am having the same internal error message. I use dnsimple and I do have a certificate on my account for my domain. Pinging the address you said above is successful. If your container accessible from the internet on port 80 ? You can check with https://www.yougetsignal.com/tools/open-ports/ Same issue here. I am able to ping with: docker exec ping acme-v02.api.letsencrypt.org No internal DNS beyond a redirect to an external PiHole server. I tried both requesting a new cert (internal error) and creating my own via a LetsEncrypt container. Both no dice. Make sure the container is reachable from the internet on port 80. You can check with https://www.yougetsignal.com/tools/open-ports/ Morning. Port shows as not reachable but both my router settings and the use of the proxy manager is present over port 80. If I use the IP:80 I go directly to where I set redirect to. 443 shows open and listening. Could I just direct through that? 80 is still working fine, It appears my isp blocks port 80, though I can go it via both directly through the IP and via my duckdns dyndns. If your ISP blocks port 80, then I guess your cannot use Nginx Proxy Manager with automatic certificate generation from LetsEncrypt. Port 80 is required. This is how LetsEncrypt validates that your are the owner of the DNS name. Odd given I can use port 80 for the hosts from Nginx Proxy Manager. I'll do some more research and report back on a fix for future reference.
gharchive/issue
2020-03-18T17:06:22
2025-04-01T06:39:12.966448
{ "authors": [ "TheZoker", "bookandrelease", "jlesage", "mdisieno" ], "repo": "jlesage/docker-nginx-proxy-manager", "url": "https://github.com/jlesage/docker-nginx-proxy-manager/issues/57", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
199320982
[WIP] A Proposal For Testing Things As we talked about in https://github.com/jlord/git-it-electron/pull/178#issuecomment-270530180, it'd be nice to be able to run the verify tests automagically. This PR is my attempt to tackle the problem, which required some refactoring of how the verify scripts work. Current Limitations A couple of issues that I've had to work through: verify scripts touch the DOM as part of rendering the results - I've attempted to move this code back up to the challenge script, but it's only done for one verify script so far (the repository tests). some verify scripts are written as if they are blocking but due to using exec they need to be asynchronous - we need a way to inspect the results, so I've introduced a callback that takes an array of results. Maybe you have other ideas in mind. How This Works We're using tape to run scripts (I'm more familiar with mocha, so I'm open to other ways to get the best out of the tape API), but here's a simple test: test('verifies the hello world repository', function (t) { var folder = helper.extractFixture('hello-world') verify(folder, function (result) { var expected = 'This is a Git repository!' var first = result[0] t.ok(first.result) t.equal(expected, first.message) t.end() }) }) For scenarios where we need the repository to be in a certain state, we can splat the folders onto disk under the tests/fixtures folder - and then extract them to a temporary folder when running the tests. To get this working, you can just rename the .git folder to _git and the repository state is then version-controllable. The other notable change here is that we have a callback to poke at the results from the verify script (because exec is asynchronous), to ensure we're getting the expected results. I'm using the same shape of "message is a string" and "result is a boolean" that the existing verify scripts use. To get this working in a test harness, I needed to move the code for addToList and challengeIncomplete/markChallengeCompleted out from the verify script to where the verify method is invoked. A partial implementation of this in challenge.js works for the repository script, but it definitely needs more work and testing as I might not be understanding it correctly. What's Left To Do? A rough list: [ ] get feedback on this approach [ ] sketch out the changes to test a path-less script, update challenge.js to handle this [ ] propose a way to test request invocations (nock? something else?) [ ] ??? verify scripts touch the DOM as part of rendering the results Yeah, this is something I've wanted to improve; to separate these concerns. Which is probably a big enough project in itself to be a different PR and a thing to figure out before this? I haven't gone through all this with a fine tooth comb but in general I feel like it's what I've been thinking! 👍 I have a couple things left to clean up, cause I don't think master is quite stable right now but then I'll pick up from what you've started here in moving out the DOM touching stuff and I'll get a PR going for that. ✨ :cool: Which is probably a big enough project in itself to be a different PR and a thing to figure out before this? Up to you - if you feel confident enough to do that refactoring before introducing tests then :thumbsup:
gharchive/pull-request
2017-01-07T00:08:40
2025-04-01T06:39:13.001096
{ "authors": [ "jlord", "shiftkey" ], "repo": "jlord/git-it-electron", "url": "https://github.com/jlord/git-it-electron/pull/180", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
79449040
View not being update with angular-cache with localStorage I'm building an Ionic app the uses angular-cache to leverage data persistence. This data is displayed in a ng-repeat list. Either by user interaction (in another view!) or new data being loaded in the background, I need the view to update to reflect this. When developing my app everything works perfectly, however, as soon as I start using angular-cache with localStorage, views stop being updated. When I turn off localStorage (but still use in memory caching from angular-cache) everything works as expected: the data updates and so is the DOM. I was able to reproduce the issue with this CodePen http://codepen.io/anon/pen/qdayMW. To demonstrate the issue you switch tabs (from home to about). If you look in the console you can see that new items are being added to the localStorage cache. However the homescreen doesn't update. If you then refresh the page (within the 10 second cache magAge) the page shows the correct list of items... However, if we comment out line 81, the caching is no longer persistent after a refresh, however the new items are added to the DOM every time the view changes. Note: I also posted this issue to the Ionic forum as I'm not sure which exactly is the root of the problem: http://forum.ionicframework.com/t/view-not-being-update-with-angular-cache-with-localstorage/25236 This must be something weird with ionic. I've been playing with your codepen, and as I click between the home and about tabs, the home controller code is only executed once at the very beginning, so while more items are being added to the cache, the resolve code that would get the latest items isn't run again. The latest ionic implemented view caching (which is causing your "bug"), which you have to disable: http://ionicframework.com/docs/nightly/api/directive/ionNavView/
gharchive/issue
2015-05-22T13:08:57
2025-04-01T06:39:13.013024
{ "authors": [ "danielcha", "jmdobry" ], "repo": "jmdobry/angular-cache", "url": "https://github.com/jmdobry/angular-cache/issues/181", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
909577463
[Bug] Kitsu anime tracking not working Device information Tachiyomi version: Latest stable and preview version Android version: 10 Device: Oneplus 5T The kitsu anime tracking only shows the manga's and not the anime. Checked it with myanimelist but that one does work, but it's not the site I use to keep my library updated. Would be awesome if it can be fixed. Keep up the great work! forgot to close. was fixed with https://github.com/jmir1/aniyomi/commit/9e2db50b77fdd99578eddc55d07cc343bacb049b.
gharchive/issue
2021-06-02T14:41:20
2025-04-01T06:39:13.017310
{ "authors": [ "Reacien", "jmir1" ], "repo": "jmir1/aniyomi", "url": "https://github.com/jmir1/aniyomi/issues/39", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1814461043
error loading model: unexpectedly reached end of file On a couple of models I am receiving this error: llama.cpp: loading model from /Users/REDACTED/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 error loading model: unexpectedly reached end of file llama_load_model_from_file: failed to load model This happens with a couple of the larger models: nous-hermes:latest llama2:13b If I do ollama pull against them, the manifests match up and it doesn't re-pull anything. Since this looks like docker under the hood, are the models corrupt? or? Any thoughts? FWIW, llama2:latest and wizard-vicuna:latest work fine. M2 Macbook Pro 32 Gigs of ram. Thanks @bkruger99, will check out why this is happening Let me know if you need any additional debugging data from my side. You'll have to tell me how to enable other than running server via cli :) Great! @bkruger99 is this on Mac? Thanks! Yes! Hardware: Model Name: MacBook Pro Model Identifier: Mac14,10 Model Number: Z174000EBLL/A Chip: Apple M2 Pro Total Number of Cores: 12 (8 performance and 4 efficiency) Memory: 32 GB OS: Ventura 13.4.1 (c) @bkruger99 can you run: sha2 -256 ~/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 Check to see that sha sum matches, and if it doesn't you can rm ~/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 and then re-pull the image. There's a fix that I think was just merged that will make certain the sha sum is verified correctly when you're pulling the layers. Yeah. there's something w/ manifest not verifying the sha256 when pulling. These two models did have a network interruption as the laptop went to sleep. ❯ shasum -a 256 sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 f2a1788633ddf3edef0ee4d9d4e93c399bfeeeb7363015d7c1b630ff268cdcf5 sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998 I re-pulled llama2:12b and it's happy, I'll do the same with the rest of 'em. The next version will check the SHAs; the re-pull is pretty tolerant of network interruptions, but wondering if the buffer wrote garbage onto the end of the partial file somehow. I haven't (yet) tested with sleeping the machine though, so that could have been the reason. I'm going to go ahead and close the issue. Feel free to re-open it though.
gharchive/issue
2023-07-20T17:15:59
2025-04-01T06:39:13.040101
{ "authors": [ "bkruger99", "jmorganca", "pdevine" ], "repo": "jmorganca/ollama", "url": "https://github.com/jmorganca/ollama/issues/143", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2768101
Raise ArgumentError when passing invalid keys to has_selector? It's a relatively common misconception that you can use :content as a key - e.g. has_selector?('h1', :content=>'wooo!') which would silently pass, even when 'wooo!' is nowhere to be found. This change should make that error much more obvious. Related issues : https://github.com/jnicklas/capybara/issues/525 It would be nice to raise an error on passing a block in - eg @session.should have_selector("//p") do |p_tag| p_tag.should have_selector("//a") end which seems to be a common error on people switching to capybara via webrat. However, that's a bit more tricky and would probably need fixing at the rspec-matcher level, since the block doesn't make it as far as has_selector? and gets silently discarded. ValidKeys on Selector could be replaced with something like this : diff --git a/lib/capybara/selector.rb b/lib/capybara/selector.rb index 1bff91d..0b85a04 100644 --- a/lib/capybara/selector.rb +++ b/lib/capybara/selector.rb @@ -15,6 +15,10 @@ module Capybara def remove(name) all.delete(name.to_sym) end + + def valid_keys + all.map{|selector| selector.custom_filters.keys}.flatten.uniq + end end def initialize(name, &block) ? I'm not sure what you mean about the implementation for checking query keys, though. There's 3 places that keys passed to has_selector?() are used - node/matchers.rb, query.rb, and selector.rb. Whichever one is responsible for checking the keys will have to also check the keys of the other 2. I could move assert_valid_options to a class method to Query if you like? Closing this pull request, we already have an open issue for this in #525, and I don't like this implementation
gharchive/issue
2012-01-09T12:20:18
2025-04-01T06:39:13.102597
{ "authors": [ "jdelStrother", "jnicklas" ], "repo": "jnicklas/capybara", "url": "https://github.com/jnicklas/capybara/issues/602", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1512494687
🛑 MegaConstruct is down In 8d9ab08, MegaConstruct (https://www.megaconstruct.ro) was down: HTTP code: 0 Response time: 0 ms Resolved: MegaConstruct is back up in edc67a4.
gharchive/issue
2022-12-28T07:27:55
2025-04-01T06:39:13.109536
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/1865", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2025303728
🛑 CEI Sales Wiki is down In 8230d21, CEI Sales Wiki (https://wiki.ceisales.com/login) was down: HTTP code: 0 Response time: 0 ms Resolved: CEI Sales Wiki is back up in 2b0ac7d after 14 minutes.
gharchive/issue
2023-12-05T05:13:31
2025-04-01T06:39:13.111958
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/2883", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2038683456
🛑 Laravel Enso is down In c645c42, Laravel Enso (https://laravel-enso.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Laravel Enso is back up in fa7077b after 22 minutes.
gharchive/issue
2023-12-12T23:09:57
2025-04-01T06:39:13.114576
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/2892", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2471341990
🛑 CEI Sales App is down In 83575f8, CEI Sales App (https://app.ceisales.com/login) was down: HTTP code: 0 Response time: 0 ms Resolved: CEI Sales App is back up in 2091cef after 20 minutes.
gharchive/issue
2024-08-17T05:05:41
2025-04-01T06:39:13.116945
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/3630", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1253103493
🛑 Laravel Enso is down In 8a5476a, Laravel Enso (https://laravel-enso.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Laravel Enso is back up in 654d5f1.
gharchive/issue
2022-05-30T20:38:24
2025-04-01T06:39:13.119241
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/719", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
860806476
[Feature Request] Change blue loading and badge accent colors in Midnight Dusk theme to something similar to J2K Midnight Dusk? Why/User Benefit/User Problem Improves the theme by keeping the color scheme. What/Requirements The blue accent (and possibly black background on loading) would be changed to better fit the Midnight Dusk theme. SY Midnight Dusk J2K Midnight Dusk Why/User Benefit/User Problem Improves the theme by keeping the color scheme. What/Requirements The blue accent would be changed to better fit the Midnight Dusk theme. Someone can do that for the other themes too and why the text in black-red are red instead of white ? I think its better to just replace everything that blue to red and change the colour of the download badge to one that look good for each theme Already addressed by main tachi
gharchive/issue
2021-04-19T00:31:09
2025-04-01T06:39:13.151373
{ "authors": [ "CaptainHalflife", "ImRyley", "jobobby04" ], "repo": "jobobby04/TachiyomiSY", "url": "https://github.com/jobobby04/TachiyomiSY/issues/277", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
157378264
async requests Is any chance an option to use async requests? like asp.net core rc2 standard template public async Task<IActionResult> Index() { return View(await _context.Reason.ToListAsync()); } _context.Reason.ToPagedList(0, DefaultPageSize) does not have ToListAsync method honestly PagedList is a completely separate thing from the pagertaghelper, pagertaghelper does not depend on PagedList and really the only reason pagedlist exists in this library is because it was based on Martijn Boland's MVCPaging project and it had a pagedlist which I copied. PagedList is kind of a hack for when you are retrieving more than a page of data but want to do paging against the result set. I don't use PagedList at all myself and don't really care to try to support it or update it. There are other implementations of PagedList out there that anyone could also use, so it is completely independent from the pagertaghelper. From a brief search I found this pagedlist which has async support, but not sure it has a compatible nuget for aspnetcore rc2. But you can search and maybe find a better compatible one, I did not spend much time looking. Or if anyone is interested in implementing that and submitting a pull request I'm open to that as well. I do async methods to get data in my own use of pagertaghelper, and I just return one page of data at a time from the database so I don't need any pagedlist. maybe could do it like this: public async Task<IActionResult> Index() { var data = await _context.Reason.ToListAsync(); return View(data.ToPagedList(0, DefaultPageSize)); } but again that looks like it returns more than a page of data and then uses pagedlist to page over it which is what PagedList is for but also why PagedList is a hack compared to just return a page of data from the database using .Skip .Take etc only thing I've ever used PagedList for is the demo pages because it was convenient to make a demo without any database, and much of the demo code was also copied from Martijn's project and that is how PagedList ended up in this project. I never use that PagedList in my own projects, I only use the taghelpers hm, sorry I a bit confused here. if you dont use PagedList what you use in your Pagination project? I mean then you retrieve data from DB. you can see an example here a controller which has methods for user admin and paging of users with search, and the corresponding partial view here as I said the use of PagedList in the pager demo pages was mainly because I copied existing demo pages and that makes it easy to have demo without any db, you just make a big list of data and use pagedlist to page over it. but again PagedList is only for if you already retrieved (or created as in the demo code) more than a page of data and now you want to page over that data. It is much better to not use PagedList and only retrieve data one page at a time from the db. PageTagHelper doesn't know about PagedList at all. not that in the example I linked, you will see model.Paging.* properties being set, in that model Paging is an instance of PaginationSettings it is simply a model to pass settings on to the pagertaghelper this makes me think I should get rid of PagedList and not use it in the demos because it makes people think they need to use that when in fact it is best not to fyi, I've reworked the demo app and removed all use of PagedList from the demo so as not to encourage use of that Many tnks. checking
gharchive/issue
2016-05-29T09:46:12
2025-04-01T06:39:13.165262
{ "authors": [ "dvdobrovolskiy", "joeaudette" ], "repo": "joeaudette/cloudscribe.Web.Pagination", "url": "https://github.com/joeaudette/cloudscribe.Web.Pagination/issues/12", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
301313799
Multiple connections to the same instance with different DBs not working Hello, I'm try to run the docker container with the REDIS_HOSTS env variable witch looks like this: REDIS_HOSTS='dev:redis:6379:0,test:redis:6379:10' but it seems that the given database is ignored, since the UI shows me for both connections the database 0. This leads to an instable system, since after you click on one of the two connections the the view stucks on "loading ...." instead of showing the connection parameters. I also tried to configure multiple connections via REDIS_HOSTS environment variable. No luck for me too. The undefined entries at the bottom are also a bit irritating. @hirnschmalz You are right, i've seen this too but i forgot to mentioned the undifined fields @tburschka The latest image solved this issue @joeferner This issue is back in the latest docker image. 👎 Seems to be an issue with the with the change of the namespace @awearsolutions/redis-commander. It's a bit annoying to create a PR to fix a problem which is back one version later. Should be working with latest github/docker versions. At least we use it without any problem... Can issue be closed? This problem should be fixed in current version. If you think its not feel free to reopen issue with updated informations based on latest version.
gharchive/issue
2018-03-01T08:17:05
2025-04-01T06:39:13.176382
{ "authors": [ "hirnschmalz", "sseide", "tburschka" ], "repo": "joeferner/redis-commander", "url": "https://github.com/joeferner/redis-commander/issues/236", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
573023714
Docker Container image for PytaskIO Make a PyTaskIO Docker image available to download from the Docker registry. Also update the README with the docker registry install details.
gharchive/issue
2020-02-28T21:49:43
2025-04-01T06:39:13.182106
{ "authors": [ "joegasewicz" ], "repo": "joegasewicz/pytask-io", "url": "https://github.com/joegasewicz/pytask-io/issues/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1090171830
v0.30.1 won't publish - issue with dependency When running cargo publish, the pin-project fails to build, because of some issue with a macro: error[E0659]: `parse_quote_spanned` is ambiguous (`macro_rules` vs non-`macro_rules` from other module) --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pin_project/derive.rs:859:67 | 859 | proj_generics.make_where_clause().predicates.push(parse_quote_spanned! { span => | ^^^^^^^^^^^^^^^^^^^ ambiguous name | note: `parse_quote_spanned` could refer to the macro defined here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/utils.rs:22:1 | 22 | / macro_rules! parse_quote_spanned { 23 | | ($span:expr => $($tt:tt)*) => { 24 | | syn::parse2(quote::quote_spanned!($span => $($tt)*)).unwrap_or_else(|e| panic!("{}", e)) 25 | | }; 26 | | } | |_^ note: `parse_quote_spanned` could also refer to the macro imported here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pin_project/derive.rs:7:5 | 7 | *, | ^ = help: use `self::parse_quote_spanned` to refer to this macro unambiguously error[E0659]: `parse_quote_spanned` is ambiguous (`macro_rules` vs non-`macro_rules` from other module) --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pinned_drop.rs:108:21 | 108 | *path = parse_quote_spanned! { path.span() => | ^^^^^^^^^^^^^^^^^^^ ambiguous name | note: `parse_quote_spanned` could refer to the macro defined here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/utils.rs:22:1 | 22 | / macro_rules! parse_quote_spanned { 23 | | ($span:expr => $($tt:tt)*) => { 24 | | syn::parse2(quote::quote_spanned!($span => $($tt)*)).unwrap_or_else(|e| panic!("{}", e)) 25 | | }; 26 | | } | |_^ note: `parse_quote_spanned` could also refer to the macro imported here --> /Users/joep/.cargo/registry/src/github.com-1ecc6299db9ec823/pin-project-internal-0.4.23/src/pinned_drop.rs:3:50 | 3 | use syn::{spanned::Spanned, visit_mut::VisitMut, *}; | ^ = help: use `self::parse_quote_spanned` to refer to this macro unambiguously For more information about this error, try `rustc --explain E0659`. error: could not compile `pin-project-internal` due to 2 previous errors warning: build failed, waiting for other jobs to finish... error: failed to verify package tarball This seems to be caused by this: https://github.com/taiki-e/pin-project/blob/eedd0885afc4bebad728f0adb1acffaaaed0d31a/pin-project-internal/src/pin_project/derive.rs Which seems to have been changed just days ago: https://github.com/taiki-e/pin-project/commit/ba7d5fd13f81092e64c04a30fd0dd7dd70cb87a8 And was mentioned in this issue: https://github.com/taiki-e/pin-project/issues/337 cargo update did not work. cargo publish --lockfile dit not work. I don't know what I can try next. Removing Cargo.lockl and running publish again worked.
gharchive/issue
2021-12-28T23:15:57
2025-04-01T06:39:13.190726
{ "authors": [ "joepio" ], "repo": "joepio/atomic-data-rust", "url": "https://github.com/joepio/atomic-data-rust/issues/264", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
181796481
Getting this error with my rogue... Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value) Ok. Will look into it (will be later this evening as I'm out and about at present :)) -- Joe Vaughan On 8 October 2016 at 03:41:10, lkj61 (notifications@github.com) wrote: Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value) — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/joev/SVUI-Temp/issues/80, or mute the thread https://github.com/notifications/unsubscribe-auth/AATrWRguVwSt-gouFgw6WMK71PE4WqrFks5qxvS2gaJpZM4KRlUg . Thanks! Much appreciate all the work you are doing for us fans of SVUI. 😀 On Saturday, October 8, 2016, Joe Vaughan notifications@github.com wrote: Ok. Will look into it (will be later this evening as I'm out and about at present :)) -- Joe Vaughan On 8 October 2016 at 03:41:10, lkj61 (notifications@github.com <javascript:_e(%7B%7D,'cvml','notifications@github.com');>) wrote: Frames [Load]: @(10/07/16 20:39:20) - ...ace\AddOns\SVUI_UnitFrames\class_resources\rogue.lua:124: attempt to index field '?' (a nil value) — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/joev/SVUI-Temp/issues/80, or mute the thread <https://github.com/notifications/unsubscribe-auth/AATrWRguVwSt- gouFgw6WMK71PE4WqrFks5qxvS2gaJpZM4KRlUg> . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/joev/SVUI-Temp/issues/80#issuecomment-252416152, or mute the thread https://github.com/notifications/unsubscribe-auth/AJEkVZg5YQOMoY-j_UDeUkY3O-EwXyhGks5qx2q4gaJpZM4KRlUg .
gharchive/issue
2016-10-08T01:41:10
2025-04-01T06:39:13.202385
{ "authors": [ "joev", "lkj61" ], "repo": "joev/SVUI-Temp", "url": "https://github.com/joev/SVUI-Temp/issues/80", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1795017583
🛑 dr.veoj.me (my gaming persona) is down In 777b0d1, dr.veoj.me (my gaming persona) (https://dr.veoj.me) was down: HTTP code: 0 Response time: 0 ms Resolved: dr.veoj.me (my gaming persona) is back up in 71fff75.
gharchive/issue
2023-07-08T17:01:27
2025-04-01T06:39:13.205727
{ "authors": [ "joevaughan" ], "repo": "joevaughan/upptime", "url": "https://github.com/joevaughan/upptime/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
747964988
[BUG] No up-to-date light and temperature values in Homebridge anymore Describe the bug Temperature and light don't get updated in Homebridge and Home app anymore, although http://<PIR-IP>/api/v1/sensors shows up-to-date temperature and light values. To Reproduce Steps to reproduce the behavior: In the Homebridge GUI navigate to Accessories Light will have an old value Open http://<PIR-IP>/api/v1/sensors in browser Light value will be different, but correct Expected behavior The light and temperature values in Homebridge should have the same values as the ones provided by the API. Plugin environment: myStrom PIR Firmware Version: 3.8.2 Homebridge Server: Raspberry Pi 4 Plugin Version: 1.8.0 Node.js Version: 14.15.1 Homebridge Version: 1.1.6 Additional context The issue appears only after some time. Upon re-starting Homebridge it works again, but will stop working after a couple of hours. The change related to reachability (73298a58f2cfbf3dff2a12e146ab1d137b2aac64) introduced in v1.8.0 seems to break quite a few things, but only after some time running. 🤦 A downgrade to v1.7.1 should work around this while I work on a fix. I've released homebridge-dingz@1.8.3-nightly.0 (homebridge-dingz@nightly) and am running it now. If you want to help with testing, feel free to install it and report back. the nightly reverts some of the changes that might have caused some of the problems encountered here and elsewhere 🤕🤒 I've installed 1.8.3 nightly and will report back 👍 In the meantime I've worked on a v2.0.0-nightly of the plug-in. If you reinstall the nightly w/ homebridge-dingz@nightly you should be upgraded to this version. The functionality remains largely the same but the code is less cluttered and better structured so should generally run better. There might be bugs though but you can always downgrade to v1.8.3-nightly (although I think the newest nightlies are much more robust). I've had it running over longer periods of time now on my production dingz. So far I could not observe any problems with temperature/brightness values not updating. (The easiest way to test if it still works after a few hours is to use your phone's flashlight to illuminate or your hand to darken the illumination: the LUX value should change significantly 😄) It's apparently still not fixed though. 🤦 😞 But I now have a clue where it happened (unreleased mutex locks on the request queue) in the code. :octocat: https://github.com/johannrichard/homebridge-dingz/blob/dee30b26ab348adb4db62c7f90f47c4ae107b684/src/dingzAccessory.ts#L1585-L1597 So far v2.0.0-nightly is working for me, but it has been running less than 24 hours.
gharchive/issue
2020-11-21T07:10:21
2025-04-01T06:39:13.255104
{ "authors": [ "johannrichard", "qx54" ], "repo": "johannrichard/homebridge-dingz", "url": "https://github.com/johannrichard/homebridge-dingz/issues/123", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
84032138
Warning in Google Chrome console when generating pdf/excel document There is a warning in Google Chrome console when generating document: Resource interpreted as Document but transferred with MIME type application/pdf. Request Headers: Accept:text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,/;q=0.8 Accept-Encoding:gzip, deflate Accept-Language:en-US,en;q=0.8,ru;q=0.6 Cache-Control:max-age=0 Connection:keep-alive Content-Length:305 Content-Type:application/x-www-form-urlencoded User-Agent:Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/43.0.2357.81 Safari/537.36 Response Headers: Content-Disposition:attachment; filename=report.pdf Content-Type:application/pdf;charset=windows-1251 Date:Tue, 02 Jun 2015 13:43:47 GMT Expires:Thu, 02 Jun 2016 06:50:44 GMT Server:Apache-Coyote/1.1 Transfer-Encoding:chunked Yeah I get that too, it still works though so I think its okay IMO I get this error also for PDF's. But it will not even try to download. I change the file to download to something else (a non-pdf) but get the error and it still downloads. Also do you know of a way to suppress this error. I would like to not put the download URL out in the open. This isn't a warning in chrome its logged at info. If you are worried about your download URL being logged that's not good because anyone can view it in the chrome debugger or by monitoring http traffic. Would suggest using single use url/token/cookie if you are concerned about URL re-use. The web is not inherently secure: http://johnculviner.com/dont-trust-the-client/ understandable that there will always be a way to view the url, but I just dont want to hand it to them, would rather make them work just a little. If you have any thoughts as to suppress it that would be great. But there was still the issue of pdf's not downloading. do you happen to get this in your last push of the script? On your dont trust the client. Believe me I do not really. I treat the client as if they have some knowledge on web development so I close those holes when I can. But then I am not a security expert and cannot get everything. Though that would be nice. BTW I am in chrome.
gharchive/issue
2015-06-02T13:49:44
2025-04-01T06:39:13.277595
{ "authors": [ "johnculviner", "mike192020", "shustinho" ], "repo": "johnculviner/jquery.fileDownload", "url": "https://github.com/johnculviner/jquery.fileDownload/issues/98", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
135531609
Errors Installing I have Python2.7.11 for Windows x86/x64 installed on Windows 7 x64. When I run npm install lite-server --save-dev, I get the following errors: TypeError: Request path contains unescaped characters. And a few 404 status codes downloading 32-bit node.libs. anyone else on windows able to test this? moving to https://github.com/johnpapa/lite-server/issues/54
gharchive/issue
2016-02-22T20:16:07
2025-04-01T06:39:13.283145
{ "authors": [ "billfranklin", "johnpapa" ], "repo": "johnpapa/lite-server", "url": "https://github.com/johnpapa/lite-server/issues/39", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
132752012
feat(bin): added run time option to config file path and name Hey, It would be really good to have an option for custom path to configuration especially if you hate dozen of files in your project root folder (like me) or have to run different configuration for different needs Thanks in advance Regards, Dima @johnpapa please review i like the idea. @cgmartin is this how you would like to do it? i prefer yargs myself. This looks good to me. I like the fact that it matches names with the browsersync --config option. Yargs is what I'm most comfortable with, but minimist is fine. There are several popular args modules out there that all score pretty high. What we're doing here is pretty simple/straightforward in terms of arg options. Easy to replace if need be. @johnpapa good to merge? Thanks for the PR @valorkin ! i havent had a chance to test it ... have you Chris? If so, Im cool with it as long as it ups the version number by a minor. Tested it, works fine. Merging and bumping minor version. Since it changes the API (not breaking), Shouldn't this be a minor bump to 2.1.0 ? (not a patch) i bumped it to 2.1.0 and published yes, minor, I fubar. Thanks for fix no worries. thanks for testing
gharchive/pull-request
2016-02-10T16:38:38
2025-04-01T06:39:13.287994
{ "authors": [ "cgmartin", "johnpapa", "valorkin" ], "repo": "johnpapa/lite-server", "url": "https://github.com/johnpapa/lite-server/pull/32", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1190736386
Disable automatic quotation marks for attrs I would like to an option in settings to disable automatic quotation marks - we use ' but it keeps putting " and it just grinds our gears :) Duplicate of #840 and #917. Try it with HTML: Auto Create Quotes disabled.
gharchive/issue
2022-04-02T20:50:42
2025-04-01T06:39:13.310288
{ "authors": [ "JTinkers", "yaegassy" ], "repo": "johnsoncodehk/volar", "url": "https://github.com/johnsoncodehk/volar/issues/1139", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1355008263
Type narrowing not working in templates with union types v-if is not narrowing union types. <template> <span v-if="user.id">({{ user.id }})</span> </template> <script setup lang="ts"> type User = { id: number email: string } let user: string | User </script> Property 'id' does not exist on type 'string | User' See demo on stackblitz here. That's just how TS works here, here's the same code on TS playground, raising the same errors: https://www.typescriptlang.org/play?#code/C4TwDgpgBAqgzhATlAvFA3gKClAlgEwC4oA7AVwFsAjJbKCCgQ1wBti5hFcSBzTAX0yYWEYFDIJE7Ttx5QAPrEmoMeIlADkBDVEGZcAMygAKCUgB0BAJQY6AYwD2JOA5HmWDnqcmX8VgUA This, however, would work: type User = { id: number email: string } let user: string | User = { id: 'id' } if (isObj(user)) { console.log(user.id) } function isObj(obj: any): obj is Record<string, any> { return typeof obj === 'object' } https://www.typescriptlang.org/play?#code/C4TwDgpgBAqgzhATlAvFA3gKClAlgEwC4oA7AVwFsAjJbKCCgQ1wBti5hFcSBzTAX0yYWEYFDIJE7Ttx5QAPrEmoMeIlADkBDVEGZcAMygAKXHADyVAFbGJSAJT2MdAMYB7EnDciAdCzc8tpI+BPYCQgZkJC7AuB54FtbGbtbEjCQg9sQpVglQAEoQ7oj4ADwcXLwANFDpIAB8zjiIomSIJFCgkG5GOagoaBo5RcAaAkA Try typeof user === 'object' or 'id' in user.
gharchive/issue
2022-08-30T00:19:10
2025-04-01T06:39:13.314089
{ "authors": [ "LinusBorg", "austin-agronick", "johnsoncodehk" ], "repo": "johnsoncodehk/volar", "url": "https://github.com/johnsoncodehk/volar/issues/1788", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1081165871
Absolute import i've seen that there has been an issue marked as resolved but it might have broke again because it doesn't seems to work for me. when volar auto-import a component, it's always using the relative path my config : tsconfig.json { "compilerOptions": { "baseUrl": ".", "esModuleInterop": true, "experimentalDecorators": true, "jsx": "preserve", "lib": ["esnext", "dom"], "module": "esnext", "moduleResolution": "node", "paths": { "@/*": ["src/*"], "~/*": ["src/components/*"] }, "resolveJsonModule": true, "sourceMap": true, "strict": true, "target": "esnext", "useDefineForClassFields": false, "types": ["quasar"] }, "exclude": ["node_modules"], "include": ["src/**/*.ts", "src/**/*.d.ts", "src/**/*.tsx", "src/**/*.vue"] } vite.config.ts import { quasar, transformAssetUrls } from '@quasar/vite-plugin'; import vue from '@vitejs/plugin-vue'; import { resolve } from 'path'; import { defineConfig } from 'vite'; // https://vitejs.dev/config/ export default defineConfig({ server: { fs: { allow: ['.'], }, }, resolve: { alias: [ { find: '@', replacement: resolve(__dirname, './src'), }, { find: '~', replacement: resolve(__dirname, './src/components'), }, ], }, plugins: [ vue({ template: { transformAssetUrls }, }), quasar({ autoImportComponentCase: 'pascal', sassVariables: 'src/assets/scss/variable.scss', }), ], }); Versions : "typescript": "^4.3.2", "vite": "^2.5.4", "vue-tsc": "^0.29.0", "Volar extension": "v0.29.8" You can config "typescript.preferences.importModuleSpecifier": "non-relative" in vscode. that already what i did, both in user and workspace settings, and it's still not importing using non-relative path, with or without the built-in typescript extension enabled I have the same problem, also type hinting is not working in VS Code when using an absolute import.
gharchive/issue
2021-12-15T15:19:11
2025-04-01T06:39:13.317865
{ "authors": [ "Zenthae", "johnsoncodehk", "wendyprogramic" ], "repo": "johnsoncodehk/volar", "url": "https://github.com/johnsoncodehk/volar/issues/793", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1112931625
Update balance UI Updates the balance text a bit to more closely resemble the screens in Figma and the Bitcoin Design guidelines on units and symbols and on protecting a wallet. Basically, I turned on some font features according to the Figma designs and added a better symbol for when the balance is hidden. It's not perfect yet, I also want to add: The possibility to cycle through states (BTC, sats, hidden) by clicking on a balance. For that we need some global state and I didn't want to touch that quite yet. Formatting based on the user’s locale. Currently I hardcoded the BTC symbol to be prefixed to the number (as it is done in the U.S. with USD amounts. However, I didn't know what to do with the sats symbol as it looked strange when prefixed to a number so I kept it as a postfix. Open for suggestions on this one! For a first version it should be fine though. Let me know what you think! Not sure how we want to organize custom CSS going forward. I used a CSS Module but am open to do it differently or just add everything to index.css for now. 📸 See also https://github.com/joinmarket-webui/joinmarket-webui/issues/19. Wonderful! Regarding the sat symbol: so far most interfaces still use "sat" or "sats," i.e. no symbol at all. Thunderhub—which looks quite neat in general I have to say—uses the sat symbol, and simply puts it at the end using a slightly dimmer font color: Most other interfaces, including bitcoin and lightning wallets, simply use "sat" or "sats" either before or after the number (usually after). 1ML: LightningNetwork.plus: Updated to use the sat symbol (resolves #19). The bitcoin design guidelines don't really use a specific sat symbol so maybe we should let the user pick if they want a symbol or just "sat(s)" in the upcoming settings page. Maybe @GBKS has some input on that? For now, I'd say let's keep it simple and just go with the symbol. I also added digit groups using Intl.NumberFormat as suggested by the design guides. For now I hardcoded the formatting to the en-US locale but we could do more location specific things in the future. Regarding browser compatibility of Intl.NumberFormat: It should be available on all somewhat modern browsers. Does someone have any experience with that? And a more broad question: What browser compatibility are we targeting in general? Great! Let's keep it as is for now. We can do some polishing for v4 later on, I'm sure we can improve things here and there. Different font color for the symbol would also be easier on the eyes, for example.
gharchive/pull-request
2022-01-24T17:22:05
2025-04-01T06:39:13.346183
{ "authors": [ "dergigi", "dnlggr" ], "repo": "joinmarket-webui/joinmarket-webui", "url": "https://github.com/joinmarket-webui/joinmarket-webui/pull/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
134876078
Add support for nullable property Other way for #21 WDYT @sagikazarmark ? How do you know if a multiple type is nullable? It's done in the ObjectGuesser I don't see any change in denormalization. You use isset in denormalization, but that won't work with null values. Right, should be good now Looks good to me. Although it lacks tests, doesn't it? Yes will add a better test system following this Cool.
gharchive/pull-request
2016-02-19T14:16:48
2025-04-01T06:39:13.375714
{ "authors": [ "joelwurtz", "sagikazarmark" ], "repo": "jolicode/jane", "url": "https://github.com/jolicode/jane/pull/23", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1464378660
[MCU Tray] MKS Skipr Made an MKS Skipr tray for my personal use, but wanted to share with the community. I hope this can make it into the project. MCU Tray - MKS Skipr - Heatserts.zip I made a total hash of merging this PR to the right branch, so I just cherry-picked the commits to 0.9.9-unstable instead. Let me know how the fit goes. Thanks!
gharchive/pull-request
2022-11-25T10:59:35
2025-04-01T06:39:13.392741
{ "authors": [ "Killajoedotcom", "jon-harper" ], "repo": "jon-harper/OmniBox", "url": "https://github.com/jon-harper/OmniBox/pull/71", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
299690929
TLS v1.2 for KitKat (and below?) With GitHub having removed support for TLS v1.0 and v1.1, ForkHub no longer works, it gives an error toast (Loading account & organizations failed). https://stackoverflow.com/questions/28943660/how-to-enable-tls-1-2-support-in-an-android-application-running-on-android-4-1 shows how to enable TLS v1.2 support for pre-Lollipop. There must be something working, since I can access GitHub on both Chrome Mobile and Lightning Browser. In the meantime, if anyone can direct me to how to get the "News" look for an arbitrary repo via web browser, I'd appreciate it. The closest I can find is Insights, but that doesn't list individual comments. square/okhttp#2372 may be instructive here. Noting also, this affects the version of Forkhub found in the Blackberry World App store, which is an Android app, I think maybe v1.2.0 (tagged release of apk), and the Android runtime provided in BB10 is 4.x era. It was a sad day when I opened the app and found that I couldn't get in. I will give a compliment, and say this is THE best performing Android app in terms of speed that I've found on the Blackberry BB10 platform. I'm willing to dig in here and help find a work around, maybe submit a pull request, since I don't have the option to upgrade my Android runtime, and find Forkhub to be essential to my daily work and my ability to be away from my desk but deal with the number of projects and companies I work with that have their stuff on GH.
gharchive/issue
2018-02-23T12:07:02
2025-04-01T06:39:13.397455
{ "authors": [ "Matrixcoffee", "goldfndr", "truedat101" ], "repo": "jonan/ForkHub", "url": "https://github.com/jonan/ForkHub/issues/345", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1297756534
same scoreboard for all players.... sry but a new problem. I have made a scoreboard with placeholders and for me it works. BUT my friend joined and has MY scoreboard, my name in scoreboard, my ranks, my money....not his own. sry this is a discord screenshot Ok, I have released a new version. Could you check out if the issue is still present there? no the bug isn fixed. NOW a new player joined the game, he has his own scoreboard but NOW I see his scoreboard iam xinaru he is Zeus or must i delete the old config? okay the bug isn fixed i have delete the old config and add the lines new. but the scoreboard dont show the own scoreboard always. Ok, I believe it has to do with the order the players join. I will have another look at it Alright, I have updated some logic concerning the buffering of scoreboards. In addition, this version will log if placeholder usage is detected. You should not need to delete the config file. If the issue persists, please check in the server logs if it is says "Detected use of placeholders" or "Detected no use of placeholders" EasyScoreboard-2.5.3.jar.zip sry but the problem ist there with the new version too. 08.07 17:27:59 [Server] Server thread/ERROR [Sponge]: The Scheduler tried to run the task de_yottaflops_easyscoreboard-S-3 owned by Plugin{id=de_yottaflops_easyscoreboard, name=EasyScoreboards, version=2.5.3, description=A plugin to easily create scoreboards for lobbys, etc., authors=[YottaFLOPS], source=./mods/EasyScoreboard-2.5.3.jar}, but an error occured. now the plugin is crashed Ok, this was probably due to another change I made. I apologize. Try with this version, if you could. EasyScoreboard-2.5.3.jar.zip ähm...i use 2.5.3? you have uplöoad this above? this version i use and the error comes It should be a new build with the same version number ah okay i try it okay tried. scoreboard there but my fried has MY scoreboard 08.07 19:52:04 [Server] Server thread/INFO [EasyScoreboard: Config]: Detected use of placeholders I believe that I have finally fixed the problem and in my testing it worked. Let me know if the problem persists for you. EasyScoreboard-2.5.3-2.jar.zip This time it looks like it's finally going to work. Just a friend and I on the server, everyone had their own sb. we are both separately leased and jointed it was still each his own sb. the only thing i forgot to test is the /esb reload command. Amazing. Thanks for your help debugging this issue and your patience. I have created a new release here but it should be identical (apart from the version number) to the .jar posted above. Thanks again
gharchive/issue
2022-07-07T16:45:14
2025-04-01T06:39:13.409051
{ "authors": [ "XiNaru", "jonas-peeters" ], "repo": "jonas-peeters/EasyScoreboards", "url": "https://github.com/jonas-peeters/EasyScoreboards/issues/47", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
635703781
Serialize Unit variant to/form string Closes #115 Todo [ ] Add test case bors r+
gharchive/pull-request
2020-06-09T19:32:23
2025-04-01T06:39:13.410451
{ "authors": [ "jonasbb" ], "repo": "jonasbb/serde_with", "url": "https://github.com/jonasbb/serde_with/pull/116", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2408403753
Navigating to New Product Details Page with Different Product ID Fails to Load New Data Question I am developing an e-commerce app using Flutter with GetX for state management. I have a ProductDetailsView that displays product details using a named route and arguments. In the ProductDetailsController, I load product details based on the product ID passed through arguments. Here is my controller code: import 'package:e_commerce/api/send_request.dart'; import 'package:e_commerce/app/data/color_collections.dart'; import 'package:e_commerce/app/widgets/theme_mode.dart'; import 'package:get/get.dart'; import 'package:webview_flutter/webview_flutter.dart'; class ProductDetailsController extends GetxController { WebViewController? webViewController; var webViewHeight = 100.0.obs; var productDetailsLoading = true.obs; var productDetails = <String, dynamic>{}.obs; var productId = 0.obs; @override void onInit() { super.onInit(); productId.value = Get.arguments['product_id']; loadPage(); } @override void onClose() { super.onClose(); productId.close(); } void loadPage() async { productDetailsLoading.value = true; var resp = await SendRequest("/Products/Manager/CRUD/GetProductDetails", data: { "product_id": productId.value, }); productDetails.value = resp["product_details"]; productDetailsLoading.value = false; if (webViewController == null) { webViewController = WebViewController() ..setJavaScriptMode(JavaScriptMode.unrestricted) ..setBackgroundColor(isDarkMode() ? DynamicColorsCollection.backgroundColorDark : DynamicColorsCollection.backgroundColorLight) ..setNavigationDelegate( NavigationDelegate( onProgress: (int progress) {}, onPageStarted: (String url) {}, onPageFinished: (String url) async { final result = await webViewController?.runJavaScriptReturningResult( "document.documentElement.scrollHeight.toString();"); String heightStr = result as String; double height = double.parse(heightStr.replaceAll('"', '')); updateHeight(height); }, onHttpError: (HttpResponseError error) {}, onWebResourceError: (WebResourceError error) {}, onNavigationRequest: (NavigationRequest request) { if (request.url.startsWith('https://www.youtube.com/')) { return NavigationDecision.navigate; } return NavigationDecision.prevent; }, ), ) ..loadRequest(Uri.parse( "http://some-domain.com/products/description/${resp["product_details"]["product_url"]}")); } else { webViewController?.loadRequest(Uri.parse( "http://some-domain.com/products/description/${resp["product_details"]["product_url"]}")); } } void updateHeight(double height) { webViewHeight.value = height; } } Here is the start of my view code: class ProductDetailsView extends GetView<ProductDetailsController> { const ProductDetailsView({Key? key}) : super(key: key); @override Widget build(BuildContext context) { Get.create<ProductDetailsController>(() => ProductDetailsController()); return Obx(() { if (controller.productDetailsLoading.value) { return Material( color: isDarkMode() ? DynamicColorsCollection.backgroundColorSecondaryDark : DynamicColorsCollection.backgroundColorSecondaryLight, child: LoadingShimmer(), ); } return CustomScaffold( backgroundColor: isDarkMode() ? DynamicColorsCollection.backgroundColorDark : DynamicColorsCollection.backgroundColorLight, The issue arises when I navigate to a new product details page from the current product details page using a button. The button's code is as follows: Get.toNamed( Routes.PRODUCT_DETAILS, arguments: { "product_id": id, }, preventDuplicates: false, ); When I navigate like this, the new product details page does not load the new product data. Instead, it shows the data of the previous product. How can I ensure that navigating to a new product details page with a different product ID loads the new product's data correctly? I tried navigating to the new product details page from the current product details page by using the Get.toNamed method with the new product ID as an argument. I expected the new product details page to load and display the details of the new product corresponding to the new product ID. However, instead of loading the new product's data, the page continued to display the data of the previously viewed product. I expected the ProductDetailsController to fetch and show the details of the new product based on the updated product ID passed in the arguments, but this did not happen. Because onInit olny called once, Mayby you can use GetBuilder and create new controller instance each the page is called try add global: false, to the GetBuilder Thanks, That worked 👍🏻👍🏻👍🏻👍🏻👍🏻🔥🔥🔥🔥🔥🔥🔥🔥
gharchive/issue
2024-07-15T10:34:30
2025-04-01T06:39:13.416900
{ "authors": [ "inyong1", "zenzzenpl" ], "repo": "jonataslaw/getx", "url": "https://github.com/jonataslaw/getx/issues/3146", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
328860832
Abort search when pressing escape, but accept search/input when escap… …e+enter has been pressed. Codecov Report Merging #629 into master will increase coverage by <.01%. The diff coverage is 100%. @@ Coverage Diff @@ ## master #629 +/- ## ========================================= + Coverage 71.3% 71.3% +<.01% ========================================= Files 135 135 Lines 12709 12711 +2 ========================================= + Hits 9062 9064 +2 Misses 3647 3647 Impacted Files Coverage Δ prompt_toolkit/key_binding/bindings/vi.py 57.73% <100%> (+0.05%) :arrow_up: prompt_toolkit/key_binding/bindings/emacs.py 82.46% <100%> (+0.11%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 677190f...eb90798. Read the comment docs.
gharchive/pull-request
2018-06-03T20:48:30
2025-04-01T06:39:13.436419
{ "authors": [ "codecov-io", "jonathanslenders" ], "repo": "jonathanslenders/python-prompt-toolkit", "url": "https://github.com/jonathanslenders/python-prompt-toolkit/pull/629", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
694099912
No way to remove handles or update their color position in code This is a request to add support for removing handles and/or updating their position in code with a color value Added PR #57
gharchive/issue
2020-09-05T14:44:34
2025-04-01T06:39:13.469255
{ "authors": [ "chrisCGL", "tache" ], "repo": "joncardasis/ChromaColorPicker", "url": "https://github.com/joncardasis/ChromaColorPicker/issues/54", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1111565891
Added state to the model type and updated the language docs Closes #189 by added a state item to the model type in the core AaC specification. Also started to fill in information about the AaC language in the docs by describing the model type and the data items within it. Specifically there is not a brief description of the model state and the intended utilization of state in modeling a system. Just as a side note, I was able to successfully exercise this change through the CLI in a GitPod environment but I got all sorts of errors when I tried to do it locally. I put this here for awareness that it is apparently not difficult to break or misconfigure the local development environment. I've not been able to figure out what's gone wrong yet but I'm sure it is something I've done to pollute my local environment. All in all, this has highlighted the value of strict control of the developer space for a project as it is through our GitPod configuration.
gharchive/pull-request
2022-01-22T13:37:37
2025-04-01T06:39:13.470798
{ "authors": [ "jondavid-black" ], "repo": "jondavid-black/AaC", "url": "https://github.com/jondavid-black/AaC/pull/190", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2466973465
prerelease 0.4 This will be a prerelease to help troubleshoot any GUI installation issues. The plan will be to make the full release right before the workshop at Biomag. I guess the linkcheck is going to fail for https://jonescompneurolab.github.io/hnn-core/stable/governance.html Since this is a new file that doesn't exist on the stable docs yet... What about including #860 ? @kmilo9999 I'll plan to review it so it can be included in the release before the biomag conference, but for right now I want to get this pip installable so we can test the current changes. This isn't on the release steps doc but I think @jasmainak mentioned we need to also add to doc/_templates/navbar.html. This is for the dropdown that lets you switch between docs by release version. note to self, we need to update the picture on the front page! Woohoo! v0.4! Thanks to all the contributors! @gtdang there's a slight issue with #865 The image shows up in the readme but not in the built docs, do you mind taking a look? If it calls for a different PR we can just rename this one... Are you pushing to pypi also? I see that it shows 0.3 ... I was going to until I caught the docs error,... It seemed like the image work with a local make html build but when I viewed it through circleCI the image was missing
gharchive/pull-request
2024-08-14T22:47:36
2025-04-01T06:39:13.475384
{ "authors": [ "gtdang", "jasmainak", "kmilo9999", "ntolley" ], "repo": "jonescompneurolab/hnn-core", "url": "https://github.com/jonescompneurolab/hnn-core/pull/863", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2598973776
🛑 HR - Croatia is down In ad2ae6d, HR - Croatia (https://www.skyshowtime.com/hr/help/) was down: HTTP code: 404 Response time: 774 ms Resolved: HR - Croatia is back up in 2acbcd2 after 10 minutes.
gharchive/issue
2024-10-19T09:53:31
2025-04-01T06:39:13.478034
{ "authors": [ "jonesyriffic" ], "repo": "jonesyriffic/gsp-sst", "url": "https://github.com/jonesyriffic/gsp-sst/issues/896", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2756972756
이정안님 개인 프로젝트 리뷰 1, offline 9 책전달+저녁식사, 2025-01-06 용산역 근처, 오후 6시 30분 가족 여행 일정이 길어져서 그 다음주인 13일로 변경 요청
gharchive/issue
2024-12-24T01:15:12
2025-04-01T06:39:13.478891
{ "authors": [ "jongfeel" ], "repo": "jongfeel/Activity", "url": "https://github.com/jongfeel/Activity/issues/500", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
222599420
How I can show the side menu without segue? New Issue Checklist I have read the guidelines for contributing and I understand: [x] My issue is happening in the latest version of SideMenu. [x] My issue was not solved in the README. [x] My issue can not be answered on stackoverflow.com. [x] My issue is not a request for new functionality that I am unwilling to build and contribute with a pull request. [ ] My issue is reproducible in the demo project. Issue Description I do something like this let VC1: SideMenuTableViewController = SideMenuTableViewController() let menuLeftNavigationController = UISideMenuNavigationController(rootViewController: VC1) SideMenuManager.menuLeftNavigationController = menuLeftNavigationController) self.viewController.present(SideMenuManager.menuLeftNavigationController!, animated: true, completion: nil) but doesn't work, I follow the instructions in the description of side menu but does't work, any idea? You're not following the README correctly.
gharchive/issue
2017-04-19T02:35:17
2025-04-01T06:39:13.483612
{ "authors": [ "MervinFlores", "jonkykong" ], "repo": "jonkykong/SideMenu", "url": "https://github.com/jonkykong/SideMenu/issues/180", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
33507740
Events not unbound when re-initialising tree I'm using the treeview control in a popup dialog representing a gallery selector, the treeview being the basis for the folder listing. Now, upon changing the scope of my dialog, I effectively get a new set of folders, so I need to rebuild the tree. Looking through the source, if I call .treeview({...}) on a control that has previously been initialised, it chains to Tree::_init -> Tree::_destroy(). The problem I am finding, is I rebind the onNodeSelected event when I rebuild the tree, but, the original event function is fired, and then the new event function: http://jsfiddle.net/ML5WX/1/ It seems the original events are not being removed, so the number of events that get fired per node increases with each rebind. Good spot @Antaris, thanks @Feoni4 for sharing!
gharchive/issue
2014-05-14T16:16:34
2025-04-01T06:39:13.486346
{ "authors": [ "Antaris", "jonmiles" ], "repo": "jonmiles/bootstrap-treeview", "url": "https://github.com/jonmiles/bootstrap-treeview/issues/4", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
90012622
Q: Updating via Web Sockets If I define my data statically in a page and call $('#treeview1').treeview({ data: dataToUse }); such as in your examples, the tree works fine, but when I try to update the tree's data in an "onmessage" function in web sockets by invoking $('#treeview1').treeview({ data: dataToUse }); again with the new data, the tree does not update. Any suggestions? Sounds like an issue with your implementation, not the treeview. I would suggest debugging your code, perhaps take a look at the context it's running in and/or verify the data being returned is the correct format. Check the console for errors, ask a question on stackoverflow. Yes, my fault. The data I sent to the page was very subtly malformed. Thanks. hiee i also have same issue...when i use the data being returned (that i printed on console and i copied it) it works fine...but when i return data from any function the tree does not update. Any suggestions?
gharchive/issue
2015-06-22T05:02:45
2025-04-01T06:39:13.488566
{ "authors": [ "BioCompDevSvcs", "jonmiles", "vivek144" ], "repo": "jonmiles/bootstrap-treeview", "url": "https://github.com/jonmiles/bootstrap-treeview/issues/94", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
267206253
Swap @if burguers statement into a @mixins Hello, nice burguers 👍 🍔 ! Usually each project has just one burguer. And of course we can go into the files and comment the @import, but this means touching files in node_modules ( that problably will rely on a problem if we try a package version update ) 😢 . Wouldn't be a better approach just have @mixins for each burguer ( and have them listed in the DOC) so i can @include just the one that i want ? Keeping your files clean 😃. @HectorLS Hey Hector, thanks! This was designed in a way that you won’t have to mess with anything in node_modules and also include just the ones you want. In the Customization section of the README.md there are instructions on how to do so. In short, you basically take the $hamburger-types map, remove all the ones you don’t want and just leave the ones that you want included. For example, if you only want the squeeze variant, you would put the following in your settings file or one of your .scss files for your project: $hamburgers-types: ( squeeze ); When the CSS is compiled, only that type will be included. Hope that helps.
gharchive/issue
2017-10-20T15:05:22
2025-04-01T06:39:13.500852
{ "authors": [ "HectorLS", "jonsuh" ], "repo": "jonsuh/hamburgers", "url": "https://github.com/jonsuh/hamburgers/issues/47", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
186053760
fixed relative url to http url Github markdown treated the url without the http prefix as a relative link. Added the http prefix. Thanks very much for this, it was fixed on the develop branch a couple of weeks ago (236f082 Documentation Tweaks) and is now in master.
gharchive/pull-request
2016-10-29T02:12:05
2025-04-01T06:39:13.743783
{ "authors": [ "Nzen", "jorabin" ], "repo": "jorabin/KeePassJava2", "url": "https://github.com/jorabin/KeePassJava2/pull/8", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
579541254
'Command failed' in Electron app Hello and thanks for a great library! I´ve got a problem with starting and stopping my local mysql process from my electron app. ( I am on macOS Catalina and mysql.server is set in my $PATH) I can do it with both child_process spawn and child_process exec like this: exec("sudo mysql.server start", (error, stdout, stderr) => { if (error) { console.log("error: " + error); console.log("stderr: " + stderr); } console.log("stdout: " + stdout); }); and: const child = spawn('sudo', ['mysql.server', 'start']); child.on("exit", code => { console.log(`Child process exited with code ${code}`); }); child.stdout.on("data", data => { console.log(`stdout: ${data}`); }); child.stderr.on("data", data => { console.log(`stderr: ${data}`); }); They both give me the sudo password prompt in my terminal though, so its not working when building the app. So I try to do it with sudo-prompt like this: sudo.exec("mysql.server start", { name: "My App" }, (error, stdout, stderr) => { if (error) { console.log("error: " + error); console.log("stderr: " + stderr); } console.log("stdout: " + stdout); }); I get the prompt dialog like expected, but then the command fails and gives me Error: Command failed: mysql.server start in the terminal. The strange thing is that when I run apachectl start with sudo-prompt, it works like a charm! Any thoughts on what could cause this issue? Thanks in advance! @DavidForss thanks for opening this. Did you get this working? If not, the issue might be this: mysql.server is set in my $PATH) since sudo-prompt needs you to be explicit with environment variables, because of the difficulties of making this work cross-platform.
gharchive/issue
2020-03-11T20:57:18
2025-04-01T06:39:13.747812
{ "authors": [ "DavidForss", "jorangreef" ], "repo": "jorangreef/sudo-prompt", "url": "https://github.com/jorangreef/sudo-prompt/issues/120", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1429872919
Add SHA commit to page footer use the drone sha commit to view on the webpage footer https://docs.drone.io/pipeline/environment/reference/ Good starting point
gharchive/issue
2022-10-31T13:42:31
2025-04-01T06:39:13.787264
{ "authors": [ "thamudi" ], "repo": "jordanopensource/community-web", "url": "https://github.com/jordanopensource/community-web/issues/80", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1272120514
Document keyboard shortcuts Is your feature request related to a problem? Please describe. Just lost my position in a lecture recording as a result of using the cmd+1 keyboard shortcut (i'm on a mac, so i guess the linux/windows equivalents would be different). The shortcuts cmd+n (with $n \in \{1, 2, ..., 8\}$) in Chrome will usually switch to the n-th tab from the left, but it seems like the video player overrides these commands to jump to different sections of a video. However, this is not documented anywhere. Describe the solution you'd like Document the keyboard commands. Many web-based applications (such as e.g. Gmail or GitHub) are using the shift+? keyboard shortcut to present an overlay/menu listing all possible keyboard shortcuts, so this might be a good thing to do as well. Documenting the commands would: make it clearer what they do exactly (i'm still not sure what exactly the overwritten cmd+n commands do. do they always jump in 14-minute intervals, or do they jump to 1/10-of-the-whole-video timestamps?) make the commands more discoverable. (after realising that the cmd+n commands exist i continued to press some random buttons, and it seems like j, k, l, m, z, and x also are mapped to certain actions, but then again it is impossible to easily know what they do exactly, without delving into the source code of the website) Describe alternatives you've considered Just not do anything and leave it as-is Add a section listing all available keyboard shortcuts somewhere else (e.g. in the settings page) Have some kind of questionmark-button at the bottom right of the video player, next to the camera mode and feedback buttons Yeah, good point with both documentation and the player stealing the alt+n/cmd+n shortcuts! About the documentation: basically, these are 1:1 YouTube's shortcuts where those made sense in TUM-Live (see #476 and YouTube docs). I think I could add something to document them, probably building on #489. I wasn't aware how many sites had keyboard shortcuts until I tried shift+? on them :) The issue with stolen shortcuts should be easy to fix by checking that the alt modifier is not pressed. Probably checking the control key makes sense too, but not shift, otherwise shortcuts that require shift won't work anymore :( I'm not aware how to solve this issue in full generality so that shortcuts are never stolen. @just-max Is this issue solved by #587? If not why did you not mark this as a resolving PR? To anyone reading this issue: the bug of stealing shortcuts is resolved in #587 but the enhancement of documenting shortcuts is not. Therefore the issue remains open for now.
gharchive/issue
2022-06-15T11:53:16
2025-04-01T06:39:13.826184
{ "authors": [ "CommanderStorm", "just-max", "lukaskollmer" ], "repo": "joschahenningsen/TUM-Live", "url": "https://github.com/joschahenningsen/TUM-Live/issues/530", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1394752001
mediapipe fails to install on Apple Silicon When I do: python -m pip install . I get: Processing /Users/chanana/code/investigations/webcam2 Preparing metadata (setup.py) ... done Collecting numpy~=1.20 Using cached numpy-1.23.3-cp39-cp39-macosx_11_0_arm64.whl (13.4 MB) Collecting opencv-contrib-python~=4.5.4 Using cached opencv_contrib_python-4.5.5.64-cp37-abi3-macosx_11_0_arm64.whl (38.7 MB) ERROR: Could not find a version that satisfies the requirement mediapipe~=0.8 (from matrix-webcam) (from versions: none) ERROR: No matching distribution found for mediapipe~=0.8 My machine is an M1 MacBookPro with macOS 12.6 I tried the following and it worked for me: Change line 36 of setup.py to: 'mediapipe~=0.8; platform_system != "Darwin" and platform.machine != "arm64"', It's directly from this issue. After this, doing python -m pip install . as given in instructions works! actually that was fixed this morning in version 0.4.1 :) Try pip install matrix-webcam --upgrade just guessing here, but this would break on a let's say a Raspberry/Linux arm64 machine check v0.4.2 I'm getting this crash after running matrix-webcam v0.4.2 on M1 Pro: objc[25534]: Class CaptureDelegate is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee490) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_videoio.3.4.16.dylib (0x10901c860). One of the two will be used. Which one is undefined. objc[25534]: Class CVWindow is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee4e0) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_highgui.3.4.16.dylib (0x108048a68). One of the two will be used. Which one is undefined. objc[25534]: Class CVView is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee508) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_highgui.3.4.16.dylib (0x108048a90). One of the two will be used. Which one is undefined. objc[25534]: Class CVSlider is implemented in both /opt/homebrew/lib/python3.10/site-packages/cv2/cv2.abi3.so (0x1057ee530) and /opt/homebrew/lib/python3.10/site-packages/mediapipe/.dylibs/libopencv_highgui.3.4.16.dylib (0x108048ab8). One of the two will be used. Which one is undefined. Add a python dependency on "@com_google_protobuf//:protobuf_python" Traceback (most recent call last): File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/descriptor.py", line 47, in <module> from google.protobuf.pyext import _message ImportError: dlopen(/opt/homebrew/lib/python3.10/site-packages/google/protobuf/pyext/_message.cpython-310-darwin.so, 0x0002): symbol not found in flat namespace (__ZN6google8protobuf15FieldDescriptor12TypeOnceInitEPKS1_) Traceback (most recent call last): File "/opt/homebrew/bin/matrix-webcam", line 5, in <module> from matrix_webcam.__main__ import main File "/opt/homebrew/lib/python3.10/site-packages/matrix_webcam/__main__.py", line 13, in <module> from mediapipe.python.solutions.selfie_segmentation import SelfieSegmentation File "/opt/homebrew/lib/python3.10/site-packages/mediapipe/__init__.py", line 16, in <module> from mediapipe.python import * File "/opt/homebrew/lib/python3.10/site-packages/mediapipe/python/__init__.py", line 28, in <module> import mediapipe.python.packet_getter File "/opt/homebrew/lib/python3.10/site-packages/mediapipe/python/packet_getter.py", line 20, in <module> from google.protobuf import symbol_database File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/symbol_database.py", line 62, in <module> from google.protobuf import descriptor_pool File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/descriptor_pool.py", line 63, in <module> from google.protobuf import descriptor File "/opt/homebrew/lib/python3.10/site-packages/google/protobuf/descriptor.py", line 47, in <module> from google.protobuf.pyext import _message ImportError: dlopen(/opt/homebrew/lib/python3.10/site-packages/google/protobuf/pyext/_message.cpython-310-darwin.so, 0x0002): symbol not found in flat namespace (__ZN6google8protobuf15FieldDescriptor12TypeOnceInitEPKS1_) Could you help me with it? What were your steps to set it up? It seems like you're using python3.10, perhaps you could try 3.9.13 instead and see if it still fails? I think using an older version of the protobuf library (by using py3.9) might help but that's just a guess @chanana Thanks! python 3.9.13 helps me.
gharchive/issue
2022-10-03T13:37:58
2025-04-01T06:39:13.833156
{ "authors": [ "anikitin-intermedia", "chanana", "joschuck" ], "repo": "joschuck/matrix-webcam", "url": "https://github.com/joschuck/matrix-webcam/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
65886422
Bower support Please add this cool jQuery plugin to the Bower. My Fork is available as package jquery-colpick. All versions 2.* are identical to those here. @mrgrain Thank you very much!
gharchive/issue
2015-04-02T09:19:08
2025-04-01T06:39:13.837617
{ "authors": [ "mrgrain", "radekdostal" ], "repo": "josedvq/colpick-jQuery-Color-Picker", "url": "https://github.com/josedvq/colpick-jQuery-Color-Picker/issues/46", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }