added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:39:15.431469
2020-05-13T14:23:44
617492882
{ "authors": [ "brianxjx", "mchenggit" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7439", "repo": "kabanero-io/kabanero-pipelines", "url": "https://github.com/kabanero-io/kabanero-pipelines/issues/374" }
gharchive/issue
Replacing-pipelineresources-with-tasks PipelineResources didn't make the cut for beta and have been replaced by a combination of Tekton Catalog Tasks and Workspaces. v1alpha1 => v1beta1 The new odo based pipelines are already using task and workspace. There is no plan to change existing Appsody based pipeline until we know when PipelineResource is removed from Tekton.
2025-04-01T06:39:15.469361
2024-10-11T11:58:53
2581211326
{ "authors": [ "azatsafin", "odorT" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7440", "repo": "kafbat/helm-charts", "url": "https://github.com/kafbat/helm-charts/issues/29" }
gharchive/issue
Error: UPGRADE FAILED: parse error at (kafka-ui/templates/deployment.yaml:57): unclosed action Issue submitter TODO list [X] I've looked up my issue in FAQ [X] I've searched for an already existing issues here (legacy) and here [X] I've tried installing latest charts and the issue still persists there [X] I'm running a supported version of the application & chart which is listed here Describe the bug (actual behavior) it seems new version of kafbat helm chart can't be upgraded anymore Expected behavior installation without errors Your installation details https://github.com/kafbat/kafka-ui/commit/2956664 kafka-ui-1.4.2 yamlApplicationConfig: kafka: clusters: - name: test bootstrapServers: test.test:9092 schemaRegistry: http://test1.test1:8081 4. - Steps to reproduce just applying this values fails Screenshots No response Logs No response Additional context No response https://github.com/kafbat/helm-charts/blob/ab4948ba71b99576b7096a120597989a6c869a53/charts/kafka-ui/templates/deployment.yaml#L57 this change broke master any updates on this? any updates on this? Could you kindly share the complete values file and the specific error message you encountered? command: helm upgrade --install kafka-ui kafbat-ui/kafka-ui -f values.yaml -n kafka-ui output: Release "kafka-ui-deleteme" does not exist. Installing it now. Error: parse error at (kafka-ui/templates/deployment.yaml:57): unclosed action helm version: version.BuildInfo{Version:"v3.5.4", GitCommit:"1b5edb69df3d3a08df77c9902dc17af864ff05d1", GitTreeState:"clean", GoVersion:"go1.15.11"} values.yaml: replicaCount: 1 route: enabled: true env: - name: filtering.groovy.enabled value: "true" - name: DYNAMIC_CONFIG_ENABLED value: "true" resources: requests: cpu: '1' memory: '1Gi' yamlApplicationConfig: management: health: ldap: enabled: false volumes: - name: config-data persistentVolumeClaim: claimName: config-pv-claim volumeMounts: - mountPath: "/etc/kafkaui" name: config-data helm version: version.BuildInfo{Version:"v3.5.4", GitCommit:"1b5edb69df3d3a08df77c9902dc17af864ff05d1", GitTreeState:"clean", GoVersion:"go1.15.11"} Please upgrade helm to the latest version and try again. yes, it worked ! thanks a lot
2025-04-01T06:39:15.475262
2024-05-24T20:15:05
2316190045
{ "authors": [ "YpNo", "chicknlil", "kaffetorsk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7441", "repo": "kaffetorsk/arlo-streamer", "url": "https://github.com/kaffetorsk/arlo-streamer/issues/22" }
gharchive/issue
Pyaarlo Please update this to the most current version of pyaarlo. The one you are using doesn't have the most recent fixes, nor does it support the newest devices. @YpNo it seems like our fearless leader @kaffetorsk has jumped ship and that you are the only one answering questions, trying to update the code, etc. Are you able to fork this project to be able to keep on top of things? Much obliged for your consideration. Ahah, I have been thinking about it but we should find a real python developper and/or a video stream "expert" to help. I can do some things but I could reach some limits. I'll keep you inform Hi, Okay okay, I think I will fork the project :) @chicknill @bbo76 @RaidMax @xitation I'll keep you inform when the fork is ready. I don't know if I would treat all your issues or suggestions but I'll do my best. If you know a real Python developpers with asynchronous operation skill and video handling to help me, it would be great ! The fork is ready if you want to try : https://github.com/YpNo/arlo-camera-streamer It is the same version with little changes. It should run as today. But it has been built with the latest version of pyaarlo ;) Feel free to (re)open issue/discussion about your needs. Welcome back @kaffetorsk ! I will close my fork if youโ€™re really return back. Feel free to find contributors to help you maintain this great app. I can be one of them if you let me build new releases (with new version of pyaarlo first). Let us know about the future of this project. Regards. Thank you, I will take better care of this repo going forward and appreciate the effort you put in. As to the future of the project I think it is best to keep the scope to it's original intent (which is enabling arlo cameras for 3rd party NVRs (such as frigate)) and keep it maintained. I also realize the need for maintenance is higher than I expected, mostly due to pyaarlo beeing based on a reverse-engineered API, suspect to change at any given moment :-) I'll look into adding contributors Dropped release v0.7.5 btw, pyaarlo is latest bleeding edge (my fork with a yet to be merged pr)
2025-04-01T06:39:15.494823
2024-02-06T17:32:12
2121341608
{ "authors": [ "Itxaka", "jimmykarily", "mudler" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7442", "repo": "kairos-io/kairos", "url": "https://github.com/kairos-io/kairos/issues/2217" }
gharchive/issue
UKI: kcrypt unlock-all doesn't unlock TPM-bound partitions Currently unlocking partitions encrypted with TPM manually by calling kcrypt unlock-all doesn't work. Workaround exists, and documented in https://kairos.io/docs/installation/trustedboot/#mount-partitions-after-install To reproduce: In the Kairos config, try to run kcrypt unlock-all in an after-install stage (e.g. to write some files to the disk) Install Kairos in UKI mode See installation failing. Possible solution: Introduce a new stage/hook (e.g. "after-decrypt") to allow people to run code right after decrypting the disks. This stage will also make sure the disk is encrypted again when the stage is done. Also, there is a --tpm flag in kcrypt unlock-all command that might workaround the issue. kcrypt unlock-all could work if we add a --tpm flag so it knows it needs to go over the tpm unlock workflow instead of the usual one. Introduced in version 0.9.0: https://github.com/kairos-io/kcrypt/compare/v0.7.0...v0.9.0 (cut in December 18th: https://github.com/kairos-io/kcrypt/releases/tag/v0.9.0)
2025-04-01T06:39:15.495992
2024-04-11T09:00:21
2237245668
{ "authors": [ "mauromorales" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7443", "repo": "kairos-io/kairos", "url": "https://github.com/kairos-io/kairos/pull/2463" }
gharchive/pull-request
Symlink any /boot/Image* to /boot/vmlinuz The jetson produces /boot/Image which wasn't caught by the previous mechanism fixes #2461 tested manually, this tests dont' run on pr so merging
2025-04-01T06:39:15.506055
2021-07-25T11:05:46
952250620
{ "authors": [ "kaiwalyakoparkar", "vind3v17" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7444", "repo": "kaiwalyakoparkar/classroom-monitor-bot", "url": "https://github.com/kaiwalyakoparkar/classroom-monitor-bot/pull/23" }
gharchive/pull-request
fix: ๐Ÿ“Updated Discord Link Fixes Issue (#22 ) This PR fixes the following issues : Updated Discord Link In Greeting Action. hey @vind3v17, I see a different link added there. Kindly add this link (I mentioned this in the issue as well) https://discord.gg/K9kxUXvfND
2025-04-01T06:39:15.538150
2021-12-17T08:56:22
1083027564
{ "authors": [ "cnouguier" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7445", "repo": "kalisio/kapture", "url": "https://github.com/kalisio/kapture/issues/7" }
gharchive/issue
Waiting for animation When capturing a GeoJSON file we need to wait fot the animation. For now we have defined a wait duration according the activity (map or globe). It could be great to tell Kano to swith off any animation when zooming. For now, we prodive a waitDelay before taking the snapshot. It could be defined using a WAIT_DELAY environment variable or could be provided on the query
2025-04-01T06:39:15.723062
2022-12-20T03:09:05
1503917791
{ "authors": [ "duongbinh214", "kameleo-team" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7446", "repo": "kameleo-io/local-api-client-csharp", "url": "https://github.com/kameleo-io/local-api-client-csharp/issues/2" }
gharchive/issue
'No connection could be made because the target machine actively refused it. (localhost:5050)' My code : Error: var baseProfileList = await client.SearchBaseProfilesAsync(deviceType: "desktop", browserProduct: "chrome"); pls help me!!!!!! Are you sure that Kameleo.CLI.exe is running. If yes, please also make sure it runs on 5050 port. Please see this article to see how to start Kameleo.CLI.exe
2025-04-01T06:39:15.727501
2017-03-24T09:35:52
216716915
{ "authors": [ "kamilsk" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7447", "repo": "kamilsk/dzone", "url": "https://github.com/kamilsk/dzone/issues/34" }
gharchive/issue
integrate easyjson https://github.com/mailru/easyjson as a result of https://github.com/kamilsk/dzone/issues/31 blocked by https://github.com/kamilsk/shared/issues/144
2025-04-01T06:39:15.731751
2015-01-28T18:45:30
55797577
{ "authors": [ "kamisama", "shadyb" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7448", "repo": "kamisama/Fresque", "url": "https://github.com/kamisama/Fresque/issues/45" }
gharchive/issue
Specifying worker names with stop command I am going to implement a feature that allows users to specify the worker to kill in command line. This is useful if you are using some automation/CI tools (jenkins, ansible, puppet etc). There are going to be two additional options to the stop command: --worker and --count. The --worker option allows you to specify a queue name to kill. The --count option works in tandem with the --worker option and allows you to specify how many to kill. I'm not sure if you accept feature additions to your codebase, but if you want, I can make a pull request, implement it and then merge back into your repo. That way others can benefit. Let me know what you think I don't quite get what you're trying to do. The stop command already print a list of workers, and you can choose the worker you want to stop. I think there's naming issue, but are you trying to stop a worker, by the name of the polled queue ? In that case, a --queue option would make more sense. Sorry, I mean't --queue (been working very long hours as of late so am tripping up all over the place). It does print a list of workers, but then it requires stdin input and that's a problem if you are using automation tools. Take my configuration for example, we have a jenkins project which allows you to specify queues to start. I make a selection and it plugs the variables into a bash script and runs it.
2025-04-01T06:39:15.735156
2019-10-14T14:00:10
506672585
{ "authors": [ "kamleshchandnani", "rohandaxini" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7449", "repo": "kamleshchandnani/awesome-interview-process", "url": "https://github.com/kamleshchandnani/awesome-interview-process/pull/10" }
gharchive/pull-request
Fix typo and improve wording in README Hello @kamleshchandnani Great initiative ๐Ÿ‘ Just a minor suggestion if you don't mind ๐Ÿ˜„ I have fixed following in the README Typo Improved the header. You may refer hemingwayapp for the same. You may review the fixes here. PS: We will also raise a PR soon to add Kiprosh to the list. Hey @rohandaxini , Thank you so much for taking out time and contributing to this project ๐ŸŽ‰
2025-04-01T06:39:15.740871
2024-04-24T07:09:01
2260542420
{ "authors": [ "filgoBot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7450", "repo": "kamp-us/monorepo", "url": "https://github.com/kamp-us/monorepo/issues/863" }
gharchive/issue
Curriculum update needed on how_does_the_web_work.md The Odin's file, how_does_the_web_work.md is updated. Please update the Kampus' file, checkout file here how_does_the_web_work.md Latest commits: How Does the Web Work?: Update descriptive link text (#27681) (#27689) (additions: 7, deletions: 7) on Mar 26 2024, 14:43 UTC New commits have been made to the Odin's file. Please update the Kampus' file. Latest commits: Surround html tag with lines to follow lint (#28132) (additions: 6, deletions: 4) on Jun 8 2024, 02:29 UTC
2025-04-01T06:39:15.744164
2016-09-08T11:46:57
175733035
{ "authors": [ "kamsar", "waaromikniet" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7451", "repo": "kamsar/Unicorn", "url": "https://github.com/kamsar/Unicorn/issues/170" }
gharchive/issue
Getting Server cannot append header after HTTP headers have been sent Hi, I am getting Server cannot append header after HTTP headers have been sent when I try to sync the roles with unicorn. Any ideas? Regards Danny Duplicate of #155. This is fixed in the current prerelease version already. I am assuming that you're using Sitecore 8.1U3 or 8.2. You can work around this by setting the 'X-Frame-Options' header yourself prior to Sitecore trying to set it on request end.
2025-04-01T06:39:15.818956
2016-06-16T08:08:15
160601638
{ "authors": [ "chenbs", "emersion", "karalabe" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7452", "repo": "karalabe/xgo", "url": "https://github.com/karalabe/xgo/issues/55" }
gharchive/issue
windows: missing TaskDialog Fixed in a recent mingw-w64 # github.com/andlabs/ui ../../andlabs/ui/libui_windows_amd64.a(stddialogs.cpp.obj): In function `msgbox': /home/simon/projects/libui/windows/stddialogs.cpp:113: undefined reference to `__imp_TaskDialog' collect2: error: ld returned 1 exit status how to resolve ? This seems to only be part of mingw 5+, which has not yet been released.
2025-04-01T06:39:15.839083
2017-12-22T14:06:08
284171299
{ "authors": [ "Shaileshz204", "avpavlov", "ptrthomas" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7453", "repo": "karatelabs/karate", "url": "https://github.com/karatelabs/karate/issues/270" }
gharchive/issue
[feature request] Support for #notpresent keyword Use case: I'm writing tests for Spring Boot Repository. It supports "projections" so user can request different detalization, e.g. "default" projection: Fields A,B,C "more-data" projection: Fields A,B,C and array D (it is loaded from separate table) "complete-data" projection: Fields A,B,C and arrays D,E,F,G (each is loaded from separate table) I cannot check contains/!contains with single JSON so right now my tests look like Scenario "default" <request> then match response contains """ { A : '#string', B : '#string' , C : '#string' } """ and match response !contains """ { D : '#array', E : '#array', F : '#array', G : '#array' } """ Scenario "more-data" <request> then match response contains """ { A : '#string', B : '#string' , C : '#string', D : '#array' } """ and match response !contains """ { E : '#array', F : '#array', G : '#array' } """ Scenario "complete-data" <request> then match response contains """ { A : '#string', B : '#string' , C : '#string', D : '#array', E : '#array', F : '#array', G : '#array' } """ With new keyword added my scenarios will be cleaner and easier to read Scenario "default" <request> then match response contains """ { A : '#string', B : '#string' , C : '#string' D : '#undefined', E : '#undefined', F : '#undefined, G : '#undefined' } """ Scenario "more-data" <request> then match response contains """ { A : '#string', B : '#string' , C : '#string', D : '#array', E : '#undefined', F : '#undefined, G : '#undefined' } """ Scenario "complete-data" <request> then match response contains """ { A : '#string', B : '#string' , C : '#string', D : '#array', E : '#array', F : '#array', G : '#array' } """ @avpavlov actually there is an #ignore marker already, can you confirm that it is what works for you: https://github.com/intuit/karate#fuzzy-matching #ignore excludes field from verification. I do not want to exclude, I want to ensure this field is not included. In other words, #undefined is like local !contains for single field would #notnull work ? I meant #null Thank you, both work - #null and ##null. However, I would say working #null is a bug in this case, because if I expect some key with null value then missing key should fail scenario. ##null is right fit in this case - any chances you expand "contains / !contains" readme sections to propose it as way to verify key is not presented in JSON? However, I would say working #null is a bug in this case, because if I expect some key with null value then missing key should fail scenario. Yes. Or how about this, to check for a null just use null itself. So this should fail (as of now it passes, but I'm proposing to change this): * def foo = { } * match foo == { a: null } In my experience, most teams assume a null value and the key missing to be the same. Typically people set the Json marshaller config to 'strip nulls' for example, to reduce payload bloat. Can you help by suggesting what change you'd like to see in the contains readme section, I'll be happy to add. (as of now it passes, but I'm proposing to change this): Sounds good most teams assume a null value and the key missing to be the same. That's true. That's why #undefined could help ;) Can you help by suggesting what change you'd like to see in the contains readme section, Let me think @avpavlov cool, I'm beginning to agree. see I'm quite reasonable :P how about #notpresent - which could be more clear ? undefined has a certain meaning in JS also ? notpresent sounds better than undefined. I tried to invent something like na or notavailable or missed but all of these were not enough clear so finally I borrowed keyword from JS great ! looking at it now. are you able to build from source or do you prefer a release. Created PR for '##null' in README https://github.com/intuit/karate/pull/271 I can build from sources @avpavlov yes, is now in the develop branch. I actually decided that both #null and the null value will expect the JSON key to be present. Just felt that this is consistent and reduces confusion. Teams may need to use #ignore or #notpresent which is more clear. I agree, Karate tests output/protocol not how this output could be interpreted by consumer Just tested project with develop branch - it works! Thank you! Is there any way in the Github to subscribe to the notifications about releases? @avpavlov not sure, but you can watch the project on GiHub. there's a twitter account if you are in to that kind of thing: https://twitter.com/KarateDSL @avpavlov well. after thinking about it, it made sense to implement #present as well ! thanks for triggering this, I think Karate has become a little better as a result. So now here is what is possible. I recommend that folks don't use != not-equals because it gets really confusing, but hey. * def foo = { } * match foo != { a: '#present' } * match foo == { a: '#notpresent' } * match foo == { a: '#ignore' } * match foo == { a: '##null' } * match foo != { a: '#null' } * match foo != { a: '#notnull' } * match foo == { a: '##notnull' } * match foo != { a: null } * def foo = { a: null } * match foo == { a: '#null' } * match foo == { a: '##null' } * match foo != { a: '#notnull' } * match foo != { a: '##notnull' } * match foo == { a: '#present' } * match foo == { a: '#ignore' } * match foo != { a: '#notpresent' } * def foo = { a: 1 } * match foo == { a: '#notnull' } * match foo == { a: '##notnull' } * match foo != { a: '#null' } * match foo != { a: '##null' } * match foo == { a: '#present' } * match foo == { a: '#ignore' } * match foo != { a: '#notpresent' } @ptrthomas When I am validating schema using Karate one of the field as per schema is defined as String but its returning null but it returns string value also in some cases so when its returning null its failing my test as its string. How to handle this. I want it to pass if its null or string when its string in schema. @Shaileshz204 use stack overflow for questions like this please: https://stackoverflow.com/a/71522605/143475
2025-04-01T06:39:15.841799
2017-09-13T08:18:10
257294817
{ "authors": [ "brunetto", "kardianos" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7454", "repo": "kardianos/govendor", "url": "https://github.com/kardianos/govendor/issues/359" }
gharchive/issue
Add tests to vendoring only for specific packages Hi, I'm sorry if it is trivial but I was not able to find any hint. Is there a way to add tests only for specific packages in the vendor folder? I can ignore the tests for all the packages with "ignore" in vendor.json file or, without it, all the tests for all the packages are added. Nope. There is no way to do a package specific ignore requirement. Ok, thanks... and is there a way to avoid having tests ignored in vendor.json after govendor init? You can always remove that after a govendor init. It is just what I consider a "sane default". Yeah it is a sane default that unfortunately does not apply to my employer default pipeline. I'm trying to find the best way to deal with it. Thanks a lot.
2025-04-01T06:39:15.901544
2024-08-17T10:40:16
2471471473
{ "authors": [ "dlmw" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7455", "repo": "karlomikus/bar-assistant", "url": "https://github.com/karlomikus/bar-assistant/issues/312" }
gharchive/issue
Recipe images not retrieved [x] I have read the FAQ. Describe the bug I have just created my Bar Assistant instance. Everything works fine but the images don't appear. However, thumbnails work fine. After inspecting the HTML, I see that the URI of the picture is "https://bar.mydomain.com/uploads/cocktails/1/1934-cosmo-1_79lTcl.jpg", but it should be "https://bar.dlmw.ch/bar/uploads/cocktails/1/1934-cosmo-1_79lTcl.jpg". My reverse proxy is Caddy and the Caddyfile looks somewhat like this: bar.mydomain.com { handle_path /search/* { reverse_proxy meilisearch:7700 } handle_path /bar/* { reverse_proxy barassistant:3000 } handle_path /* { reverse_proxy saltrim:8080 } } To Reproduce Create a Caddyfile containing the snippet above Versions: Docker: 24.0.7 LibreWolf: 129.0-1 I managed to solve it by modifying the Caddyfile like so: bar.dlmw.ch { # here @uploads { path_regexp uploads ^/uploads/(.*) } rewrite @uploads /bar/uploads/{re.uploads.1} handle_path /search/* { reverse_proxy meilisearch:7700 } handle_path /bar/* { reverse_proxy barassistant:3000 } handle_path /* { reverse_proxy saltrim:8080 } }
2025-04-01T06:39:15.905812
2015-05-12T21:15:28
75741338
{ "authors": [ "MetaMemoryT", "maksimr", "zzo" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7456", "repo": "karma-runner/karma", "url": "https://github.com/karma-runner/karma/issues/1402" }
gharchive/issue
npm install WARN message with npm install of karma, npm warns: npm WARN engine<EMAIL_ADDRESS>wanted: {"node":"~0.8 || ~0.10"} (current: {"node":"0.12.0","npm":"2.7.2"}) #merge Will be fixed in new version karma
2025-04-01T06:39:15.912869
2016-06-27T14:01:39
162459166
{ "authors": [ "cellog", "dignifiedquire", "johnjbarton", "trusktr", "wesleycho" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7457", "repo": "karma-runner/karma", "url": "https://github.com/karma-runner/karma/issues/2211" }
gharchive/issue
when karma hangs, how do I debug it? In my project, https://github.com/cellog/react-selection , I have encountered a bug when running karma remotely on saucelabs. It is caused by code coverage. In short, karma connects to the browser, loads, and then hangs indefinitely. When I remove code coverage, the test runs in about 16 seconds. This happens only remotely, when I run coverage on my local machine, it works great. How can I debug the hang on saucelabs? I'm new to karma. When using karma remotely with saucelabs, is it loading karma from my machine? If yes, which file can I put debug code into to see if I can figure out where it is failing? Or is there another way? Thanks much. Oh and I should mention the code coverage is accomplished by a babel plugin, so there is nothing inside karma, it just serving the transpiled files directly to saucelabs setting logLevel to DEBUG should give you more details, after that regular debugging techniques for node loke console.log in code and attaching a debugger it turns out that any concurrency larger than 1 causes the entire thing to fail with disconnects to every browser. Is this expected behavior? On saucelabs that is probably caused by rate limiting from their side. fyi, the problem is in the karma-saucelabs-launcher, which is unmaintained and doesn't work any more. After 50 tests, it just dies. I switched to browserstack and it works perfectly. However, there are some side effects of running more than 1 browser test in the same karma process. I have no idea how to isolate them, but the ONLY way I can get karma to work and generate code coverage is to run each browser in its own karma process. You can see what I mean in https://github.com/cellog/react-selection Major pain. Fortunately, I can do local development with karma very quickly, and then push on commit to find those chance browser differences eventually. Test runs now take about 15 minutes each, so it's really a pain, but at least it works. So to summarize: there are several bugs that I can't fix or even track down. The first is that karma and sauce labs are no longer friends, and so I can't use them together at all. The second is that karma has some kind of strange shared stuff interfering with each other whenever I run more than 1 browser in the same karma process, even if concurrency is set to 1. That's a huge one, and might be worth investigating further, since it means the sandboxing is leaking. Let me know if you want me to try things to debug it. FWIW, I have noticed that the Angular team has set up karma with Sauce Labs to use Sauce Labs via shell script in separate processes - if I had to guess, it is to have environmental separation for purer test environments, but maybe they came across some of the same issues with multiple browsers. Separate issues should probably be opened if there are specific issues found though, that way we can track them better. Going to close this issue, but feel free to open issues for actionable problems you find with karma. attaching a debugger How exactly? I tried ndb ./node_modules/.bin/karma ... but nothing runs, and the opened devtools window is empty, no source files, nothing paused. The correct answer is to move to jest and forget about it Unfortunately I've found ndb fail sometimes. I use node --inspect-brk but I also don't use the .bin file directly.
2025-04-01T06:39:15.917482
2021-02-08T11:52:14
803491200
{ "authors": [ "AppVeyorBot", "devoto13", "jimbojw", "maksimr", "xel23" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7458", "repo": "karma-runner/karma", "url": "https://github.com/karma-runner/karma/pull/3651" }
gharchive/pull-request
fix(config): check extension before ts-node register Call require('ts-node').register() after checking configFilePath has .ts extension Fixes #3329 ัั @devoto13 :white_check_mark: Build karma 2927 completed (commit https://github.com/karma-runner/karma/commit/4157266d7a by @xel23) @googlebot I signed it! @xel23 Thanks for the PR! I guess we can land it as a workaround to solve the most painful appearance of this issue, but I don't think it "fixes" the issue. We really should implement a systematic solution as outlined in https://github.com/karma-runner/karma/issues/3329#issuecomment-772262377. Technically this is a breaking change. People, who might have relied on type-checking their karma.conf.js with allowJS: true will no longer have the type-checking. As this does not prevent them from running Karma, I am willing to do so to resolve the issue for the majority of users. @johnjbarton What do you think? Ping Thank you so much for fixing this! ๐Ÿฅณ ๐ŸŽ‰
2025-04-01T06:39:15.953840
2023-06-13T02:53:36
1753934976
{ "authors": [ "TheUltDev" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7459", "repo": "kat-tax/vslite", "url": "https://github.com/kat-tax/vslite/pull/21" }
gharchive/pull-request
Feat/collaboration This adds Figma -> RN plugin syncing as well as generic editor syncing: https://feat-collaboration.vslite.pages.dev/#/strait_domestic_heat_dean_patronage Right now you need to run your own local websocket server: HOST=localhost PORT=1234 npx y-websocket All of this is work in progress, but the PoC is working... Merged initial sync support. Will follow up on this in another PR.
2025-04-01T06:39:15.955203
2019-05-20T10:52:45
446050339
{ "authors": [ "jodh-intel" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7460", "repo": "kata-containers/documentation", "url": "https://github.com/kata-containers/documentation/pull/482" }
gharchive/pull-request
docs: Added missing doc link Added a link to the existing how-to-use-virtio-fs-with-kata.md. Fixes #481. Signed-off-by: James O. D. Hunt<EMAIL_ADDRESS> /test
2025-04-01T06:39:15.980828
2019-05-10T09:35:01
442627585
{ "authors": [ "egernst", "grahamwhaley", "jodh-intel", "zhiminghufighting" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7461", "repo": "kata-containers/runtime", "url": "https://github.com/kata-containers/runtime/issues/1661" }
gharchive/issue
qemu-lite failed to enable vga when launching Kata container Description of problem When I reconfigure Qemu-lite launch parameter from โ€œ-vga noneโ€ to โ€œ-vga stdโ€ to enable vga in Kata runtime, it reports PCI unavailable error as below. Is there any dependence to enable vga? I ask some people and there is no dependence in kvm +qemu to launch an centos VM. Expected result Launch Kata container successfully. Actual result May 10 06:31:21 localhost.localdomain kata-runtime[28005]: time="2019-05-10T06:31:21.574044938-04:00" level=info msg="launching /usr/bin/qemu-lite-system-x86_64 with: [-name sandbox-c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b -uuid 3e6695d3-d12f-4bfe-b1e5-2369c0b96df2 -machine pc,accel=kvm,kernel_irqchip,nvdimm -cpu host -qmp unix:/run/vc/vm/c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b/qmp.sock,server,nowait -m 2048M,slots=10,maxmem=8752M -device pci-bridge,bus=pci.0,id=pci-bridge-0,chassis_nr=1,shpc=on,addr=2,romfile= -device virtio-serial-pci,disable-modern=false,id=serial0,romfile= -device virtconsole,chardev=charconsole0,id=console0 -chardev socket,id=charconsole0,path=/run/vc/vm/c522a66fa441bb1eb5 be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b/console.sock,server,nowait -device nvdimm,id=nv0,memdev=mem0 -object memory-backend-file,id=mem0,mem-path=/usr/share/kata-containers/kata-containers-image_clearlinux_1.5.0_agent_a581aebf473.img,size=536870912 -device virtio-scsi-pci,id=scsi0,disable-modern=false,romfile= -object rng-random,id=rng0,filename=/dev/urandom -device virtio-rng,rng=rng0,romfile= -device virtserialport,chardev=charch0,id=channel0,name=agent.channel.0 -chardev socket,id=charch0,path=/run/vc/vm/c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2 f51d077a5dba7732987b/kata.sock,server,nowait -device virtio-9p-pci,disable-modern=false,fsdev=extra-9p-kataShared,mount_tag=kataShared,romfile= -fsdev local,id=extra-9p-kataShared,path=/ru n/kata-containers/shared/sandboxes/c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b,security_model=none -netdev tap,id=network-0,vhost=on,vhostfds=3,fds=4 -device driver=virtio-net-pci,netdev=network-0,mac=02:42:ac:12:00:02,disable-modern=false,mq=on,vectors=4,romfile= -global kvm-pit.lost_tick_policy=discard -vga std -no-user-config -nodefaults -nographic - daemonize -kernel /usr/share/kata-containers/vmlinuz-<IP_ADDRESS>-143.1.container -append tsc=reliable no_timer_check rcupdate.rcu_expedited=1 i8042.direct=1 i8042.dumbkbd=1 i8042.nopnp=1 i8042.noaux=1 noreplace-smp reboot=k console=hvc0 console=hvc1 iommu=off cryptomgr.notests net.ifnames=0 pci=lastbus=0 root=/dev/pmem0p1 rootflags=dax,data=ordered,errors=remount-ro rw rootf stype=ext4 debug systemd.show_status=true systemd.log_level=debug panic=1 nr_cpus=8 init=/usr/lib/systemd/systemd systemd.unit=kata-containers.target systemd.mask=systemd-networkd.service systemd.mask=systemd-networkd.socket -smp 1,cores=1,threads=1,sockets=1,maxcpus=8]" arch=amd64 clicreate container=c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b command= create name=kata-runtime pid=28005 source=virtcontainers subsystem=qmp level=error msg="Unable to launch /usr/bin/qemu-lite-system-x86_64: exit status 1" arc h=amd64 clicreate container=c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b command=create name=kata-runtime pid=28005 source=virtcontainers subsystem=qmp May 10 06:31:21 localhost.localdomain kata-runtime[28005]: time="2019-05-10T06:31:21.633469001-04:00" level=error msg="qemu-lite-system-x86_64: -device pci-bridge,bus=pci.0,id=pci-bridge-0,chassis_nr=1,shpc=on,addr=2,romfile=: PCI: slot 2 function 0 not available for pci-bridge, in use by VGA\n" arch=amd64 clicreate container=c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d 077a5dba7732987b command=create name=kata-runtime pid=28005 source=virtcontainers subsystem=qmp Hi @zhiminghufighting - what are you trying to achieve? Were you hoping that the container would then gain access to the hosts VGA controller, or? It might help us understand your goal and how to achieve it :-) For reference, in case it helps, there is a guide on how to pass GPU in/out of Kata at https://github.com/kata-containers/documentation/blob/master/use-cases/GPU-passthrough-and-Kata.md And there is a link to a container that can help get X11 working in a container (including a kata container) at https://github.com/kata-containers/documentation/wiki/UserGuide#x11-containers @zhiminghufighting - qemu-lite is extremely minimal (by design), so doesn't have graphical support enabled. We disable all graphics at build time - see: https://github.com/kata-containers/packaging/blob/master/scripts/configure-hypervisor.sh#L206 @grahamwhaley The background is here: i want to launch android based container image in Kata which is used to support cloud gaming. As you know, lots of gaming is running on billion mobile devices with Android OS in China. there is a big request to enable android container in Kata. After i launch Kata container with android based image with our OTC android in container team and found there is an android key process --surface finger can't be started normally caused by wrong framer buffer size. But if i need to get the framer buffer size, i need to enable vga in qemu-lite to support right framer buffer to start software render. Here android container needs a software render and framer buffer & virtual VGA is a mandatory option. The android container doesn't care about if there is a real VGA or GPU in host because it depends on software render library --opengl in android container image. In above scenario, there is no need of hardware render in android container, so it doesn't depend on GPU GVT-g or GVT-d. I know and check all the related doc of enabling GPUs for kata in your link. Thanks for your quick response. Is there any other options for me to use vga? @jodh-intel please see the detail background and root cause of this requirement in above answer to Graham's question. I think you already answer the root cause of why the error log is reported by qemu-lite. Can i enable gpu support by rebuild qemu-lite? If yes, is there any guide doc? And if there is any other ways? I think even i enable gpu support to solve this issue, it will cause qemu-lite being heavy and increase the memory footprint and resource consumption. thanks for your quick response! By the way, there would be a huge potential user scenario for Kata beside cloud is mentioned by me in above comment : gaming in cloud and cloud gaming based on android image container in chine. Billions of android mobile device and thousands of gaming based on android OS. @zhiminghufighting - qemu-lite is a highly optimised version of qemu designed for "standard" sorts of container use-cases so doesn't include graphical support. To just prove the concept, you could of course simply change your config to use the distro-packaged version of qemu which should contain graphics support: path = "/usr/bin/qemu-system-x86_64" Although you could rebuild qemu-lite to include graphic support, I don't think that is the best approach. Depending on the architecture, either NEMU or qemu-vanilla would be better I think. However, both those options are also built without graphical support so you would have to rebuild them with some of the graphical options we disable using https://github.com/kata-containers/packaging/blob/master/scripts/configure-hypervisor.sh. However, you may need to make changes to the guest kernel configuration as that too is as minimal as possible for "normal" sorts of workloads. We don't have documentation for this sort of scenario currently. Next steps This is an interesting scenario but since it is going to require changes specifically for graphical support and since those may adversely affect memory density and boot speed, we'd need to study the idea carefully. We would also need additional tests, extra testing infrastructure and documentation. As such, I suggest you raise an RFC issue using the main project repo (which we tend to use to discuss large features that potentially affect lots of different repos): https://github.com/kata-containers/kata-containers/issues/new Then, send a mail to the mailing list referring to the issue and ideally present your idea at the Architecture Committee meeting so we can get more input from the entire community. @jodh-intel Good suggestion!I will collect the detail background and estimate the potential business impact for android image in kata container firstly and then prepare some material for the this user scenario. thanks for your inputs! I will try the ways provided by you to make sure the feasibility of android in Kata container in next days. Thanks @zhiminghufighting. @zhiminghufighting - I realize this is pretty stale now, but was pretty curious about the Android in Kata use case. Have you made more progress here, or have an end-to-end demo in this space?
2025-04-01T06:39:15.997901
2021-07-27T01:29:13
953416232
{ "authors": [ "KFlash", "aladdin-add" ], "license": "ISC", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7462", "repo": "kataw/kataw", "url": "https://github.com/kataw/kataw/pull/162" }
gharchive/pull-request
chore: add linting as public API as requested by @aladdin-add this PR implements linting as public API. Linting can now be done like this import { lintModule, aladdin } from 'kataw'; lintModule('eval', /* reporter */ aladdin, { noEval: true}); or import { lintScript, aladdin } from 'kataw'; lintScript('eval', /* reporter */ aladdin, { noEval: true}); The 'reporter' argument gives the option to add any reporters and end-users can make their own reporter. I tried to figure out how Babel does this, but I failed in my research. Note This PR is blocked by #160 This is how it looks like on the command line with the aladdin reporter @aladdin-add I'm using the aladdin reporter by default here so this can't be merged before you have merged #160. In the future we will add a 3rd arg into this so the end-user can choose which reporter to use. You also requested for custom rules, right? They are coming soon. ESLint is horrible slow you know ;) lintModule('eval', /* reporter */ aladdin, { noEval: true}); the 2nd param can be an linterOptions: { reporter: ..., globals: {...}, fix: true, } @aladdin-add Good idea! But what is Globals in this context? And can you fix the conflict? @aladdin-add I changed it into what you suggested, but now we got an performance issue ... Every time we invoke the reporter we need to do 'options.reporter(...)'. It's unnecessary property access. Can you get rid of this unnecessary property access?
2025-04-01T06:39:16.024614
2020-07-18T14:12:47
660173422
{ "authors": [ "razttt", "rt-2" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7463", "repo": "katursis/Pawn.RakNet", "url": "https://github.com/katursis/Pawn.RakNet/issues/36" }
gharchive/issue
Server crashing on player connect Same as https://github.com/urShadow/Pawn.RakNet/issues/35 I can see the server, but as soon as I connect, the server crashes. There is NOTHING in the log at the moment of the crash (Not even the 'incoming connection' message). I simply removed the plugin in server.cfg and in the script and everything worked fine instantly. I will have to make more tests when I will have a couple of hours to debug this I will try to see what other includes/plugins might be interfering. content of serverlog.txt : ---------- Loaded log file: "server_log.txt". ---------- SA-MP Dedicated Server ---------------------- v0.3.7-R2, (C)2005-2015 SA-MP Team [2020-07-18 09:40:19] [2020-07-18 09:40:19] Server Plugins [2020-07-18 09:40:19] -------------- [2020-07-18 09:40:19] Loading plugin: pawnraknet [2020-07-18 09:40:19] [Pawn.RakNet] | Pawn.RakNet 1.4.1 | 2016 - 2020 |-------------------------------- | Author and maintainer: urShadow | Compiled: Jun 18 2020 at 14:19:11 |-------------------------------------------------------------- | Forum thread: https://forum.sa-mp.com/showthread.php?t=640306 |-------------------------------------------------------------- | Repository: https://github.com/urShadow/Pawn.RakNet |-------------------------------------------------------------- | Wiki: https://github.com/urShadow/Pawn.RakNet/wiki [2020-07-18 09:40:19] Loaded. [2020-07-18 09:40:19] Loading plugin: crashdetect [2020-07-18 09:40:19] CrashDetect plugin 4.19 [2020-07-18 09:40:19] Loaded. [2020-07-18 09:40:19] Loading plugin: mysql [2020-07-18 09:40:19] >> plugin.mysql: R39-6 successfully loaded. [2020-07-18 09:40:19] Loaded. [2020-07-18 09:40:19] Loading plugin: streamer [2020-07-18 09:40:19] *** Streamer Plugin v2.9.3 by Incognito loaded *** [2020-07-18 09:40:19] Loaded. [2020-07-18 09:40:19] Loading plugin: FileFunctions [2020-07-18 09:40:19] Loaded. [2020-07-18 09:40:19] Loading plugin: FCNPC [2020-07-18 09:40:19] [2020-07-18 09:40:19] ------------------------------------------------- [2020-07-18 09:40:19] FCNPC - Fully Controllable NPC v1.8.2 [2020-07-18 09:40:19] Windows SA-MP 0.3.7 R2 [2020-07-18 09:40:19] Jan 8 2018 at 01:14:48 [2020-07-18 09:40:19] [2020-07-18 09:40:19] Author: OrMisicL (2013 - 2015) [2020-07-18 09:40:19] Continued by: ziggi (2016 - present) [2020-07-18 09:40:19] Contributors: kurta999, Neutralneu [2020-07-18 09:40:19] ------------------------------------------------- [2020-07-18 09:40:20] [2020-07-18 09:40:20] Loading... [2020-07-18 09:40:20] Loaded. [2020-07-18 09:40:20] Loading plugin: PathFinder [2020-07-18 09:40:20] ========================================= [2020-07-18 09:40:20] PathFinder Plugin 1.0 MT [2020-07-18 09:40:20] by Pamdex [2020-07-18 09:40:20] [2020-07-18 09:40:20] Using MapAndreas Plugin 1.2.1 [2020-07-18 09:40:20] Waiting for Init... [2020-07-18 09:40:20] ========================================= [2020-07-18 09:40:20] Loaded. [2020-07-18 09:40:20] Loading plugin: MapAndreas [2020-07-18 09:40:20] Loaded. [2020-07-18 09:40:20] Loading plugin: SKY [2020-07-18 09:40:20] Loaded. [2020-07-18 09:40:20] Loading plugin: YSF [2020-07-18 09:40:20] ARRAY_ConsoleCommands: 4e43d8 [2020-07-18 09:40:20] [2020-07-18 09:40:20] =============================== [2020-07-18 09:40:20] YSF - kurta999's version R19 loaded [2020-07-18 09:40:20] (c) 2008 Alex "Y_Less" Cole - (c) 2010 - 2016 kurta999 [2020-07-18 09:40:20] Server version: 0.3.7 R2-1 [2020-07-18 09:40:20] Operating System: Windows [2020-07-18 09:40:20] Built on: Mar 11 2017 at 10:32:34 [2020-07-18 09:40:20] =============================== [2020-07-18 09:40:20] Loaded. [2020-07-18 09:40:20] Loading plugin: sscanf [2020-07-18 09:40:20] [2020-07-18 09:40:20] =============================== [2020-07-18 09:40:20] sscanf plugin loaded. [2020-07-18 09:40:20] Version: 2.8.2 [2020-07-18 09:40:20] (c) 2012 Alex "Y_Less" Cole [2020-07-18 09:40:20] =============================== [2020-07-18 09:40:20] Loaded. [2020-07-18 09:40:20] Loaded 11 plugins. [2020-07-18 09:40:20] [2020-07-18 09:40:20] Ban list [2020-07-18 09:40:20] -------- [2020-07-18 09:40:20] Loaded: samp.ban [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] Filterscripts [2020-07-18 09:40:20] --------------- [2020-07-18 09:40:20] Loading filterscript 'antiddos.amx'... [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] | YSI version 4.00.0001 | [2020-07-18 09:40:20] | By Alex "Y_Less" Cole | [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] [2020-07-18 09:40:20] [TESTING]: SERVER_LOG_FIRSTCHAR:22; [2020-07-18 09:40:20] Loaded! [2020-07-18 09:40:20] Loading filterscript 'fac_test.amx'... [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] | YSI version 4.00.0001 | [2020-07-18 09:40:20] | By Alex "Y_Less" Cole | [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] [2020-07-18 09:40:20] -black screens initialization... [2020-07-18 09:40:20] -players variables [2020-07-18 09:40:20] -black screens initialization completed. [2020-07-18 09:40:20] -Loading Testing Faction... [2020-07-18 09:40:20] [MYSQL]: Connection to `saarp` succesful! [2020-07-18 09:40:20] -Loading Objects... [2020-07-18 09:40:20] -Loading Pickups... [2020-07-18 09:40:20] -Testing Faction loaded correctly... [2020-07-18 09:40:20] Loading filterscript 'vehicles.amx'... [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] | YSI version 4.00.0001 | [2020-07-18 09:40:20] | By Alex "Y_Less" Cole | [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] [2020-07-18 09:40:20] -black screens initialization... [2020-07-18 09:40:20] -players variables [2020-07-18 09:40:20] -black screens initialization completed. [2020-07-18 09:40:20] ----------------------------------------- [2020-07-18 09:40:20] Stefan/Kevin974 - Speedometer | rt-2 - Fuel/Engine system [2020-07-18 09:40:20] ----------------------------------------- [2020-07-18 09:40:20] [VEHICLES]: Setting vehicle Interior positions [2020-07-18 09:40:20] [VEHICLES]: Setting vehicle Interior external doors positions [2020-07-18 09:40:20] [VEHICLES]: SA Driving Assoc [2020-07-18 09:40:20] [VEHICLES]: Spawning cars from database [2020-07-18 09:40:20] Loading filterscript 'doors.amx'... [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] | YSI version 4.00.0001 | [2020-07-18 09:40:20] | By Alex "Y_Less" Cole | [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] [2020-07-18 09:40:20] -black screens initialization... [2020-07-18 09:40:20] -players variables [2020-07-18 09:40:20] -black screens initialization completed. [2020-07-18 09:40:20] Loading City Planning faction and adresses... [2020-07-18 09:40:20] -Connecting to database [2020-07-18 09:40:20] -Acquiring vehicles coords [2020-07-18 09:40:20] -Initializing doors vars [2020-07-18 09:40:20] -Other variables and timers [2020-07-18 09:40:20] Loading filterscript 'public.amx'... [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] | YSI version 4.00.0001 | [2020-07-18 09:40:20] | By Alex "Y_Less" Cole | [2020-07-18 09:40:20] | | [2020-07-18 09:40:20] ======================================= [2020-07-18 09:40:20] [2020-07-18 09:40:21] -black screens initialization... [2020-07-18 09:40:21] -players variables [2020-07-18 09:40:21] -black screens initialization completed. [2020-07-18 09:40:21] # # # # # # # # # # # # # # # # # # # # # # [2020-07-18 09:40:21] # # # # The city have a PT systems. # # # # [2020-07-18 09:40:21] # # # # # # # # # # # # # # # # # # # # # # [2020-07-18 09:40:21] # # Initializing "PT" System [2020-07-18 09:40:21] # # Initializing other plugins [2020-07-18 09:40:21] # # # Initializing FCNPC [2020-07-18 09:40:21] # # # Initializing SAPT System [2020-07-18 09:40:21] # # # Loading SAPT Routes [2020-07-18 09:40:21] Loading filterscript 'testing.amx'... [2020-07-18 09:40:21] Loaded 6 filterscripts. [2020-07-18 09:40:22] Filterscript '../scriptfiles/callbackfix.amx' loaded. [2020-07-18 09:40:22] [2020-07-18 09:40:22] [2020-07-18 09:40:22] [2020-07-18 09:40:22] ======================================= [2020-07-18 09:40:22] | | [2020-07-18 09:40:22] | YSI version 4.00.0001 | [2020-07-18 09:40:22] | By Alex "Y_Less" Cole | [2020-07-18 09:40:22] | | [2020-07-18 09:40:22] ======================================= [2020-07-18 09:40:22] [2020-07-18 09:40:25] -AC(new) initialization... [2020-07-18 09:40:25] -players variables [2020-07-18 09:40:25] -vehicles variables [2020-07-18 09:40:25] -AC(new) initialization completed. [2020-07-18 09:40:25] -black screens initialization... [2020-07-18 09:40:25] -players variables [2020-07-18 09:40:25] -black screens initialization completed. [2020-07-18 09:40:25] -inventories initialization... [2020-07-18 09:40:25] -general variables [2020-07-18 09:40:25] -players variables [2020-07-18 09:40:25] -inventories initialization completed. [2020-07-18 09:40:26] -Logged Off Players Variables initializations... [2020-07-18 09:40:26] -melee interactions initialization... [2020-07-18 09:40:26] -players variables [2020-07-18 09:40:26] -all players variables [2020-07-18 09:40:26] -melee interactions initialization completed. [2020-07-18 09:40:26] -Mailboxs initializations... [2020-07-18 09:40:26] -SASD faction initializing [2020-07-18 09:40:26] -Pickup(s) [2020-07-18 09:40:26] -Setting routes [2020-07-18 09:40:26] -SAPO faction initializing [2020-07-18 09:40:26] -Pickup(s) [2020-07-18 09:40:26] -Setting routes [2020-07-18 09:40:26] -SATEL faction initializing [2020-07-18 09:40:26] -Pickup(s) [2020-07-18 09:40:26] -Setting routes [2020-07-18 09:40:26] -SPRU faction initializing [2020-07-18 09:40:26] -Pickup(s) [2020-07-18 09:40:26] -Setting routes [2020-07-18 09:40:26] -Missions initializations... [2020-07-18 09:40:26] -Initializing player status. [2020-07-18 09:40:26] -Initializing timer(s). [2020-07-18 09:40:26] -Initializing text strings. [2020-07-18 09:40:26] -Gang Wars initializations... [2020-07-18 09:40:26] [GANGWAR]: -Initializing Gang War system. [2020-07-18 09:40:26] [GANGWAR]: -Initializing Areas. [2020-07-18 09:40:26] [GANGWAR]: -Initializing Gangs Skins. [2020-07-18 09:40:26] [GANGWAR]: -Initializing Gangs Cribs. [2020-07-18 09:40:26] [GANGWAR]: -Initializing Gangs Pickups. [2020-07-18 09:40:26] [GANGWAR]: -Initializing Territories Wars Timer. [2020-07-18 09:40:26] [GANGWAR]: -Initializing Other Gangs Vars. [2020-07-18 09:40:26] -Interactive Menus initializations... [2020-07-18 09:40:26] -Shops initializations... [2020-07-18 09:40:26] -Sending query [2020-07-18 09:40:26] -tutorial initialization... [2020-07-18 09:40:26] -players variables [2020-07-18 09:40:26] -office pickup [2020-07-18 09:40:26] -slides variables [2020-07-18 09:40:26] -Transport system initializations... [2020-07-18 09:40:26] -Impex Automatic Sales initialization; [2020-07-18 09:40:26] -stockid is 53 after load [2020-07-18 09:40:26] -creating container objects; [2020-07-18 09:40:26] -creating pickup; [2020-07-18 09:40:26] -creating container variables; [2020-07-18 09:40:26] -spawning cargos; [2020-07-18 09:40:26] -digestion initialization... [2020-07-18 09:40:26] -players variables [2020-07-18 09:40:26] -digestion initialization completed. [2020-07-18 09:40:26] -machines initialization... [2020-07-18 09:40:26] -players variables [2020-07-18 09:40:26] -machines variables [2020-07-18 09:40:26] -machines recipes variables [2020-07-18 09:40:26] -machine initialization completed. [2020-07-18 09:40:26] -fire initialization... [2020-07-18 09:40:26] -vehicles variables [2020-07-18 09:40:26] -previous fires [2020-07-18 09:40:26] -fire initialization completed. [2020-07-18 09:40:26] -underground races initialization... [2020-07-18 09:40:26] -general variables [2020-07-18 09:40:26] [TESTING]: urace_racevar_reinit() called. [2020-07-18 09:40:26] -racers variables [2020-07-18 09:40:26] -players variables [2020-07-18 09:40:26] -underground races initialization completed. [2020-07-18 09:40:26] -fishing initialization... [2020-07-18 09:40:26] -catch types [2020-07-18 09:40:26] -players variables [2020-07-18 09:40:26] -fishing initialization completed. [2020-07-18 09:40:26] -elevators initialization... [2020-07-18 09:40:26] -loading elevators... [2020-07-18 09:40:26] -elevators initialization completed. [2020-07-18 09:40:26] -Scripted Casinos initializations... [2020-07-18 09:40:26] ---------------------------------- [2020-07-18 09:40:26] [2020-07-18 09:40:26] STREAMER_OBJECT_SD =<PHONE_NUMBER> [2020-07-18 09:40:26] STREAMER_OBJECT_DD = 0 [2020-07-18 09:40:26] NB OF OBJ LOADED: 5126/8000 [2020-07-18 09:40:26] ---------------------------------- [2020-07-18 09:40:26] -removed buildings initialization... [2020-07-18 09:40:26] -vars initializing... [2020-07-18 09:40:26] -building listing... [2020-07-18 09:40:26] -testing configuration: [2020-07-18 09:40:26] -679 building removed; [2020-07-18 09:40:26] -modelid:923 is removed 3 times: [2020-07-18 09:40:26] -modelid:1216 is removed 14 times: [2020-07-18 09:40:26] -modelid:'traffic light'(1283) is removed 239 times: [2020-07-18 09:40:26] -modelid:1284 is removed 17 times: [2020-07-18 09:40:26] -modelid:1315 is removed 62 times: [2020-07-18 09:40:26] -modelid:1350 is removed 20 times: [2020-07-18 09:40:26] -modelid:1373 is removed 8 times: [2020-07-18 09:40:26] -modelid:1374 is removed 8 times: [2020-07-18 09:40:26] -modelid:'interior boxes'(1421) is removed 2 times: [2020-07-18 09:40:26] -modelid:'interior boxes'(1431) is removed 2 times: [2020-07-18 09:40:26] -modelid:1440 is removed 2 times: [2020-07-18 09:40:26] -modelid:1441 is removed 2 times: [2020-07-18 09:40:26] -modelid:2647 is removed 7 times: [2020-07-18 09:40:26] -modelid:2663 is removed 6 times: [2020-07-18 09:40:26] -modelid:2672 is removed 4 times: [2020-07-18 09:40:26] -modelid:2673 is removed 2 times: [2020-07-18 09:40:26] -modelid:2674 is removed 2 times: [2020-07-18 09:40:26] -modelid:2675 is removed 3 times: [2020-07-18 09:40:26] -modelid:2676 is removed 3 times: [2020-07-18 09:40:26] -modelid:2677 is removed 3 times: [2020-07-18 09:40:26] -modelid:3377 is removed 8 times: [2020-07-18 09:40:26] -modelid:3378 is removed 8 times: [2020-07-18 09:40:26] -modelid:3474 is removed 2 times: [2020-07-18 09:40:26] -modelid:3516 is removed 4 times: [2020-07-18 09:40:26] -modelid:3567 is removed 5 times: [2020-07-18 09:40:26] -modelid:3569 is removed 5 times: [2020-07-18 09:40:26] -modelid:'cargos'(3574) is removed 32 times: [2020-07-18 09:40:26] -modelid:'abandonned car'(3593) is removed 11 times: [2020-07-18 09:40:26] -modelid:'abandonned car'(3594) is removed 10 times: [2020-07-18 09:40:26] -modelid:3621 is removed 5 times: [2020-07-18 09:40:26] -modelid:3625 is removed 7 times: [2020-07-18 09:40:26] -modelid:3664 is removed 4 times: [2020-07-18 09:40:26] -modelid:3665 is removed 3 times: [2020-07-18 09:40:26] -modelid:3688 is removed 5 times: [2020-07-18 09:40:26] -modelid:3744 is removed 32 times: [2020-07-18 09:40:26] -modelid:3747 is removed 5 times: [2020-07-18 09:40:26] -modelid:3769 is removed 6 times: [2020-07-18 09:40:26] -modelid:3780 is removed 3 times: [2020-07-18 09:40:26] -Administration initializations... [2020-07-18 09:40:26] -SAARP_fac_cityp_tcmd initializations... [2020-07-18 09:40:26] [2020-07-18 09:40:26] -------------------------------------- [2020-07-18 09:40:26] Anticheat Nex-AC loaded! [2020-07-18 09:40:26] Anticheat version: 1.9.53 [2020-07-18 09:40:26] Author: Nexius [2020-07-18 09:40:26] -------------------------------------- [2020-07-18 09:40:26] -Initializing principal game mode. [2020-07-18 09:40:26] -Mysql connection. [2020-07-18 09:40:26] -Setting Variables. [2020-07-18 09:40:26] -Preparing streamer plugin... [2020-07-18 09:40:26] -tick rate: 50; [2020-07-18 09:40:26] -max pickups: -1; [2020-07-18 09:40:26] -cell distance:<PHONE_NUMBER>; [2020-07-18 09:40:26] -cell size: 300.000000; [2020-07-18 09:40:26] -Preparing Main Textdraws... [2020-07-18 09:40:26] -Spawning map icons... [2020-07-18 09:40:26] -Plants initializations... [2020-07-18 09:40:26] -Races initializations... [2020-07-18 09:40:26] -Blood Stains initializations... [2020-07-18 09:40:26] -Cheats initializations... [2020-07-18 09:40:26] -Noob Path Helper initializations... [2020-07-18 09:40:26] -Death Reasons initializations... [2020-07-18 09:40:26] -Objects initializations... [2020-07-18 09:40:26] -Acquiring object types [2020-07-18 09:40:26] -Spawning ground objects models [2020-07-18 09:40:26] -Spawning posts [2020-07-18 09:40:26] -Vehicles interior initializations... [2020-07-18 09:40:26] -Acquiring vehicles coords [2020-07-18 09:40:26] -Spawning vehicle interior pickups [2020-07-18 09:40:26] -Dialog Menus initializations... [2020-07-18 09:40:26] -LS Stadiums initializations... [2020-07-18 09:40:26] -Loading Police Forces faction [2020-07-18 09:40:26] -Loading objects [2020-07-18 09:40:26] -Loading pickups [2020-07-18 09:40:26] [TESTING]: LSPD_Weapons_Pickup = 89 [2020-07-18 09:40:26] [TESTING]: SAMA_Armour_Pickup = 94 [2020-07-18 09:40:26] -Taxi faction initializing [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Loading SF Airport Management faction [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Loading LS International Airport faction [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Loading SF Military Police faction [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Loading LS Military Police faction. [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Loading SA Anti Terrorists faction. [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Spawning objects [2020-07-18 09:40:26] -Mush faction initializing [2020-07-18 09:40:26] -Spawning objects [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Mush faction initializing [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -City Planning faction initializing [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Spawning objects [2020-07-18 09:40:26] -Gouv faction initializing (SADA/SAAA/SABA/SACB) [2020-07-18 09:40:26] -Spawning objects [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] -Xoomer faction initializing [2020-07-18 09:40:26] -Spawning pickups [2020-07-18 09:40:26] [TESTING]: LSPD_Weapons_Pickup = 89 [2020-07-18 09:40:26] -IG Player Variables initializations... [2020-07-18 09:40:26] -Vehicle Variables initializations... [2020-07-18 09:40:26] -Other Variables initializations... [2020-07-18 09:40:26] -Principal game mode initializing completed. [2020-07-18 09:40:26] Game mode ready! [2020-07-18 09:40:26] Number of vehicle models: 0 [2020-07-18 09:40:26] [SERVER]: Vehicles are loading... [2020-07-18 09:40:26] [MYSQL]: Receiving response on "SpawnVehicles", 861 rows. [2020-07-18 09:40:36] [VEHICLES]: Vehicle spawned, 861 vehicles. [2020-07-18 09:40:36] Spawning errors: -Vehicule id:1240(m:588) does not have a vehicle interior view associated!; -Vehicule id:1250(m:558) does not have a vehicle interior view associated!; [2020-07-18 09:40:36] [MYSQL]: Receiving response on "getZonesFromDatabase", 43 rows. [2020-07-18 09:40:36] [MYSQL]: Receiving response on "SpawnGazstations", 33 rows. [2020-07-18 09:40:36] [DOORS]: Doors are loading... [2020-07-18 09:40:36] [MYSQL]: Receiving response on "LoadDoors", 1486 rows. [2020-07-18 09:40:46] [DOORS]: loaded 16/20 map icons. [2020-07-18 09:40:46] [MYSQL]: Receiving response on "LoggedOff_InitAllAtStart_res", 857 rows. [2020-07-18 09:40:46] -All Player Variables initializations... [2020-07-18 09:40:51] [MYSQL]: Receiving response on "getAreasFromDatabase", 43 rows. [2020-07-18 09:40:51] [MYSQL]: Receiving response on "ReceiveShopsList", 157 rows. [2020-07-18 09:40:52] [MYSQL]: Receiving response on "GetGroundCargos", 193 rows. [2020-07-18 09:40:52] [MYSQL]: Receiving response on "Machines_LoadFromDB", 16 rows. [2020-07-18 09:40:53] [MYSQL]: Receiving response on "elevators_loadFromDatabase", 2 rows. [2020-07-18 09:40:53] [MYSQL]: Receiving response on "LoadSAAGFields", 11 rows. [2020-07-18 09:40:53] [MYSQL]: Receiving response on "LoadSAAGPlants", 73 rows. [2020-07-18 09:40:53] [MYSQL]: Receiving response on "ReceiveGovOffices", 69 rows. [2020-07-18 09:40:53] [MYSQL]: Receiving response on "ReceiveObjectTypes", 66 rows. [2020-07-18 09:40:53] [MYSQL]: Receiving response on "SpawnGroundObjects", 4770 rows. [2020-07-18 09:41:03] [MYSQL]: Receiving response on "SpawnPosts", 4 rows. [2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(0)", 24 rows. [2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(1)", 61 rows. [2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(2)", 20 rows. [2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(3)", 189 rows. [2020-07-18 09:41:04] ## SASD ROUTE TOO LARGE ## ## will truncate r:1,t:fac_sasd_jy_lsscres from 189 to 65 ## [2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(4)", 10 rows. [2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(5)", 26 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(6)", 2 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(7)", 150 rows. [2020-07-18 09:41:05] ## SASD ROUTE TOO LARGE ## ## will truncate r:2,t:fac_sasd_jy_lsnebiz from 150 to 65 ## [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(8)", 0 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(9)", 31 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(10)", 19 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(11)", 12 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(12)", 9 rows. [2020-07-18 09:41:05] [TESTING]: Finishing route, thisroute_max_leg:14, thisroute_total_leg:16 [2020-07-18 09:41:05] [TESTING]: Finishing route, thisroute_max_leg:5, thisroute_total_leg:7 [2020-07-18 09:41:05] [TESTING]: Finishing route, thisroute_max_leg:5, thisroute_total_leg:7 [2020-07-18 09:41:05] [MYSQL]: Receiving response on "elevator_floor_loadFromDatabase", 21 rows. [2020-07-18 09:41:05] [MYSQL]: Receiving response on "elevator_floor_loadFromDatabase", 11 rows. [2020-07-18 09:41:05] [SERVER]: The time is 9:41 [2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:247, toggle:1) called. [2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:318, toggle:1) called. [2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:357, toggle:1) called. [2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:418, toggle:1) called. [2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:461, toggle:1) called. [2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:559, toggle:1) called. [2020-07-18 09:41:28] [SERVER]: The time is 9:41 [2020-07-18 09:41:51] [SERVER]: The time is 9:41 [2020-07-18 09:42:13] [SERVER]: The time is 9:42 [2020-07-18 09:42:36] [SERVER]: Executing 'serverStart1'. [2020-07-18 09:42:36] [2020-07-18 09:42:36] [2020-07-18 09:42:36] [2020-07-18 09:42:36] ======================================= [2020-07-18 09:42:36] | | [2020-07-18 09:42:36] | YSI version 4.00.0001 | [2020-07-18 09:42:36] | By Alex "Y_Less" Cole | [2020-07-18 09:42:36] | | [2020-07-18 09:42:36] ======================================= [2020-07-18 09:42:36] [2020-07-18 09:42:36] -black screens initialization... [2020-07-18 09:42:36] -players variables [2020-07-18 09:42:36] -black screens initialization completed. [2020-07-18 09:42:36] ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ [2020-07-18 09:42:36] ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ The state is being infected ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ [2020-07-18 09:42:36] ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ ร‚ยฐ [2020-07-18 09:42:36] ร‚ยฐ ร‚ยฐ Initializing "Infected" System [2020-07-18 09:42:36] ร‚ยฐ ร‚ยฐ Initializing other plugins [2020-07-18 09:42:36] ร‚ยฐ ร‚ยฐ ร‚ยฐ Initializing Map Andreas [2020-07-18 09:42:37] PathFinder Plugin -> Creating New Thread [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ Initializing FCNPC [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ Initializing PathFinder [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ Generating spawn loacations. [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ initializing infected [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ initializing reserves [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ initializing timers [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ ร‚ยฐ initializing SQL connection [2020-07-18 09:42:37] ร‚ยฐ ร‚ยฐ Script initialized [2020-07-18 09:42:37] Filterscript 'infected.amx' loaded. [2020-07-18 09:42:37] ---------------------------------- [2020-07-18 09:42:37] GOUVERNEMENT OBJECT SCRIPT [2020-07-18 09:42:37] nz = 12.557999, stream is [2020-07-18 09:42:37] NB OF OBJ LOADED: 4338/X [2020-07-18 09:42:37] ---------------------------------- [2020-07-18 09:42:37] Filterscript 'mobjects.amx' loaded. [2020-07-18 09:42:37] [2020-07-18 09:42:37] [2020-07-18 09:42:37] [2020-07-18 09:42:37] ======================================= [2020-07-18 09:42:37] | | [2020-07-18 09:42:37] | YSI version 4.00.0001 | [2020-07-18 09:42:37] | By Alex "Y_Less" Cole | [2020-07-18 09:42:37] | | [2020-07-18 09:42:37] ======================================= [2020-07-18 09:42:37] [2020-07-18 09:42:37] -black screens initialization... [2020-07-18 09:42:37] -players variables [2020-07-18 09:42:37] -black screens initialization completed. [2020-07-18 09:42:37] -LSPD Tickets initialization... [2020-07-18 09:42:37] -Loading Testing1 Faction... [2020-07-18 09:42:37] -Loading Testing1337 Faction... [2020-07-18 09:42:37] [MYSQL]: Connection to `saarp` succesful! [2020-07-18 09:42:37] -Loading Objects... [2020-07-18 09:42:37] -Loading Areas... [2020-07-18 09:42:37] -Loading Vehicles vars... [2020-07-18 09:42:37] -Testing1 Faction loaded correctly... [2020-07-18 09:42:37] Filterscript 'fac_test1.amx' loaded. [2020-07-18 09:42:37] [2020-07-18 09:42:37] [2020-07-18 09:42:37] [2020-07-18 09:42:37] ======================================= [2020-07-18 09:42:37] | | [2020-07-18 09:42:37] | YSI version 4.00.0001 | [2020-07-18 09:42:37] | By Alex "Y_Less" Cole | [2020-07-18 09:42:37] | | [2020-07-18 09:42:37] ======================================= [2020-07-18 09:42:37] [2020-07-18 09:42:38] -black screens initialization... [2020-07-18 09:42:38] -players variables [2020-07-18 09:42:38] -black screens initialization completed. [2020-07-18 09:42:38] -setting BIOS environments [2020-07-18 09:42:38] -computers variables [2020-07-18 09:42:38] -players variables [2020-07-18 09:42:38] -computer initialization... [2020-07-18 09:42:38] -computer initialization completed. [2020-07-18 09:42:38] -tele-comunication system initializations... [2020-07-18 09:42:38] -Phone Objects adjustments [2020-07-18 09:42:38] -Phone Variables [2020-07-18 09:42:38] -Starting timers [2020-07-18 09:42:38] -Loading Test2 Script... [2020-07-18 09:42:38] -Loading Map Andreas... [2020-07-18 09:42:38] [MYSQL]: Connection to `saarp` succesful! [2020-07-18 09:42:38] -Test2 Script loaded correctly... [2020-07-18 09:42:38] Filterscript 'fac_test2.amx' loaded. [2020-07-18 09:42:38] [MYSQL]: Receiving response on "cptr_LoadFromDatabase", 3 rows. [2020-07-18 09:42:38] [MYSQL]: Receiving response on "GetDbPhones", 46 rows. [2020-07-18 09:42:38] [SERVER]: The time is 9:42 [2020-07-18 09:42:41] [SERVER]: Executing 'serverStart2'. [2020-07-18 09:42:41] Server password has been removed. [2020-07-18 09:42:41] Filterscript 'testing.amx' unloaded. [2020-07-18 09:42:44] [INFECTED]: Infected must be spawned. [2020-07-18 09:42:44] [npc:join] Infected has joined the server (203:<IP_ADDRESS>) [2020-07-18 09:42:44] [INFECTED]: Connecting player is infected. [2020-07-18 09:42:52] [INFECTED]: Infected must be spawned. [2020-07-18 09:42:52] [npc:join] Infected has joined the server (202:<IP_ADDRESS>) [2020-07-18 09:42:52] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:01] [SERVER]: The time is 9:43 [2020-07-18 09:43:01] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:01] [npc:join] Infected has joined the server (201:<IP_ADDRESS>) [2020-07-18 09:43:01] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:09] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:09] [npc:join] Infected has joined the server (200:<IP_ADDRESS>) [2020-07-18 09:43:09] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:18] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:18] [npc:join] Infected has joined the server (199:<IP_ADDRESS>) [2020-07-18 09:43:18] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:24] [SERVER]: The time is 9:43 [2020-07-18 09:43:26] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:26] [npc:join] Infected has joined the server (198:<IP_ADDRESS>) [2020-07-18 09:43:26] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:35] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:35] [npc:join] Infected has joined the server (197:<IP_ADDRESS>) [2020-07-18 09:43:35] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:43] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:43] [npc:join] Infected has joined the server (196:<IP_ADDRESS>) [2020-07-18 09:43:43] [INFECTED]: Connecting player is infected. [2020-07-18 09:43:44] [TESTING]: RefreshDoorMapIcons() called (Initializer, interval=90). [2020-07-18 09:43:46] [SERVER]: The time is 9:43 [2020-07-18 09:43:51] [INFECTED]: Infected must be spawned. [2020-07-18 09:43:51] [npc:join] Infected has joined the server (195:<IP_ADDRESS>) [2020-07-18 09:43:51] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:00] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:00] [npc:join] Infected has joined the server (194:<IP_ADDRESS>) [2020-07-18 09:44:00] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:08] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:08] [npc:join] Infected has joined the server (193:<IP_ADDRESS>) [2020-07-18 09:44:08] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:09] [SERVER]: The time is 9:44 [2020-07-18 09:44:17] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:17] [npc:join] Infected has joined the server (192:<IP_ADDRESS>) [2020-07-18 09:44:17] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:25] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:25] [npc:join] Infected has joined the server (191:<IP_ADDRESS>) [2020-07-18 09:44:25] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:31] [SERVER]: The time is 9:44 [2020-07-18 09:44:33] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:33] [npc:join] Infected has joined the server (190:<IP_ADDRESS>) [2020-07-18 09:44:33] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:41] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:41] [npc:join] Infected has joined the server (189:<IP_ADDRESS>) [2020-07-18 09:44:42] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:50] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:50] [npc:join] Infected has joined the server (188:<IP_ADDRESS>) [2020-07-18 09:44:50] [INFECTED]: Connecting player is infected. [2020-07-18 09:44:54] [SERVER]: The time is 9:44 [2020-07-18 09:44:58] [INFECTED]: Infected must be spawned. [2020-07-18 09:44:58] [npc:join] Infected has joined the server (187:<IP_ADDRESS>) [2020-07-18 09:44:58] [INFECTED]: Connecting player is infected. [2020-07-18 09:45:06] [INFECTED]: Infected must be spawned. [2020-07-18 09:45:06] [npc:join] Infected has joined the server (186:<IP_ADDRESS>) [2020-07-18 09:45:06] [INFECTED]: Connecting player is infected. [2020-07-18 09:45:15] [INFECTED]: Infected must be spawned. [2020-07-18 09:45:15] [npc:join] Infected has joined the server (185:<IP_ADDRESS>) [2020-07-18 09:45:15] [INFECTED]: Connecting player is infected. [2020-07-18 09:45:16] [SERVER]: The time is 9:45 [2020-07-18 09:45:23] [INFECTED]: Infected must be spawned. [2020-07-18 09:45:23] [npc:join] Infected has joined the server (184:<IP_ADDRESS>) Thank you rt-2 I fixed this problem by experimenting with the include orders. Here is what I done: I moved the include from line 26, now to almost the bottom of my includes. Note that I could not place it before "fixes.inc", otherwise, "fixes.inc" would give me an ALS error. Thank you, rt-2 I fixed this problem by experimenting with the include orders. Here is what I done: I moved the include from line 26, now to almost the bottom of my includes. Note that I could not place it before "fixes.inc", otherwise, "fixes.inc" would give me an ALS error. Thank you, rt-2 this fix dont work for me, do u try another?
2025-04-01T06:39:16.029184
2020-04-21T13:43:11
604009534
{ "authors": [ "aaronbriel", "kaushaltrivedi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7464", "repo": "kaushaltrivedi/fast-bert", "url": "https://github.com/kaushaltrivedi/fast-bert/pull/205" }
gharchive/pull-request
added call to convert posix path to string to fix TypeError in save_pโ€ฆ โ€ฆretrained call. Fixes Issue #200 Merged. Thanks.
2025-04-01T06:39:16.031088
2024-12-20T13:23:35
2752738033
{ "authors": [ "mwestphal", "theusst" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7465", "repo": "kaust-vislab/MidSurfer", "url": "https://github.com/kaust-vislab/MidSurfer/issues/2" }
gharchive/issue
Ship a binary version of the plugin Since midsurfer only depends on ParaView, creating (and shipping) a binary version of this plugin should be possible by using the tools developed by Kitware: https://github.com/Kitware/paraview-ci-example Let me know what you think Dear Mathieu, Thank you for the suggestion! This is definitely on my list. The paper is currently under review, we will decide on how to proceed once we receive the reviews. I will keep this issue open for now. Thomas
2025-04-01T06:39:16.033537
2019-05-28T17:34:22
449392464
{ "authors": [ "CyrilDebon", "Soviut" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7466", "repo": "kavalcante/vue-truncate-collapsed", "url": "https://github.com/kavalcante/vue-truncate-collapsed/pull/14" }
gharchive/pull-request
FEATURE property to control truncation state This allows the truncation to be expanded or collapsed via a property. Contracted: <truncate :truncated="true">...</truncate> Expanded: <truncate :truncated="false">...</truncate> Bound: <truncate :truncated="isExpanded">...</truncate> data() { isExpanded: true } NOTE: I'm unsure if the property should be named to something truthy by default, like expanded so that true is expanded and false is collapsed Could someone can merge this ? I need it haha Or did @Soviut did u found an alternative solution ?
2025-04-01T06:39:16.043818
2024-03-17T13:28:32
2190672499
{ "authors": [ "ndepomereu", "palexdev" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7467", "repo": "kawansoft/SympleGit-Java", "url": "https://github.com/kawansoft/SympleGit-Java/issues/1" }
gharchive/issue
Some feedback... Hello, I just wanted to leave some feedback on this library. Context I'm developing a tool which needs Git functionalities. In particular, but not limited to, it has to clone a huge amount of repositories from different sources. I started by using JGit but it's so bad it's almost ridiculous. Very often, clone operations are super slow and would end with a EOF exception. Of course, it does not happen with the native command. Switching to SympleGit Let's say that the README is quite catchy: However, JGit's API comes with a learning curve and lacks direct, one-to-one support for CLI actions. Therefore, SympleGit is likely to be a more straightforward option for simple Git integration in many Java projects, particularly those utilizing basic Git functionalities. Let's delve into the details! In my honest opinion, it cannot be more false. JGit is super easy to use, at least in my use case, every git command is a class in JGit. In SympleGit the clone command class does not exist. The great thing about SympleGit is that it uses the native command to perform operations, which is so much faster and performant. The bad thing, again in my opinion, is that the API is not so well thought, and it's lacking. The idea of creating custom git commands with executeGitCommand(...) surely is good to cover all cases even those that are not implemented yet, but still having them implemented as classes would make everything much easier to use. It's so confusing to do this: SympleGit sympleGit = SympleGit.custom() .setDirectory(repoDirectoryPath) .build(); Why do I have to give the directory here instead of giving it to the command directly? Like this for JGit: CloneCommand cmd = new CloneCommand() .setDirectory(destDir.resolve(path).toFile()) .setURI(url) .setRemote(remote) .setBranch(branch); // In this case, destDir is the base dir in which I want to store all the projects // path is the where I want to clone the repository, the last part of the path will be the name of the directory I find it a bit more intuitive 3) I read the README multiple times, but I still don't know how I can track the progress of a command. Probably because, yet again, it's not very intuitive. By nature, working with Process and ProcessBuilder in Java is a cumbersome task, processing the output of a process properly is hard. And for this very reason, a library that uses such APIs should make it as easy as possible for the end user to use it. The first thing that comes to my mind when I want to track the progress of an external process is something like this: // I get the why of the check... if (! gitCommander.isResponseOk()) { System.out.println("An Error Occured: " + gitCommander.getProcessError()); return; } // Then... while (command.output ...) { // print output } JGit here is doing a far better job here, I can modify the above command like this: CloneCommand cmd = new CloneCommand() .setProgressMonitor(new TextProgressMonitor()) // And boom, I automatically have output to the console .setDirectory(destDir.resolve(path).toFile()) .setURI(url) .setRemote(remote) .setBranch(branch); // Not only that, I can even make custom monitor implementations Conclusion I believe SympleGit could become a very good alternative to JGit, but it definitely needs to grow, improve and expand the API. In the meantime, I think I'll implement a custom solution that uses the native command just like SympleGit because it's simply much much better than JGit which crashes all the time. Hi, Sorry for very late reply & thanks for the comments. I was a bid dubitative, about why it's confusing to do that: SympleGit sympleGit = SympleGit.custom() .setDirectory(repoDirectoryPath) .build(); Could you please elaborate a little? I'm not sure of the importance of the issue. About the progress monitor, I clearly understand the need, but can you tel when it's important to set up monitor? Maybe my projects are not big enough, I never had the need... Of course, we will do it cleanly, extendable like you wish. Best, N.
2025-04-01T06:39:16.048738
2018-05-18T05:07:51
324271876
{ "authors": [ "kayex", "peter279k" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7468", "repo": "kayex/http-codes", "url": "https://github.com/kayex/http-codes/pull/3" }
gharchive/pull-request
Test enhancement Changed log Set the multiple PHPUnit version for the different PHP versions. Add the white filter list in phpunit.xml setting. Thank you for the pull request! ๐Ÿ™‚ Do you mind explaining the reason for these changes? Especially the added PHP version constraint, since it breaks compatibility with PHP 5. I know PHP 5 is reaching its end of life very soon, but I don't see a reason to break compatibility unless it's required. I also think the phpunit version string should be "^6.2 || ^7.0" rather than "^6.2|^7.0", at least according to the Composer docs. As you say, I set the required PHP version at least 7.0+ because the PHP 5 is EOL. How about setting the >=5.6 in composer.json require key? The PHPUnit version is "^6.2 || ^7.0" is the correct defined version description.
2025-04-01T06:39:16.060860
2016-08-26T22:33:03
173556237
{ "authors": [ "ShibiHe", "eggie5" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7470", "repo": "kayzhu/LSHash", "url": "https://github.com/kayzhu/LSHash/issues/15" }
gharchive/issue
projection type The code is using np.random.randn() times input vector. In the LSH paper survey, we are using either (Gaussian Distribution * input + bias)/W or (Uniform Distribution * input). I was wondering if we should change the distribution to uniform in the code? I have seen that Gaussian Random Projection is on way to implement the random projections. What is this LSH paper survey?
2025-04-01T06:39:16.064261
2022-11-30T09:23:04
1469309694
{ "authors": [ "kazurayam" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7471", "repo": "kazurayam/inspectus", "url": "https://github.com/kazurayam/inspectus/issues/40" }
gharchive/issue
Use materialstore-0.13.0-SNAPSHOT or higher The materialstore-0.13.0-SNAPSHOT resolved the issue The pom.xml has no Now most of the external dependencies for the inspectus project can be resolved by Gradle automatically. done at 0.6.0
2025-04-01T06:39:16.066148
2018-04-29T22:51:08
318757159
{ "authors": [ "alexanderreiff" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7472", "repo": "kbacha/stellar-sdk", "url": "https://github.com/kbacha/stellar-sdk/pull/201" }
gharchive/pull-request
Add operations for account to CLI Adds a command to request all operations relating to a given account to the CLI application. Fixes #163 Is there a GIF that reflects how this work made you feel? @correlator @kbacha can you guys take another look at this? Thanks! @kbacha can you take a look at this refactor? Took me a bit of Rust-y try and error to figure out the indent levelโ€“aware nesting, but in the end, the solution I got was pretty simple.
2025-04-01T06:39:16.066985
2024-09-04T13:26:21
2505377078
{ "authors": [ "TomAugspurger", "minrk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7473", "repo": "kbatch-dev/helm-chart", "url": "https://github.com/kbatch-dev/helm-chart/pull/6" }
gharchive/pull-request
add extraPodLabels, hub access label by default ensures network access to the hub with the default hub networkPolicy in the jupyterhub helm chart Thanks!
2025-04-01T06:39:16.117525
2022-04-26T08:15:06
1215578937
{ "authors": [ "MichalMed", "ahmadjana", "blcham", "ledsoft" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7474", "repo": "kbss-cvut/kbss-website", "url": "https://github.com/kbss-cvut/kbss-website/pull/13" }
gharchive/pull-request
Patch 1 this branch for our open mic sessions: It should contains the schedule and contents of the seminars @blcham Does it make sense to have the open mic schedule publicly visible, when the sessions are internal to KBSS? Or is there a plan to make the sessions public? Does it make sense to have the open mic schedule publicly visible, when the sessions are internal to KBSS? Or is there a plan to make the sessions public? Yes, this was the idea, see #7. I would like to publish some of the presentations and possibly invite somebody to join our presentation online. Good thing is to have some static page with programme and link it every time like three days before.
2025-04-01T06:39:16.120976
2018-05-09T14:59:34
321606485
{ "authors": [ "kc9eye" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7475", "repo": "kc9eye/UData", "url": "https://github.com/kc9eye/UData/issues/5" }
gharchive/issue
Create Installer Form Create an installer form to collect information on settings that writes the config file and tests the database, creates objects, and initializes the admin. It should also test whether or not it can send mail, and write to the disk for file uploads. this is not going to happen by me
2025-04-01T06:39:16.130758
2023-07-03T15:38:34
2097506867
{ "authors": [ "Peefy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7476", "repo": "kcl-lang/tree-sitter-kcl", "url": "https://github.com/kcl-lang/tree-sitter-kcl/issues/6" }
gharchive/issue
[Feature] Tree-sitter grammar Discussed in https://github.com/kcl-lang/kcl/discussions/595 Originally posted by matoous July 3, 2023 Hi team! Are there any plans to offer Tree-sitter grammer for KCL? I would be happy to contribute but wonder whether the repository should be maintained under the kcl-lang organization. Hi there @octonawish-akcodes If you are willing, you can work in this repo. Thank you! โค๏ธ
2025-04-01T06:39:16.143584
2022-10-27T14:45:36
1425760800
{ "authors": [ "hardys", "ncdc", "p0lyn0mial" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7477", "repo": "kcp-dev/kcp", "url": "https://github.com/kcp-dev/kcp/issues/2274" }
gharchive/issue
bug: Describe the bug In e2e-sharded runs we're seeing an error like: PROXY W1025 18:20:39.100256 31146 reflector.go:324] k8s.io/client-go@v0.0.0-20221025160842-38c73163e766/tools/cache/reflector.go:167: failed to list *v1alpha1.ClusterWorkspace: Unauthorized PROXY E1025 18:20:39.101457 31146 reflector.go:138] k8s.io/client-go@v0.0.0-20221025160842-38c73163e766/tools/cache/reflector.go:167: Failed to watch *v1alpha1.ClusterWorkspace: failed to list *v1alpha1.ClusterWorkspace: Unauthorized After debugging locally, it appears to be because we're using the root shard-admin token all shards when creating the clients used by the clusterworkspace informer We can see here in some local debug log output that the shard-admin token hitting kcp-1 is actually the token from kcp-0: $ grep "invalid bearer token" kcp-1.log | head -n1 E1027 14:04:04.549596 929593 authentication.go:63] "Unable to authenticate the request SHDEBUG" err="invalid bearer token" req=&{Method:GET URL:/apis/tenancy.kcp.dev/v1alpha1/clusterworkspaces:5678dc3ffb5f4156924efc5c9a4732f12e73fadca3592faa5c762f91fd0fc3f5?limit=500&resourceVersion=0 Proto:HTTP/2.0 ProtoMajor:2 ProtoMinor:0 Header:map[Accept:[application/json, */*] Accept-Encoding:[gzip] Authorization:[Bearer 4ae09225-6a3b-4413-a5a7-d105b45a4466] User-Agent:[kcp-front-proxy/v1.24.3+kcp (linux/amd64) kubernetes/574fe23]] Body:0xc0071db1a0 GetBody:<nil> ContentLength:0 TransferEncoding:[] Close:false Host:<IP_ADDRESS>:6445 Form:map[] PostForm:map[] MultipartForm:<nil> Trailer:map[] RemoteAddr:<IP_ADDRESS>:37122 RequestURI:/clusters/%2A/apis/tenancy.kcp.dev/v1alpha1/clusterworkspaces:5678dc3ffb5f4156924efc5c9a4732f12e73fadca3592faa5c762f91fd0fc3f5?limit=500&resourceVersion=0 TLS:0xc00e689080 Cancel:<nil> Response:<nil> ctx:0xc00fe05d70} $ grep -B2 4ae09225-6a3b-4413-a5a7-d105b45a4466 ../.kcp-0/admin.kubeconfig - name: shard-admin user: token: 4ae09225-6a3b-4413-a5a7-d105b45a4466 Steps To Reproduce Check e2e-sharded logs, or run make test-e2e-sharded locally Expected Behaviour IIUC we need the proxy informer to be able to list/watch clusterworkspace objects on all shards, so we have to use an authentication method which isn't limited to the root shard. Some ideas were discussed on slack and it seems like switching to cert based auth may be the best option - this should mean we can authenticate with any shard, and avoid modifying the proxy config when a shard gets added/removed. Additional Context No response /assign sgtm, thanks for picking it up. So looking at the history here, I see that the current per-shard client code landed in https://github.com/kcp-dev/kcp/pull/1203 - so it would be helpful to get input from @sttts before I start changing it :) If we switch this to use client certs, then I think we need to handle the case where each shard has a different cert, and the method to enable that appears to be the mapping file, however in the sharded-test-server case, we seem to only create mappings to the root shard: $ grep "server:" .kcp-0/admin.kubeconfig server: https://<IP_ADDRESS>:6444 server: https://<IP_ADDRESS>:6444/clusters/root $ grep "server:" .kcp-1/admin.kubeconfig server: https://<IP_ADDRESS>:6445 server: https://<IP_ADDRESS>:6445/clusters/root $ cat .kcp-front-proxy/mapping.yaml - backend: https://localhost:6444 backend_server_ca: .kcp/serving-ca.crt path: /services/ proxy_client_cert: .kcp-front-proxy/requestheader.crt proxy_client_key: .kcp-front-proxy/requestheader.key - backend: https://localhost:6444 backend_server_ca: .kcp/serving-ca.crt path: /clusters/ proxy_client_cert: .kcp-front-proxy/requestheader.crt proxy_client_key: .kcp-front-proxy/requestheader.key So I think for multi-shard deployments we need to specify each shard as a backend (using the IP, not localhost), then we can look up the necessary client cert/key using the shard BaseURL, does that sound reasonable? If we switch this to use client certs, then I think we need to handle the case where each shard has a different cert, I think that should be okay, I think that the proxy could specify a well-known ServerName during bootstrapping a TLS connection to a shard. That would give us a certificate that would much a well-known CA. Does it make sense? If we switch this to use client certs, then I think we need to handle the case where each shard has a different cert, I think that should be okay, I think that the proxy could specify a well-known ServerName during bootstrapping a TLS connection to a shard. That would give us a certificate that would much a well-known CA. Does it make sense? Thanks for the feedback @p0lyn0mial - I get that the ServerName has to match the CN/SAN in the server cert, but I'm unclear if the proxy has to do anything special here - the user provides the server/client certs, and the ServerName will be derived from the ClusterWorkspaceShard spec in the indexer, so the hostname (or potentially IP for CI/dev testing) just has to match? I think leveraging the client certs from the mapping file is definitely viable, and probably optimal if we want to ensure the informer uses per-shard credentials. I can't help wondering if we should just pass an admin kubeconfig to the proxy though which would be potentially much simpler, any thoughts on that? The main disadvantage AFAICS is that we'll need to rename/replace the --root-kubeconfig CLI flag /cc @ncdc who mentioned he has some work in progress related to this Summary of some discussion with @ncdc and @p0lyn0mial @ncdc has some related work in progress which we can possibly revive then use the same client-cert pattern to solve this (convert the informer client to use a proxy->shard client cert instead of the shard-admin token) We can't easily use an admin kubeconfig to switch to client certs, since in the sharded case the client cert terminates at the proxy, then we use request-header at the shard (so, we can't use the same kubeconfig for the proxy->shard auth) Accidentally marked as fixed by 2297 /reopen /assign @p0lyn0mial This should be resolved via https://github.com/kcp-dev/kcp/pull/2382
2025-04-01T06:39:16.163683
2024-04-24T19:17:43
2262005918
{ "authors": [ "souravroy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7478", "repo": "kdhrubo/db2rest", "url": "https://github.com/kdhrubo/db2rest/issues/502" }
gharchive/issue
MongoDB - Order Tests Need to mark tests with @Order() to avoid intermittent test failure. PR https://github.com/kdhrubo/db2rest/pull/503 is merged. Closing the ticket.
2025-04-01T06:39:16.227083
2018-10-10T08:04:32
368540274
{ "authors": [ "Actimel", "tomasfejfar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7479", "repo": "keboola/php-component", "url": "https://github.com/keboola/php-component/issues/47" }
gharchive/issue
Add support for state file Comes up from https://github.com/keboola/db-extractor-mysql/pull/81#discussion_r223825923 Would be nice to have getter for state file. Besides there could be also method for writing into outcomes state. Uลพ tam byla issue https://github.com/keboola/php-component/issues/46 ;)
2025-04-01T06:39:16.229555
2020-11-25T17:07:51
750987183
{ "authors": [ "odinuv", "pivnicek" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7480", "repo": "keboola/storage-api-php-client-branch-wrapper", "url": "https://github.com/keboola/storage-api-php-client-branch-wrapper/pull/5" }
gharchive/pull-request
Update composer.json for latest sapi client Required for https://github.com/keboola/input-mapping/pull/68 Pending sapi-client release https://github.com/keboola/storage-api-php-client/releases/tag/untagged-3d86260daac18b393369 @odinuv what should we call this, 0.0.3? hmm, test failed on "branch not found". is there a possible concurrency issue? they don't run in parallel, it's quite suspicous feel free to make it 1.0.0 :) it's already in production
2025-04-01T06:39:16.246661
2024-08-07T09:41:52
2453042770
{ "authors": [ "JorTurFer", "LY-today", "enaguo", "guanqinglin" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7481", "repo": "kedacore/keda", "url": "https://github.com/kedacore/keda/issues/6045" }
gharchive/issue
create ScaledObject when triggers type is prometheus metricType is Value err Report when triggers type is prometheus๏ผŒ metricType is Value๏ผŒcreate ScaledObject ๏ผŒkeda-operator is err 2024-08-07T09:33:45Z ERROR Reconciler error {"controller": "scaledobject", "controllerGroup": "keda.sh", "controllerKind": "ScaledObject", "scaledObject": {"name":"my-helm-demo","namespace":"crane-system"}, "namespace": "crane-system", "name": "my-helm-demo", "reconcileID": "f750f79f-5a1b-4fc9-88d2-b136a11b65ff", "error": "HorizontalPodAutoscaler.autoscaling \"keda-hpa-my-helm-demo\" is invalid: spec.metrics[0].external.target.type: Invalid value: \"value\": must be either Utilization, Value, or AverageValue"} apiVersion: keda.sh/v1alpha1 kind: ScaledObject metadata: name: my-helm-demo namespace: crane-system spec: scaleTargetRef: apiVersion: apps/v1 kind: Deployment name: my-helm-demo pollingInterval: 15 minReplicaCount: 2 maxReplicaCount: 50 advanced: horizontalPodAutoscalerConfig: behavior: scaleDown: stabilizationWindowSeconds: 0 policies: - type: Percent value: 100 periodSeconds: 15 scaleUp: stabilizationWindowSeconds: 0 policies: - type: Percent value: 500 periodSeconds: 15 triggers: - type: prometheus metricType: Value metadata: serverAddress: xxx metricName: portrait_pod_cpu_predict_compression_all_num query: portrait_pod_cpu_predict_compression_all_num{} threshold: "50" Expected Behavior hpa resources can be created normally Actual Behavior hpa not create Name: my-helm-demo Namespace: crane-system Labels: scaledobject.keda.sh/name=my-helm-demo Annotations: <none> API Version: keda.sh/v1alpha1 Kind: ScaledObject Metadata: Creation Timestamp: 2024-08-07T09:33:24Z Finalizers: finalizer.keda.sh Generation: 1 Resource Version: 24908 UID: 0d11f0d0-216e-4696-be18-51a49cad0bd3 Spec: Advanced: Horizontal Pod Autoscaler Config: Behavior: Scale Down: Policies: Period Seconds: 15 Type: Percent Value: 100 Stabilization Window Seconds: 0 Scale Up: Policies: Period Seconds: 15 Type: Percent Value: 500 Stabilization Window Seconds: 0 Max Replica Count: 50 Min Replica Count: 2 Polling Interval: 15 Scale Target Ref: API Version: apps/v1 Kind: Deployment Name: my-helm-demo Triggers: Metadata: Metric Name: portrait_pod_cpu_predict_compression_all_num Query: portrait_pod_cpu_predict_compression_all_num{} Server Address: xxxx Threshold: 50 Metric Type: Value Type: prometheus Status: Conditions: Message: Failed to ensure HPA is correctly created for ScaledObject Reason: ScaledObjectCheckFailed Status: False Type: Ready Message: ScaledObject check failed Reason: UnkownState Status: Unknown Type: Active Status: Unknown Type: Fallback External Metric Names: s0-prometheus-portrait_pod_cpu_predict_compression_all_num Original Replica Count: 8 Scale Target GVKR: Group: apps Kind: Deployment Resource: deployments Version: v1 Scale Target Kind: apps/v1.Deployment Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning ScaledObjectCheckFailed 59s (x17 over 6m27s) keda-operator Failed to ensure HPA is correctly created for ScaledObject Steps to Reproduce the Problem Logs from KEDA operator 2024-08-07T09:33:45Z ERROR Reconciler error {"controller": "scaledobject", "controllerGroup": "keda.sh", "controllerKind": "ScaledObject", "scaledObject": {"name":"my-helm-demo","namespace":"crane-system"}, "namespace": "crane-system", "name": "my-helm-demo", "reconcileID": "f750f79f-5a1b-4fc9-88d2-b136a11b65ff", "error": "HorizontalPodAutoscaler.autoscaling \"keda-hpa-my-helm-demo\" is invalid: spec.metrics[0].external.target.type: Invalid value: \"value\": must be either Utilization, Value, or AverageValue"} KEDA Version < 2.11.0 Kubernetes Version < 1.28 Platform None Scaler Details prometheus Anything else? $ kubectl version Client Version: v1.29.2 Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3 Server Version: v1.21.1 WARNING: version difference between client (1.29) and server (1.21) exceeds the supported minor version skew of +/-1 kubectl apply --server-side -f https://github.com/kedacore/keda/releases/download/v2.8.0/keda-2.8.0.yaml Hello, KEDA v2.8 is quite old (almost 2 years). I don't remember any issue related but just in case, have you tried using v2.8.2 -> https://github.com/kedacore/keda/releases/tag/v2.8.2? Sadly, that version is out of support, and we won't ship any new releases for it. I know that recent versions don't face with that issue, I'd suggest upgrading to a recent version. Hello, KEDA v2.8 is quite old (almost 2 years). I don't remember any issue related but just in case, have you tried using v2.8.2 -> https://github.com/kedacore/keda/releases/tag/v2.8.2? Sadly, that version is out of support, and we won't ship any new releases for it. I know that recent versions don't face with that issue, I'd suggest upgrading to a recent version. The problem was indeed solved by switching to v2.8.2, but isnโ€™t this the recommended version officially provided by keda? https://keda.sh/docs/2.15/operate/cluster/#kubernetes-compatibility The problem was indeed solved by switching to v2.8.2 Nice! Happy to read it ๐Ÿ˜„ I guess that you are using old k8s cluster, so probably v2.8 is the greatest version that you can use (if you are running k8s < 1.23 KEDA > 2.9 won't work). The point is that we don't ship fixes for old versions, we could ship a fix for the previous version but 2.8 is 7 versions far from v2.15 and although we will try to support with misconfigurations, if there is a real bug unresolved, we won't ship any fix for old versions. The problem was indeed solved by switching to v2.8.2 Nice! Happy to read it ๐Ÿ˜„ I guess that you are using old k8s cluster, so probably v2.8 is the greatest version that you can use (if you are running k8s < 1.23 KEDA > 2.9 won't work). The point is that we don't ship fixes for old versions, we could ship a fix for the previous version but 2.8 is 7 versions far from v2.15 and although we will try to support with misconfigurations, if there is a real bug unresolved, we won't ship any fix for old versions. @JorTurFer OK, I understand what you mean. There is another situation for you, that is, our latest version of keda is found to be incompatible on the cluster of version 1.21, and panic will occur. The reason is that the v2 version of hpa cannot be detected. I also encountered the same problem which caused the function not to work properly๏ผŸ Yes, I said, KEDA >= v2.9 requires k8s >= 1.23. This is because k8s introduced a breaking change when they released HPA v2 and removed HPA v1beta1. v2.8 uses v1beta1 and v2.9 uses v2. There isn't any fix or action to do about this, v2.8 uses v1beta1 and v2.9 uses v2. @JorTurFer OK, thank you for your answer and wish you a happy life @JorTurFer Excuse me, when I only configure cron, will the replicas of the workload become minReplicaCount if it is not in the cron period? @JorTurFer Excuse me, when I only configure cron, will the replicas of the workload become minReplicaCount if it is not in the cron period? @JorTurFer The test is like this. When I manually modify replicas during non-cron periods, it will eventually become minReplicaCount, which is not very friendly to the business. Based on version 1.8.2, is there any way to allow the business to still have the right to make decisions in this situation? Excuse me, when I only configure cron, will the replicas of the workload become minReplicaCount if it is not in the cron period? Yes, outside the cron period and assuming the cron as the only scaler, it'll return 0 so it'll scale to minReplicaCount @JorTurFer่ฏท้—ฎไธ€ไธ‹๏ผŒๅฝ“ๆˆ‘ๅช้…็ฝฎcron็š„ๆ—ถๅ€™๏ผŒๅฆ‚ๆžœไธๅœจcronๅ‘จๆœŸๅ†…๏ผŒๅทฅไฝœ่ดŸ่ฝฝ็š„replicasไผšๅ˜ๆˆminReplicaCountๅ—๏ผŸ @JorTurFerๆต‹่ฏ•ๆ˜ฏ่ฟ™ๆ ท็š„๏ผŒๅฝ“ๆˆ‘ๅœจ้žcronๆœŸ้—ดๆ‰‹ๅŠจไฟฎๆ”นreplicasๆ—ถ๏ผŒๆœ€ๅŽไผšๅ˜ๆˆminReplicaCount๏ผŒๅฏนไธšๅŠกไธๅคชๅ‹ๅฅฝ๏ผŒๅŸบไบŽ1.8.2็‰ˆๆœฌ๏ผŒๆœ‰ๆฒกๆœ‰ๅŠžๆณ•่ฎฉไธšๅŠกๅœจ่ฟ™็งๆƒ…ๅ†ตไธ‹ไพ็„ถๆ‹ฅๆœ‰ๅ†ณ็ญ–ๆƒ๏ผŸ I also encountered the same problem๏ผŒHow to solve it
2025-04-01T06:39:16.254467
2015-07-16T19:43:41
95511708
{ "authors": [ "aendrew", "dustinlarimer", "josephwegner", "standaniels", "timfalls" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7482", "repo": "keen/dashboards", "url": "https://github.com/keen/dashboards/issues/79" }
gharchive/issue
Bower package is empty I don't know why, but bower install keen-dashboards results in a license, a README, and nothing else. Which is a shame, really; you could easily install the necessary CSS and whathaveyou by specifying it as such in the main stanza, which would mean folks like me who use Yeoman could use your layouts far more easily. Might write a PR. Thanks for writing this up @aendrew. Admittedly, we haven't spent much time on more involved ways of distributing Keen Dashboards (bower, yeoman, etc.). That would be super cool if you wrote up a PR! @aendrew just wanted to echo @josephwegner's encouragement toward the PR. We really appreciate your thoughts and any potential contribution! I don't know anything about bower, but I just came across this SO topic and noticed that there is only a bower.json file in the master branch of this project. @standaniels Ah! That would make sense why there's nothing when pulling from bower. If the gh-pages branch was merged with master, that would probably resolve this. Resolved- thanks, everyone!
2025-04-01T06:39:16.258400
2017-03-23T17:47:06
216516094
{ "authors": [ "amounierlltech", "matm" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7483", "repo": "keeneyetech/ecli", "url": "https://github.com/keeneyetech/ecli/issues/42" }
gharchive/issue
Can't use --config to define login file location I use release 0.10 for windows 64. I try to login using --config flag, but always have error 1 in return. I created a .ecli.json somewhere as defined in login standard output, but json format is not exact in this message. missing " after password, and additionnal coma after url. But fixing format didn't allow me to connect. I type : ecli login platform --config "D:\myfolder.ecli.json" What should be --config argument syntax, are windows paths accepted ? is it folder of file path ? Looks like your JSON config file is malformed maybe. It must look like {"platform": {"login": "yourlogin", "password": "yourpassword", "url": "https://yoururl/api/v2"}} Just create it in your local directory and call it by filename: ecli login platform --config myfile.json You can remove the dot . at the beginning on the file name and use any name like myfile.json. It's working with this syntax, thx for your support.
2025-04-01T06:39:16.259592
2020-12-18T14:04:38
770910665
{ "authors": [ "r-czajkowski" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7484", "repo": "keep-network/keep-core", "url": "https://github.com/keep-network/keep-core/pull/2224" }
gharchive/pull-request
Liquidity rewards fetching data Ref: #2216 This PR adds support for fetching and displaying the rewards data from the LPRewards contracts. All data are stored in the redux store. @michalsmiarowski I'm ready for the second round!
2025-04-01T06:39:16.262899
2021-07-26T13:17:51
952895281
{ "authors": [ "michalinacienciala" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7485", "repo": "keep-network/npm-version-bump", "url": "https://github.com/keep-network/npm-version-bump/pull/2" }
gharchive/pull-request
Stabilize failing test The test checking if npm-version-bump action "updates version for already published environment that don't match initial preid" was unstable - was refering to actively worked on npm packages versions, making it necessary to update the action every time new package with that base version is released on ropsten. In this commit we change the reference to different version, which is no longer actively worked on. @nkuba, I wasn't sure what to pick as a base branch (main or v2) for this PR. I see that main is a couple of commits behind v2 and there's no open PR for merging v2 to main. Should I keep the base branch as v2 and after the merge of this PR should I tag the v2 branch code with v2 tag?
2025-04-01T06:39:16.342303
2024-12-11T06:53:18
2731986306
{ "authors": [ "augustin-v", "feltroidprime" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7486", "repo": "keep-starknet-strange/garaga", "url": "https://github.com/keep-starknet-strange/garaga/issues/270" }
gharchive/issue
feat: Centralize RiscZero constants definition by creating Rust bindings for py/ts. Currently, the same constants (CONTROL ROOT & CONTROL_ID) are defined in multiple locations. In python : https://github.com/keep-starknet-strange/garaga/blob/073843663fd253ad33a697a6aae285704df45b21/hydra/garaga/starknet/groth16_contract_generator/parsing_utils.py#L20-L24 In rust : https://github.com/keep-starknet-strange/garaga/blob/073843663fd253ad33a697a6aae285704df45b21/tools/garaga_rs/src/calldata/full_proof_with_hints/groth16.rs#L297-L311 In typescript : https://github.com/keep-starknet-strange/garaga/blob/073843663fd253ad33a697a6aae285704df45b21/tools/npm/garaga_ts/src/node/starknet/groth16ContractGenerator/parsingUtils.ts#L7-L9 Keep them in rust only and create a python / wasm binding so that updating them is easier. Note : the tools/garaga_rs package has already multiple examples of python / wasm bindings. Hello, I would love to work on this. I'm really excited about Garaga with the latest Noir compatibility update and would be honored to be a part of it. I have just completed implementing a L1 gas prices sampling oracle to Katana in Rust so I believe to be able to tackle this issue. I'll make sure to communicate if I encounter any problems. @feltroidprime Thank you, on it!
2025-04-01T06:39:16.344986
2023-06-13T16:02:41
1755228327
{ "authors": [ "MdTeach", "abdelhamidbakhta", "greged93" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7487", "repo": "keep-starknet-strange/madara", "url": "https://github.com/keep-starknet-strange/madara/issues/626" }
gharchive/issue
dev: improve testing of to_invoke_tx Improve the testing of the rpc-core/src/utils.rs file by adding unit tests for to_invoke_tx. The tests are expected to cover edge cases in order to be accepted. @greged93 I would love to take this! @greged93 I would love to take this! Assigned to you @MdTeach need to wait on #597 to be merged Hey @MdTeach, #597 was merged and moved the logic to the file primitives/starknet/src/transation/types.rs and the implementation changed to impl TryFrom<BroadcastedInvokeTransaction> for InvokeTransaction. Would be great if you are still able to provide tests for this. sure @greged93
2025-04-01T06:39:16.347869
2023-06-21T16:23:39
1767970229
{ "authors": [ "lambda-0x" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7488", "repo": "keep-starknet-strange/madara", "url": "https://github.com/keep-starknet-strange/madara/pull/721" }
gharchive/pull-request
dev: use Felt252Wrapper instead of [u8; 32] Pull Request type Please add the labels corresponding to the type of changes your PR introduces: Refactoring (no functional changes, no API changes) What is the current behavior? Resolves: #716 Does this introduce a breaking change? Not sure if this counts as breaking change. Other information (OT: we should add clippy to husky hooks as well)
2025-04-01T06:39:16.351213
2024-09-11T10:48:40
2519390257
{ "authors": [ "0xChqrles", "ikemHood", "manlikeHB", "mubarak23", "ugur-eren" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7489", "repo": "keep-starknet-strange/zkramp", "url": "https://github.com/keep-starknet-strange/zkramp/issues/65" }
gharchive/issue
[CONTRACT] setup liquidity addition Complete this issue according to this article: How to manage liquidity in a P2P ramping app Please add integration tests @0xChqrles i Hi @0xChqrles, an i work on this? Hi @0xChqrles Can i work on this issue if it's available? Hey @0xChqrles I would like to work on this so I can work on #67 which is a dependant. @0xChqrles, can I work on this? @0xChqrles am available to work on this , With previous experience working on this codebase, I have read the article I can implement this with 3 days ETH: 72hrs @0xChqrles am available to work on this
2025-04-01T06:39:16.394902
2019-05-21T01:50:27
446380147
{ "authors": [ "keiohta" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7490", "repo": "keiohta/tf2rl", "url": "https://github.com/keiohta/tf2rl/issues/21" }
gharchive/issue
Reproduce DQN paper results for Atari Reproduce following paper results (at least 1 game, because @keiohta does not have enough computation resource) Human-level control through deep reinforcement learning Preprocessing Working directly with raw Atari 2600 frames, which are 210 x 160 pixel images with a 128-colour palette, can be demanding in terms of computation and memory requirements. We apply a basic preprocessing step aimed at reducing the input dimensionality and dealing with some artefacts of the Atari 2600 emulator. First, to encode a single frame we take the maximum value for each pixel colour value over the frame being encoded and the previous frame. This was necessary to remove flickering that is present in games where some objects appear only in even frames while other objects appear only in odd frames, an artefact caused by the limited number of sprites Atari 2600 can display at once. Second, we then extract the Y channel, also known as luminance, from the RGB frame and rescale it to 84 x 84. The function w from algorithm 1 described below applies this preprocessing to the m most recent frames and stacks them to produce the input to the Q-function, in which m = 4, although the algorithm isrobust to different values of m (for example, 3 or 5).
2025-04-01T06:39:16.425009
2023-07-05T17:48:18
1790015887
{ "authors": [ "keltia" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7491", "repo": "keltia/fetiche-rs", "url": "https://github.com/keltia/fetiche-rs/issues/20" }
gharchive/issue
Statistics/metrics ought to be gathered at Engine level. Right now, sources::access::opensky has some stats gathering code. This should be expanded into an Engine-wide system. tracing might be a good crate to use. More likely as a separate Actor within fetiched.
2025-04-01T06:39:16.433226
2018-05-13T12:23:23
322589437
{ "authors": [ "kemokemo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7492", "repo": "kemokemo/gckdir", "url": "https://github.com/kemokemo/gckdir/issues/18" }
gharchive/issue
failed to create the hash value of large files https://github.com/kemokemo/gckdir/blob/1fb158b19ff1160d2a23244216b076f4c52e3ae1/lib/hash.go#L96-L103 The ioutil.ReadFile() is easy to use, but this read all of the files in memory. If you have any large files (ex. several gigabytes), the function to create hash value will be very slowly. In some cases, gckdir application crashes. ๐Ÿ˜ข Let's fix it! The official godoc of the "crypto/sha256" package is most valuable. https://godoc.org/crypto/sha256#ex-New--File
2025-04-01T06:39:16.440307
2024-10-01T19:24:17
2560016431
{ "authors": [ "CXWorks", "kenba" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7493", "repo": "kenba/opencl3", "url": "https://github.com/kenba/opencl3/issues/69" }
gharchive/issue
Unsound issue in SvmVec Hi, thanks for your time to read this issue. Our static analysis tool found there might be an unsound issue in your set_len implementation of buffer: https://github.com/kenba/opencl3/blob/1d7d74da2ff76c93c7418dd016386a8b907e2bd3/src/svm.rs#L300-L311 As mentioned in the comments, this may introduce uninitilaized memory and reading it is considered as undefined behavior in Rust. As a reference, in std library, all the set_len method are marked as unsafe: https://github.com/rust-lang/rust/blob/63a0bdd5622eaf6b9524702f055bb4525acfc9f2/library/alloc/src/vec/mod.rs#L1849-L1853 Thanks again for your time. Change incorporated in version 0.10.0.
2025-04-01T06:39:16.487044
2016-02-08T23:11:45
132283591
{ "authors": [ "kennethreitz", "victor1969" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7494", "repo": "kennethreitz/requests", "url": "https://github.com/kennethreitz/requests/issues/3000" }
gharchive/issue
Problem with Content-Lenght is missing I'm playing clonk rage but i can't play online because this happend Content-Lenght is missing pls help me PS: i don't speak english so well @victor1969 can you please share the code you are using so we can understand what you're asking? where i get the code? @victor1969 is this an error you saw while playing a game? Why did you decide to post about it here? i'm just trying to play online Clonk rage but says on the server list Internet server on league.clonkspot.org Invalid server response: Content-Length is missing! @victor1969 unfortunately, that has nothing to do with project. There's probably somewhere else online for you to report this, though! :(
2025-04-01T06:39:16.490377
2016-04-21T15:21:26
150106470
{ "authors": [ "Lukasa", "kevinburke", "sigmavirus24" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7495", "repo": "kennethreitz/requests", "url": "https://github.com/kennethreitz/requests/pull/3108" }
gharchive/pull-request
Flip conditional in session.send() Previously we checked that the request being sent was an instance of a PreparedRequest. If a user somehow created a PreparedRequest using a different Requests library instance, this check makes the request un-sendable. (This happened recently - unbeknownst to me, my server was running an outdated version of pip, vulnerable to this issue - pypa/pip#1489, which creates multiple subdirectories (src/requests, src/requests/requests) when you rerun pip install --target. So the PreparedRequest was being created in one version of the library and compared against the other version of the library, and throwing this exception, even though they were both PreparedRequest instances!) It would probably be preferable to check the object's behavior (instead of its type), but a PreparedRequest has a lot of behavior, and it wouldn't be really feasible or allow us to provide a helpful error message to check all of it here. Instead flip the conditional to guard against the user sending an unprepared Request, which should still give us most of the benefits of the better error message. Fixes #3102 I'm happy with this! Go for it @kennethreitz, merge if you'd like to. =D Not if I merge it first ;)
2025-04-01T06:39:16.502134
2017-10-24T00:23:57
267858266
{ "authors": [ "coveralls", "obi1kenobi" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7496", "repo": "kensho-technologies/graphql-compiler", "url": "https://github.com/kensho-technologies/graphql-compiler/pull/49" }
gharchive/pull-request
Update filter handler functions to use FilterOperationInfo objects. Since we now have a lot of different filter operators, it has become clear that the old filter function signature can use some streamlining. This PR adds a new FilterOperationInfo object that holds all non-contextual data that is relevant to a filtering operation, such as the directive object itself, or the type and field name being filtered on. This new abstraction: simplifies the function signature of filtering functions to just 3 arguments, while getting rid of all unused variables; and simplifies the reasoning for which filters are valid in which situations. Coverage decreased (-0.5%) to 92.553% when pulling f50fc3a51842cc330bf735048237f23bfecf0b8a on filter_operation_info into da6e03c424abc17cc242341674823ea8c21c060d on edge_degree_filtering.
2025-04-01T06:39:16.511739
2015-06-19T13:25:02
89564228
{ "authors": [ "jdart", "kentcdodds" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7497", "repo": "kentcdodds/kcd-angular", "url": "https://github.com/kentcdodds/kcd-angular/pull/3" }
gharchive/pull-request
bower was removing all the good stuff constants, directives, filters and services were not available after bower installing. I've merged it, but to be perfectly honest, I don't recommend you use this module. I will not be supporting it. I recommend you copy and paste and modify the code.
2025-04-01T06:39:16.516208
2016-09-07T10:55:58
175473095
{ "authors": [ "craigcosmo", "leggomuhgreggo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7498", "repo": "kenwheeler/slick", "url": "https://github.com/kenwheeler/slick/issues/2508" }
gharchive/issue
how to center item and some this is how I want it to be displayed this is how I got so far What I want to do is show one center image, and half of the right and left images, Problem: In my solution, I was able to show one center image, but the right and left image only show little, I want them to show more Here is my code: http://jsfiddle.net/fp1kcah6/ Check out centerPadding
2025-04-01T06:39:16.561042
2020-11-25T11:15:19
750728349
{ "authors": [ "ankitjain28may", "bacherfl", "codecov-io", "johannes-b" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7499", "repo": "keptn/keptn", "url": "https://github.com/keptn/keptn/pull/2733" }
gharchive/pull-request
Keptn Smart Auth: Autofetch keptn endpoint and api-token Signed-off-by: ankitjain28may<EMAIL_ADDRESS> [ ] Can you please add the message "CLI is not authenticated" in case of ClusterIP / NotePort. (to inform the user about the current status) Codecov Report Merging #2733 (18ba47f) into master (9bd9c79) will decrease coverage by 13.64%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #2733 +/- ## =========================================== - Coverage 36.26% 22.61% -13.65% =========================================== Files 161 50 -111 Lines 9288 1110 -8178 Branches 197 197 =========================================== - Hits 3368 251 -3117 + Misses 5381 832 -4549 + Partials 539 27 -512 Flag Coverage ฮ” moduleA 22.61% <รธ> (-0.27%) :arrow_down: Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage ฮ” bridge/client/app/_services/data.service.ts 0.98% <0.00%> (-0.12%) :arrow_down: ...lient/app/project-board/project-board.component.ts 2.83% <0.00%> (-0.03%) :arrow_down: bridge/client/app/_models/service.ts 0.00% <0.00%> (รธ) helm-service/controller/action_handler.go platform-support/openshift-route-service/main.go api/ws/hub.go ...ce/pkg/handler/approval_triggered_event_handler.go helm-service/pkg/helm/helm_helper.go cli/cmd/update.go .../pkg/configurationchanger/configuration_changer.go ... and 100 more Continue to review full report at Codecov. Legend - Click here to learn more ฮ” = absolute <relative> (impact), รธ = not affected, ? = missing data Powered by Codecov. Last update ba0b2ae...18ba47f. Read the comment docs. @bacherfl The unit tests are failing because it needs this PR to be merged - https://github.com/keptn/kubernetes-utils/pull/13 @bacherfl The unit tests are failing because it needs this PR to be merged - keptn/kubernetes-utils#13 Ok, got it. Just had one comment about the other PR - once that has been adressed we can merge that one @bacherfl The unit tests are failing because it needs this PR to be merged - keptn/kubernetes-utils#13 Ok, got it. Just had one comment about the other PR - once that has been adressed we can merge that one @ankitjain28may I have just merged the kubernetes-utils PR. Can you please update the kubernetes-utils dependency to the latest status from the master branch? When that is done we can merge this PR as well @bacherfl Have updated the dependency to latest master, Please review :)
2025-04-01T06:39:16.569686
2021-11-20T14:03:57
1059152027
{ "authors": [ "fchollet", "fmerizzi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7500", "repo": "keras-team/keras-io", "url": "https://github.com/keras-team/keras-io/pull/716" }
gharchive/pull-request
fixing issue in GCN example The small issue prevented the code from running in Colab environment, because of inconsistencies in the dimension of the output layer. Thanks for the PR. This is fixed by #717 717
2025-04-01T06:39:16.589182
2023-10-06T19:27:58
1930825996
{ "authors": [ "fchollet", "qlzh727", "sampathweb" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7501", "repo": "keras-team/keras", "url": "https://github.com/keras-team/keras/issues/18567" }
gharchive/issue
TensorFlow GPU - Fix keras/layers/merging/merging_test.py Fix failing test - keras/layers/merging/merging_test.py::MergingLayersTest::test_sparse_dot_2d Fatal Python error: Aborted and update TODO in https://github.com/keras-team/keras/blob/master/keras/kokoro/github/ubuntu/gpu/build.sh#L39 https://source.cloud.google.com/results/invocations/9df9ee7e-5666-4644-abd2-01a10771faeb/targets/keras%2Fgithub%2Fubuntu%2Fgpu%2Ftensorflow%2Fpresubmit/log keras/layers/merging/merging_test.py::MergingLayersTest::test_sparse_dot_2d Fatal Python error: Aborted Current thread 0x00007f51610f0740 (most recent call first): File "/tmpfs/venv/lib/python3.9/site-packages/tensorflow/python/ops/linalg/sparse/gen_sparse_csr_matrix_ops.py", line 1114 in sparse_matrix_sparse_mat_mul File "/tmpfs/src/github/keras/keras/backend/tensorflow/numpy.py", line 119 in sparse_sparse_matmul File "/tmpfs/src/github/keras/keras/backend/tensorflow/numpy.py", line 156 in matmul File "/tmpfs/src/github/keras/keras/ops/numpy.py", line 3431 in matmul File "/tmpfs/src/github/keras/keras/layers/merging/dot.py", line 171 in batch_dot File "/tmpfs/src/github/keras/keras/layers/merging/dot.py", line 320 in _merge_function File "/tmpfs/src/github/keras/keras/layers/merging/base_merge.py", line 189 in call File "/tmpfs/src/github/keras/keras/ops/operation.py", line 47 in __call__ File "/tmpfs/src/github/keras/keras/utils/traceback_utils.py", line 114 in error_handler File "/tmpfs/src/github/keras/keras/layers/layer.py", line 810 in __call__ File "/tmpfs/src/github/keras/keras/utils/traceback_utils.py", line 114 in error_handler File "/tmpfs/src/github/keras/keras/testing/test_case.py", line 380 in run_layer_test File "/tmpfs/src/github/keras/keras/layers/merging/merging_test.py", line 240 in test_sparse File "/tmpfs/venv/lib/python3.9/site-packages/absl/testing/parameterized.py", line 319 in bound_param_test File "/usr/lib/python3.9/unittest/case.py", line 550 in _callTestMethod File "/usr/lib/python3.9/unittest/case.py", line 592 in run File "/usr/lib/python3.9/unittest/case.py", line 651 in __call__ File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/unittest.py", line 333 in runtest File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__ File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 262 in <lambda> File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 341 in from_call File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 261 in call_runtest_hook File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 222 in call_and_report File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 133 in runtestprotocol File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 114 in pytest_runtest_protocol File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__ File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 350 in pytest_runtestloop File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__ File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 325 in _main File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 271 in wrap_session File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 318 in pytest_cmdline_main File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__ File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/config/__init__.py", line 169 in main File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/config/__init__.py", line 192 in console_main File "/tmpfs/venv/bin/pytest", line 8 in <module> github/keras/keras/kokoro/github/ubuntu/gpu/build.sh: line 34: 4954 Aborted (core dumped) The culprit is between 2.15.0.dev20230918 (good) and 2.15.0.dev20230919 (bad). Tested via: pytest keras/layers/merging/merging_test.py::MergingLayersTest::test_basic_add Culprit in one of the changes in this range: git log e4a6720f42a..dfcf1d40e46 --oneline Thanks Ramesh for the repo, we will revisit this during the triage meeting. On a side note, I didn't find any change on the sparse side between those two date. Will need to dig deep for the root cause. Here's a small code snippet to reproduce the issue in Colab with Keras Master and TF-Nightly - !pip uninstall -y keras tensorflow !pip install tf-nightly[and-cuda]==2.15.0.dev20231009 --extra-index-url https://pypi.nvidia.com !pip uninstall -y keras-nightly # Install Keras from Master via `python pip_build.py --install` input = keras.layers.Input(shape=(2,)) x1 = keras.layers.Dense(4, activation='relu')(input) x2 = keras.layers.Dense(4, activation='relu')(input) added = keras.layers.Add()([x1, x2]) out = keras.layers.Dense(1)(added) model = keras.models.Model(inputs=input, outputs=out) x = np.random.randn(8, 2) y = np.random.randn(8, 1) model.compile(optimizer='sgd', loss='mse') model.fit(x, y, epochs=1) If I replace Add with Concatenate also it breaks. This is a high priority error since it breaks very important layer in TF GPU. Also, the same test fails for JAX GPU as well. @fchollet - If you have any thoughts or suggestions to try let me know. The example you provided doesn't even use sparse inputs, which is different from the error on top. The error DNN library initialization failed somehow indicate that its a setup issue for GPU. The example you provided doesn't even use sparse inputs, which is different from the error on top. The error DNN library initialization failed somehow indicate that its a setup issue for GPU. There are multiple failures in merging_test.py. I tried to run the basic test case with add and that fails. Initially I reported on the sparse test which actually aborts with core dump keras/layers/merging/merging_test.py::MergingLayersTest::test_sparse_dot_2d Fatal Python error: Aborted TF Nightly 09/18 works for ALL the tests in merging_test.py. So I think its a common issue due to change in TF on 09/19 between these commits in TF: git log e4a6720f42a..dfcf1d40e46 --oneline Somehow I wasn't able to produce the on colab with T4 GPU. https://colab.sandbox.google.com/drive/1_hMJieL_6DobTPUbZ6BRZIEVz0YRHhBo#scrollTo=GM2B7qEqNYqk Maybe I didn't config the GPU properly? @sampathweb do u have a testable env that I can run with? Also seems to be failing with JAX-GPU now: github/keras/keras/kokoro/github/ubuntu/gpu/build.sh: line 57: 4493 Aborted (core dumped) pytest keras --ignore keras/applications --ignore keras/layers/merging/merging_test.py --ignore keras/trainers/data_adapters/py_dataset_adapter_test.py --ignore keras/backend/jax/distribution_lib_test.py --cov=keras I wil work on this tomorrow. I used Colab v100 as my test env Seems to be a Cudnn TF compilation issue. 2023-10-17 20:23:09.628643: I external/local_xla/xla/service/service.cc:176] StreamExecutor device (0): Tesla V100-SXM2-16GB, Compute Capability 7.0 2023-10-17 20:23:10.277194: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:447] Loaded runtime CuDNN library: 8.7.0 but source was compiled with: 8.9.4. CuDNN library needs to have matching major version and equal or higher minor version. If using a binary install, upgrade your CuDNN library. If building from sources, make sure the library loaded at runtime is compatible with the version specified during compile configuration. 2023-10-17 20:23:10.278786: W tensorflow/core/framework/op_kernel.cc:1839] OP_REQUIRES failed at xla_ops.cc:574 : FAILED_PRECONDITION: DNN library initialization failed. Look at the errors above for more details. Tested via pytest keras/layers/merging/merging_test.py::MergingLayersTest::test_basic_add Don't have a resolution yet. But might be related to this change that's within the range - git log e4a6720f42a..dfcf1d40e46 --oneline commit 3de44168950a5972ba4cfa7e3c6cbf4cffa67fe6 Author: A. Unique TensorFlower<EMAIL_ADDRESS>Date: Mon Sep 18 13:50:11 2023 -0700 Upgrade to LLVM 17, CUDA 12.2, and CuDNN 8.9.4 This is updating TF's default toolchain to LLVM 17, as well as CUDA and cuDNN to the latest releases. PiperOrigin-RevId: 566403707
2025-04-01T06:39:16.607783
2017-10-18T08:58:28
266412679
{ "authors": [ "lfoppiano", "stzellerhoff" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7502", "repo": "kermitt2/grobid", "url": "https://github.com/kermitt2/grobid/issues/249" }
gharchive/issue
Build error duplicate entry Installing version 0.4.4 fails reporting this error: Failed to execute goal org.apache.maven.plugins:maven-shade-plugin:3.0.0:shade (default) on project grobid-core: Error creating shaded jar: duplicate entry: META-INF/services/shadedgrobid.org.apache.lucene.codecs.PostingsFormat How can I avoid duplicate entries? Stephan hi @stzellerhoff could you paste here the whole log? Thanks Luca
2025-04-01T06:39:16.652767
2024-06-26T03:17:52
2374096242
{ "authors": [ "aarontrowbridge", "andgoldschmidt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7503", "repo": "kestrelquantum/QuantumCollocation.jl", "url": "https://github.com/kestrelquantum/QuantumCollocation.jl/issues/132" }
gharchive/issue
[Bug]: namespace includes common variable names What happened? In _integrators.jl, export state export controls export timestep export comps export dim Version stable release What does this bug affect? [ ] quantum system construction [ ] problem setup [ ] problem solution [ ] problem performance [ ] solution analysis [ ] plotting [ ] documentation [ ] tests [ ] other (please specify below) Other information No response these should all be prefixed with get_ probably closed by #149
2025-04-01T06:39:16.681900
2024-02-07T01:46:04
2121997626
{ "authors": [ "kevincobain2000", "vunhatchuong" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7504", "repo": "kevincobain2000/gobrew", "url": "https://github.com/kevincobain2000/gobrew/issues/174" }
gharchive/issue
Bug: Windows 10 Symlink required privilege Problem Windows 10 symlink required privilege: Looks like it hasn't been resolved: https://github.com/golang/go/issues/22874 Solution You can either require privilege or use cmd through os.exec.Command: if runtime.GOOS == "windows" { if err = exec.Command("cmd", "mklink", "/d", link , destination).Run(); err == nil { return nil } } Thanks for reporting. Pull req too please? Ok I'll try
2025-04-01T06:39:16.688436
2022-03-20T08:43:10
1174476354
{ "authors": [ "b0o", "kevinhwang91" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7505", "repo": "kevinhwang91/nvim-bqf", "url": "https://github.com/kevinhwang91/nvim-bqf/issues/60" }
gharchive/issue
Support previewing vim-fugitive :Gclog entries Feature description Currently, when the quickfix window is populated by vim-fugitive's :Gclog command, vim-bqf's preview window is blank unless the highlighted entry has been manually opened first. I'm guessing it has to do with the special file URIs fugitive uses, which look like fugitive:///home/maddy/.config/nvim/.git//f0e7c59ec59a25bfe928b555dd8387242d2b810f. Describe the solution you'd like It would be nice if nvim-bqf could properly preview these vim-fugitive quickfix items. Additional context If the highlighted entry hasn't been manually opened yet: Once the highlighted entry has been opened: nvim-bqf will never support this feature because of the performance issue. You can use should_preview_cb hack what you want. You can also use the below setting to experience how slow it is. require('bqf').setup({ should_preview_cb = function(bufnr) local bufname = vim.api.nvim_buf_get_name(bufnr) if bufname:match('^fugitive://') then vim.api.nvim_buf_call(bufnr, function() cmd(('noa do fugitive BufReadCmd %s'):format(bufname)) end) end return true end }) For me, I use https://github.com/rbong/vim-flog which is a git log wrapper and has the ability to limit the commits. More importantly, it is an extension based on fugitive. Cool, thank you for the tips, and for the wonderful plugin! I took your hack and hacked it even further - this loads the fugitive buffer asynchronously: require('bqf').setup { preview = { should_preview_cb = function(bufnr) local bufname = vim.api.nvim_buf_get_name(bufnr) if bufname:match '^fugitive://' then local pvs = require 'bqf.preview.session' local extmark = require 'bqf.preview.extmark' local ts = require 'bqf.preview.treesitter' vim.schedule(function() vim.api.nvim_buf_call(bufnr, function() local fbufnr = pvs.float_bufnr() vim.cmd(('do fugitive BufReadCmd %s'):format(bufname)) pvs.floatbuf_reset() ts.disable_active(fbufnr) extmark.clear_highlight(fbufnr) vim.cmd(('silent lua require"bqf.utils".transfer_buf(%d, %d)'):format(bufnr, fbufnr)) end) end) end return true end, }, } I copied some code from bqf.preview.handler.open() to do this, I'm sure there's a better way but this works well enough for me. Please update the code which is convenient to hack. local bqf_pv_timer require('bqf').setup { preview = { should_preview_cb = function(bufnr, qwinid) local bufname = vim.api.nvim_buf_get_name(bufnr) if bufname:match '^fugitive://' and not vim.api.nvim_buf_is_loaded(bufnr) then if bqf_pv_timer and bqf_pv_timer:get_due_in() > 0 then bqf_pv_timer:stop() bqf_pv_timer = nil end bqf_pv_timer = vim.defer_fn(function() vim.api.nvim_buf_call(bufnr, function() vim.cmd(('do fugitive BufReadCmd %s'):format(bufname)) end) require('bqf.preview.handler').open(qwinid, nil, true) end, 60) end return true end } } Awesome, much cleaner. Thank you!
2025-04-01T06:39:16.696267
2023-07-13T14:15:19
1803128402
{ "authors": [ "kevinscaria", "noshaq" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7506", "repo": "kevinscaria/InstructABSA", "url": "https://github.com/kevinscaria/InstructABSA/issues/16" }
gharchive/issue
Missing create_data_in_joint_task_format method at DatasetLoader Hi. Thanks for sharing the code of your papers. I'm trying to reproduce the training for the joint task, but this error appears: AttributeError: 'DatasetLoader' object has no attribute 'create_data_in_joint_task_format' There is any way to solve it with the other methods in the class? Thanks! Hi, There has been some API changes. I have not updated the run_model.py script yet. But for joint task, it should be: create_data_in_aspe_format() From here on joint task will be called as ASPE (Aspect Sentiment Pair Extraction), since new tasks are being added. Great! Thanks for your answer.
2025-04-01T06:39:16.780386
2017-06-07T12:21:49
234190595
{ "authors": [ "bharathvaj1995", "kevva" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7531", "repo": "kevva/brightness", "url": "https://github.com/kevva/brightness/issues/13" }
gharchive/issue
(node:8987) UnhandledPromiseRejectionWarning: Unhandled promise rejectio n (rejection id: 1): Error: No backlight device found This is the error that occurs when executing on linux platform. Please update the module to the latest node.js with catch() Error: No backlight device found Using the CLI or the regular API? Anyway, you can see why it fails in the error message. Please update the module to the latest node.js with catch() You should catch errors yourself. Try running ls /sys/class/backlight. Does it list anything? Nothing. But i installed backlight using apt-get install xbacklight What distro are you running? And I assume you can change brightness otherwise?
2025-04-01T06:39:16.783143
2016-11-02T04:58:14
186712651
{ "authors": [ "shinnn", "tunnckoCore" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7532", "repo": "kevva/decompress", "url": "https://github.com/kevva/decompress/issues/44" }
gharchive/issue
Decouple builtin plugins? Currently i just uses decompress-unzip, but it needs some more workarounds it because of the dest. I don't need the others. Yea they are pretty small deps, but... would be better. :) FYI you can directly use a plugin without decompress. Example: https://github.com/shinnn/get-zonetab/blob/ddf68e876dcc93cc891d3c289c45f97ea215d3ad/index.js#L15 Yea, that's what I'm doing now, but i should handle files manually. utils.unzip()(buf).then((files) => { // handle / decompress files }) maxogden/extract-zip seems to fit your use case.
2025-04-01T06:39:16.784607
2017-03-07T19:52:17
212540055
{ "authors": [ "maxtaco", "thibaultdej" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7533", "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/6160" }
gharchive/issue
Can't log into the Keybase app Keybase GUI Version: 1.0.18-20170223121400+af0e640 I tried to log in through the shell after resetting my keys and it returned 'ERROR SecretSyncer: no device found for ID=660e***'. I can log in in the browser, but not on the desktop. Thanks for the report, this should be fixed in our next release.
2025-04-01T06:39:16.821384
2017-09-21T16:46:35
259562025
{ "authors": [ "AngryAnt", "Avamander", "Hexstream", "IpsmLorem", "ItsExtra", "Mikaela", "QuentinBellus", "TheKangaroo", "bradmurray", "cbluth", "cjb", "dabura667", "dlumma", "domq", "fortran77", "getraf", "grilldos", "heronhaye", "hexagile", "jcausey-astate", "jfinstrom", "junderw", "mahmad-nutrien", "maxtaco", "mvdan", "neunenak", "rsoltanzadeh", "srd424", "supakeen", "tstrijdhorst", "w3bb", "wolfmcnally" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7534", "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/8586" }
gharchive/issue
Please Support Multiple Identities or Personas I have two personas online, my professional persona and my personal/family personas are kept separate online. In this real-life example, I will never associate my facebook with my github via way of keybase (i wont proof facebook and github on the same keybase account, because that then associates them). Are my concerns common, is anyone else concerned about the same thing? Can/Will keybase support multiple personas or identities? We do support this. You can keybase logout of one account and then log in to the second. We don't plan to make it much easier than it already is, because most people don't do this. I also would love the ability to be logged in to multiple accounts simultaneously. I understand Keybase is โ€œsingle identityโ€ driven, but I think that idea is misguided. The equivalent in GnuPG would be to say โ€œERROR: โ€”genkeys: you can only generate one private key ring per instance of gnupg, please log out of current OS user and log into a new user to manage a second private keypairโ€ You are just creating an artificial hurdle that wonโ€™t prevent people from doing it, and will only hurt the people who try jumping over that hurdle and trip and fall and break their face. People who want to do this will do it, so why not make it simpler for them? If the answer is โ€œitโ€™s not a priorityโ€ I totally understand and can put up with the inconvenience of switching between identities. But if the answer is โ€œmultiple email addresses are bad. multiple personas are bad, everyone should know about all your emails and PGP keys and social accountsโ€ then I think that is misguided. The whole โ€œdiscourage multiple accountsโ€ thing makes sense for Facebook and twitter etc. but not for keybase imo. The logout/login process is not very user friendly because if people want to run multiple identities they usually want to be able to be logged into all of them at the same time which would mean the clients should be able to handle it. Running multiple instances on win/osx/linux would work, but multiple instances on a phone are problematic. @bradmurray I agree. Logging in/out is not representative of real-life, for example, any one of my acquaintances should be able to reach me at any time, despite whether they are professional acquaintances or otherwise. Just because i'm "at work" doesnt mean i shouldnt receive notifications from anyone else. Excellent software but honestly I'm surprised this wasn't a must-have in version 1. Part of managing one's privacy these days is managing one's identities. I know many people who would balk at using Keybase because they can't easily manage multiple identities. We already support this. On Wed, Feb 14, 2018 at 1:32 AM Wolf McNally<EMAIL_ADDRESS>wrote: Excellent software but honestly I'm surprised this wasn't a must-have in version 1. Part of managing one's privacy these days is managing one's identities. I know many people who would balk at using Keybase because they can't easily manage multiple identities. โ€” You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/keybase/client/issues/8586#issuecomment-365509735, or mute the thread https://github.com/notifications/unsubscribe-auth/AA05_5nmKbxpy6Qo4gSBEbrGYXtP5js0ks5tUn35gaJpZM4Pfm9d . @maxtaco I just checked the iPhone app and logging in to multiple accounts simultaneously (the subject of this issue) is not yet supported. I need to log out and log back in with a different account. During which time I can no longer get notifications for the previous account. We already support this. How is this already supported? I run keybase on Android, Windows and OSX and on all of them I have to logout and log back in to change identities. I'm a new keybase user using it for multiple identities and I too would like to see the UX around being logged into multiple identities simultaneously become better. This is a sorely missing feature, imo. I'll add my vote for this one as well, and also point out that the Firefox container solution doesn't help with chat -- chat seems to be only supported with the native app, with the exception of the limited chat enabled by the browser extension. So, the browser can't really fix this issue. I have a similar use case to others here; I have a need to keep my work-related communication separate from personal accounts. I'm happy to see the new "exploding messages" feature; I hope this one might get some consideration now that that one is done. (Maybe also chat search via the GUI, but that's a different issue.) Hey, I just started to use keybase and the first thing I checked was if I can create separate accounts for private and business use cases. It would be great to have these two logged in in the app simultaneously. Jumping on this bandwagon as well. A lot of services nowadays do support multiple identities and/or personas logged in at the same time. It'd be awesome if keybase could do the same for those of us who keep separate online identities for separate areas of work. N+1'd. An important use case is the ability to make "bot" accounts to be members in a team (e.g. to auto-push DevOps secrets into a Keybase Git repo) +1 on this. Multiple identities should be supported. We don't plan to make it much easier than it already is, because most people don't do this. Many people who already have accounts in different places do have one for work and one for their personal identity. I know many, MANY people in this case, and looking at the list of people in this thread, there are indeed many of us. Please bring this to Keybase +1 I also have personal (where I am generally opinionated on things that matter to me) and my company account (fun but filtered) Conflicted if the "already supported" responses ought to be treated as depressing or hilarious. Dropping in to +1 here. Keybase looks fascinating but I have had my fingers burnt in the past by letting different areas of my life overlap (new rule: colleagues only get added on social media when one or other of us leaves!) Not doing that again! Want this as well. The logout/login process is not very user friendly because if people want to run multiple identities they usually want to be able to be logged into all of them at the same time which would mean the clients should be able to handle it. Running multiple instances on win/osx/linux would work, but multiple instances on a phone are problematic. If it is not feasible on phone, then don't try to implement this for phone. Do it for the desktop app. It is (obviously) a largely overlooked feature. +1000 For me, this is a dealbreaker - as much as I love keybase, until multiple personas will be properly supported, it will be just a toy app I'm checking out every now and then. +1 This is the main reason I'm not using keybase tbh. It's quite understandable that software developers want to discourage or prevent account creation spam. It can clog up the namespace, for one thing. It allows Sybil attacks in dicsussions, for another, so a person can pretend to be multiple people all arguing the same side (happens on Reddit and Twitter all the time). Making it too easy for a person to have many logins will also make it too easy for the person to take up many times the free allocation of cloud storage per person. So I definitely will not argue for multiple identities without limit. However, two is a bare minimum, because many of us find it desirable to keep our personal and professional lives separate. If carrying a single mobile device, you would want to have both identities logged in. Or you might have separate business and personal devices, and in that case, you can much more easily implement security procedures, so certain business-related files are never allowed to touch a personal device. You might need a third identity to say and do naughty things, not naughty enough to make you a criminal in the civilized world, but naughty enough to damage your personal or professional life if found out and, in some less civilized parts of the world, be stoned to death. You can probably imagine reasons for additional identities. I think about five or six would be a reasonable limit. Definitely a must have ! I just joined a workplace that uses Keybase and would love to have the ability to simultaneously sign into my work and personal accounts. I'd understand if it wasn't a priority though. We're close on this, either in this release or the next. In https://github.com/keybase/client/issues/2104#issuecomment-493079134 it's said that you can currently keybase login -s anotheraccount to switch account Is there any UI support planned for this? yes. however I currently can't get desktop notifications for any of the other logged in accounts. I have to switch and visually check for unread chats. Please add support for this! I honestly can't believe someone, never mind someone responsible for a product like keybase, would honestly think "most people don't have multiple identities on the internet" There's a subculture of people who seem to have melted their selves into the internet like butter on toast. People who do this will naturally interact with lots of others who have done the same to a greater or lesser extent so they come to believe it is normal. If a primary way you interact with other humans is online, then you might eventually arrive at the conclusion that the people posting prolifically online are representative of all humans. Because how else would you ever comes across anyone? It's a weird self perpetuating bubble. That said I am not sure if the devs/bosses here think this (though I haven't read all the other related issues yet so maybe they do) or if they are merely indifferent or uninterested. IMHO security and privacy/anonymity are intrinsically linked in every practical way. Like the others, I would also consider this tool much more useful if I didn't have to sit and think which portion of my life it is most suited for, which is my train of thought at the moment. Feature: is in current release version. Feature Request Issue: continues discussion as if it is not there. Feature: Feature: is in current release version. I am new to this whole github thing but why is the issue still marked as "open"? Also #2104 "Feature Request: Logging into multiple keybase accounts locally". In my limited experience, mostly lurking, devs/admins are extremely happy to close these things once they are resolved. Found this thread (and the many others which link here) trying to determine if it is possible or no. All I see is many people stating it's not possible. I can't see anything obvious on my phone about how to do this. "Add identities" lets me verify on about 800 mastodon servers. Because these threads get long, and after a while it's no longer clear what feature they were talking about to begin with. Also, Keybase client has 2.6k issues open... which tells you that this project is not focused much on closing / managing issues and is more focused on coding / reviewing code. The closed issues mostly tend to be bugs that were fixed, since they are much easier to put a binary value on completion. I think this issue is still open and people are still talking about this because currently, there is no Graphical User Interface (GUI) support for multiple identities or personas through the Keybase app. Here's my +1 for GUI support in addition to the Command Line Interface (CLI) implementation. @junderw Feature: is in current release version. The feature definitely doesn't exist in current version in an acceptable manner. +1. Especially with chat there's times where I'd want to use alternative identities. If Keybase is aiming for businesses to support Keybase, then it would be incredibly useful to have different accounts for work and private use. Note that these days twitter allows to simultaneously connect to up to 5 accounts/identities at the same time, right there through the normal official GUI. Just click on "More" then the blue "โŠ•" to add more accounts. (I never use that feature, but I'm glad it exists.) Also note that these days, gmail (or only G Suite?) supports multiple accounts natively: Kebase technically supports it, but having to re-enter your password every time is a massive pain. On Wed., Nov. 6, 2019, 7:19 a.m. Jean-Philippe Paradis, < <EMAIL_ADDRESS>wrote: Also note that these days, gmail (or only G Suite?) supports multiple accounts natively: [image: Multiple gmail accounts] https://camo.githubusercontent.com/5646d7d9503d7d5513b0623695dfd33968a03675/68747470733a2f2f7777772e68657873747265616d2e6e65742f746d702f676d61696c2d6d756c7469706c652d6163636f756e74732e706e67 โ€” You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/keybase/client/issues/8586?email_source=notifications&email_token=AIEGR7W5YENP5SGDKQTYRS3QSKY5DA5CNFSM4D36N5O2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEDGK3GY#issuecomment-550284699, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIEGR7U5DEVHW4FWYV6FJFLQSKY5DANCNFSM4D36N5OQ . Being able to be logged in to multiple accounts simultaneously is distinctly different. I think this issue is still open and people are still talking about this because currently, there is no Graphical User Interface (GUI) support for multiple identities or personas through the Keybase app. Here's my +1 for GUI support in addition to the Command Line Interface (CLI) implementation. We just got our wish. Multiple identities or personas support through the Keybase app Graphical User Interface (GUI). That's in addition to the Command Line Interface (CLI) implementation. Yes! Sounds great, but as soon as I select "Log in as another user" it immediately logs me out and I have to enter my password again to log back in... This is now in the app in version 5.0 on all platforms. You should be able to switch users without entering your password every time. @Hexstream, if you have "Always stay logged in" checked in the Advanced Settings tab, that sounds like a bug. Please send feedback after reproducing and make another issue. Thanks. You are right, checking "Always stay logged in" solves the problem, but this behavior is not intuitive at all. I created a new issue. Thanks so much for fixing this issue! I use this feature on a daily basis, and it works great. I have two suggestions to make it even more awesome: https://github.com/keybase/client/issues/23850 to show notifications for all accounts configured as "always stay logged in". https://github.com/keybase/client/issues/23851 to allow switching between "always stay logged in" accounts with just a keyboard shortcut, or a single click, and remember what tab was last open for each account.
2025-04-01T06:39:16.823953
2018-03-30T17:49:36
310127146
{ "authors": [ "chrisnojima" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7535", "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/11157" }
gharchive/pull-request
fix jump to new conversation of kbfs chats @keybase/react-hackers this fixes the jump to new convo flow I renamed the payload param and flow didn't catch it because the payload types weren't exact. i don't remember why we undid that (it used to be) but it really should go back i think
2025-04-01T06:39:16.826098
2019-12-09T22:17:54
535371730
{ "authors": [ "buoyad", "maxtaco" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7536", "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/21496" }
gharchive/pull-request
fix loading team on create Messy because team get takes a name Spawn a getDetails on team created because the component will not have the teamname Store entry teamNameToID on team created so we have it as soon as possible cc @keybase/y2ksquad We can easily expose an RPC that takes an ID (if we donโ€™t have one already) @maxtaco changing the gui to use the team ID here is part of y2k-974
2025-04-01T06:39:16.831689
2016-10-12T07:47:11
182457338
{ "authors": [ "oconnor663", "sijad" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7537", "repo": "keybase/kbfs", "url": "https://github.com/keybase/kbfs/pull/419" }
gharchive/pull-request
Update kbfstool URL it has been changed in https://github.com/keybase/kbfs/commit/3cf9045546540bb72435bb48cd53471773a46c76 but URL didn't updated. can be related to #309 Thanks!
2025-04-01T06:39:16.836158
2016-10-28T03:44:06
185832218
{ "authors": [ "strib" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7538", "repo": "keybase/kbfs", "url": "https://github.com/keybase/kbfs/pull/473" }
gharchive/pull-request
cr: speed up while journaling by avoiding extraneous block uploads and downloads This PR enables a journaled TLF to use only local information during conflict resolution, and completely avoids creating a conflict branch on the servers. Getting there required quite a few changes (let me know if you'd prefer them broken out into separate PRs): tlfJournal pauses flushing as soon as it notices a conflict. conflictResolver figures out which blocks don't need to be uploaded, and passes that to MDOps.ResolveBranch so the journal can ignore them. We now write the resolution MD to the journal, rather than directly to the server (due to the ResolveBranch semantics). Since the journal might have uploaded some blocks before discovering a conflict, BlockServer now has an IsFlushed method that the conflictResolver can use to tell which blocks have already been flushed. blockJournal now has a way to avoid removing the data for flushed blocks from disk, until the next successful MD flush. This helps us avoid re-downloading blocks in the case where the resolution MD itself hits a conflict during flushing, and we have to re-do CR all over again. (This is possible now because we are writing the resolution MD to the journal, and won't know if there's a conflict until later.) conflictResolver needed several tweaks to work correctly when the unmerged branch contains a resolutionOp, which wasn't possible before. RootMetadata.deepCopy needs to also copy the cached block changes -- otherwise when the journal converts a branch and replaces the cached MD entries, it could wind up forgetting the cached block change pointer. This fixes a bug in master that we just didn't have test coverage for, but now we do. Ok fixed up, ready for another look please. Known unrelated CI failures. Merging.
2025-04-01T06:39:16.838675
2017-04-16T05:12:02
221991159
{ "authors": [ "akalin-keybase", "coveralls" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7539", "repo": "keybase/saltpack", "url": "https://github.com/keybase/saltpack/pull/28" }
gharchive/pull-request
Switch on major versions when reading Otherwise, we may panic when reading a header from an unknown minor version. Move some non-common functions out to better places. Coverage increased (+0.04%) to 86.244% when pulling 72453ddb0f9d49ab06adb214e119187bb313b94d on akalin/fix-switch into 5ff37ca6c69573e0786615153f260a45b266476b on master. Hmm. I should probably write a test for this. I'll prep one tomorrow.
2025-04-01T06:39:16.900387
2023-02-23T11:52:19
1596719277
{ "authors": [ "THS-on", "codecov-commenter", "maugustosilva" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7540", "repo": "keylime/keylime", "url": "https://github.com/keylime/keylime/pull/1323" }
gharchive/pull-request
roadmap: update for 2023 This updates the roadmap with the major changes planned for Keylime in 2023. Please have a look and see, if the goals are defined well enough and if I forgot something to add. Codecov Report Merging #1323 (cc93e8a) into master (8ddfd12) will decrease coverage by 0.49%. The diff coverage is n/a. :mega: This organization is not using Codecovโ€™s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more Additional details and impacted files Flag Coverage ฮ” packit-e2e 74.15% <รธ> (-0.49%) :arrow_down: testsuite 11.07% <รธ> (รธ) unittests 1.92% <รธ> (รธ) Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage ฮ” keylime/keylime_agent.py 72.55% <0.00%> (-0.58%) :arrow_down: Just to make clear the key realization on the parenthetical comment on the 4th bullet on @THS-on answer. It seems clear to me there is the need for separating the "ascertaining the integrity" of the attestation artifacts (i.e., tpm2_checkquote, the replaying of both MB and IMA log and matching of PCRs) and the subsequent validation against policy.
2025-04-01T06:39:16.949687
2016-05-24T15:17:22
156535078
{ "authors": [ "morenoh149", "pll33" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7541", "repo": "keystonejs/generator-keystone", "url": "https://github.com/keystonejs/generator-keystone/issues/200" }
gharchive/issue
Installation failed with "yo keystone" Expected behavior I was following the "Get Started" page, "Commands". I wanted to install the latest version of keystone and startup new project. Actual behavior After the keystone-generator installed, tried to do "yo keystone" and the installation has started but it does not generate the project. I got the following error: screen shot 2016-05-24 at 10 21 49 Thanks for helping me out! Issue was fixed with #195, the Gruntfile is out of date on the 0.3.16 with npm install @mxstbr
2025-04-01T06:39:16.956905
2021-06-25T09:20:40
929998322
{ "authors": [ "Jytesh", "Kikobeats" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7542", "repo": "keyvhq/keyv", "url": "https://github.com/keyvhq/keyv/pull/14" }
gharchive/pull-request
(WIP) build: restructure project Changes [x] Setup Dependabot [x] Use standard instead of xo [x] Unify Licenses [x] Setup npm metadata [x] Setup editorconfig Should homepage be changed to whatever site we deploy docs to?
2025-04-01T06:39:16.963016
2024-07-17T10:47:51
2413277227
{ "authors": [ "chrdebru", "dachafra", "frmichel" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7544", "repo": "kg-construct/rml-cc", "url": "https://github.com/kg-construct/rml-cc/pull/49" }
gharchive/pull-request
Included RML Strategy, fixed some mistakes, and updated sources to respect RML-IO @anaigmo: I have removed "join" from the comment in the ontology to make rml:Strategy and rml:strategy implicitly specific to RML-CC. @frmichel Can you check? Hi @chrdebru, sorry I've been away from the last discussions. I'm just confused about commit https://github.com/kg-construct/rml-cc/pull/49/commits/29216b27393ed1b1ab2a83e8fd94cd09e1b394a1, I though the plan was to move rml:Strategy to core, but it is reintegrated into CC here. Why is that? @frmichel the explanation is here https://github.com/kg-construct/rml-core/issues/51 (last comment from Ben) Ok got it, thx @dachafra, sorry for missing that. no worries! :-) Just one tiny remark, the rendered.html mentions 10 January 2024 as the spec date. This is on purpose or should be updated? Apart from that I've good, you can merge the PR whenever you like.
2025-04-01T06:39:16.972887
2023-01-18T12:50:53
2101722153
{ "authors": [ "bjdmeest", "dachafra", "elsdvlee", "pmaria" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7545", "repo": "kg-construct/rml-core", "url": "https://github.com/kg-construct/rml-core/issues/74" }
gharchive/issue
Join specification when logical source is the same Let's say we have two triple maps that refer to the same logical source (and with same, we really mean same URI, not "same because the descriptions lead to the semantically same logical source"). Sample source (CSV) id,parent_id 1,2 2,1 Base mapping (YARRRML) prefixes ex: http://example.com# sources: test: [data.csv] mappings: test1: s: ex:$(id) po: p: ex:parent o: mapping: test2 test2: s: ex:$(parent_id) We have following use cases that are underspecified in de spec the spec currently says If the logical source of the child triples map and the logical source of the parent triples map of a referencing object map are not identical, then the referencing object map must have at least one join condition. If a join condition is specified AND the logical source is not the same: common case, execute join condition between each iteration pair If a join condition is specified AND the logical source is the same: same as above If no join condition is specified AND the logical source is not the same: do a full join (i.e., take all iterations into account) example output: ex:1 ex:parent ex:2, ex:1 ex:parent ex:1, ex:2 ex:parent ex:2, ex:2 ex:parent ex:1 If no join condition is specified AND the logical source is the same: don't do a full join, but take the current iteration into account example output: ex:1 ex:parent ex:2, ex:2 ex:parent ex:1 this last one is the edge case, but allows to 'join per iteration'. Question is: should we make this edge case explicit, or should there be a different way to tackle this edge case? To continue the discussion of this issue, and considering that there is already a spec written, I would suggest making a PR @elsdvlee so the rest can review it and provide comments! See https://github.com/kg-construct/rml-core/pull/78 awesome! Please assign @andimou @pmaria @bjdmeest @DylanVanAssche as potential reviewers My view on defining equality of logical sources: Object equality in programming languages is used as the basis for many things. For example comparison in different data structures for uniqueness and hashing. (Think dictionaries, sets etc.) I strongly believe we should be able to leverage this for logical sources. I think source and logical source equality is something that is very useful to have when building RML processors. Therefor, I would propose to come up with a definition of equality which can be implemented as such. My proposal would be to define a logical source or source to be equal to another logical source or source if the RML-defined properties of the description of both are equal. RML-defined: Those properties that are defined by a specification to have behavior that influences the behavior of an RML processor. These properties MUST be listed for the rml:LogicalSource specification. These properties MUST be listed for any rml:Source description. Doing so will allow RML processors to map these descriptions to standard object equality mechanisms in their respective programming languages to best leverage the language's abilities.
2025-04-01T06:39:17.016576
2017-05-19T12:56:03
229966649
{ "authors": [ "khorolets" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7548", "repo": "kiasaki/docker-alpine-postgres", "url": "https://github.com/kiasaki/docker-alpine-postgres/issues/33" }
gharchive/issue
Fatal error with latest update Hi! Thanks for your image, I'm happy using it! Unfortunately, with the latest update, I'm not able to run postgres in docker-compose.yml The error is FATAL: could not create lock file "/run/postgresql/.s.PGSQL.5432.lock": No such file or directory Your image part from docker-compose.yml postgres: restart: always image: kiasaki/alpine-postgres:latest environment: - POSTGRES_PASSWORD=test - POSTGRES_DB=test_db - POSTGRES_USER=test_user volumes: - ./persistant/pgdata:/var/lib/postgresql/data I've switched to 9.5 tag and everything works fine. Sorry looks like the same as #32 You can close the issue if you decide that it's a duplicate. Thank you!
2025-04-01T06:39:17.018193
2016-11-15T11:15:14
189364456
{ "authors": [ "Scirrio", "Yukun99", "kicjow" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7549", "repo": "kicjow/Crazy-Enchantments", "url": "https://github.com/kicjow/Crazy-Enchantments/issues/110" }
gharchive/issue
Telepathy does not work with Fortune enchantment The telepathy enchantment disables fortune from allowing ore blocks to yield more than 1 of their resource, whereas the fortune enchantment without telepathy would allow ores to yield 2-3. Is there any possible fix to the telepathy enchantment to allow more than 1 resource to be sent to the inventory when fortune is being used on the same tool? Ill be fixing it in the next update. #Badbones69totherescue This has been fixed and will be in the next update.
2025-04-01T06:39:17.027636
2019-02-08T20:24:19
408310364
{ "authors": [ "dannyalright", "eoji" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7550", "repo": "kickstarter/android-oss", "url": "https://github.com/kickstarter/android-oss/pull/471" }
gharchive/pull-request
Logged in drawer updates What โ“ making avatar 48 by 48 making name bigger Story ๐Ÿ“– Trello before and after ๐Ÿ‘€ killing it! One tweak: could you please increase the distance from the top of the screen to the header avatar? Between top-of-header-avatar and system bar: 16dp (or 18 if you want to use our grid) Between top-of-header-avatar and top-of-screen: 40dp (or 42 if you want to use our grid) That space is 16dp, my avi has a white background so it's hard to see Improve navigation and filtering on Android app
2025-04-01T06:39:17.035892
2021-06-16T17:20:33
922880477
{ "authors": [ "msadoon", "nativeksr" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7551", "repo": "kickstarter/ios-oss", "url": "https://github.com/kickstarter/ios-oss/pull/1521" }
gharchive/pull-request
[Ignore] CI Failure Validation Just drafting PR to verify if a CI failure is related to changes in Cartfile or general environment issue. SwiftFormat found issues: File Rules Kickstarter-iOS/Library/OptimizelyResultTypeTests.swift:13:1 warning: (spaceInsideComments) Add leading and/or trailing space inside comments. Generated by :no_entry_sign: Danger SwiftFormat found issues: File Rules Kickstarter-iOS/AppDelegate.swift:10:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:11:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:12:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:13:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:14:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:15:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:16:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:17:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:18:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:19:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:20:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:21:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:22:1 warning: (sortedImports) Sort import statements alphabetically. Generated by :no_entry_sign: Danger No longer needed to check ci validation
2025-04-01T06:39:17.038781
2018-05-30T19:09:57
327861802
{ "authors": [ "Scollaco", "ifbarrera" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7552", "repo": "kickstarter/ios-oss", "url": "https://github.com/kickstarter/ios-oss/pull/368" }
gharchive/pull-request
Password reveal functionality What Show/hide password on login. ๐Ÿ‘ See ๐Ÿ‘€ Native: Password reveal option to make entering passwords easier
2025-04-01T06:39:17.130269
2022-01-21T06:20:06
1110099930
{ "authors": [ "cristianonicolai", "kie-ci", "nmirasch" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7553", "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/1916" }
gharchive/pull-request
KOGITO-6606 KOGITO-6446 Produce event listener before starting Data Iโ€ฆ โ€ฆndex The (build) kogito-runtimes check has failed. Please check the logs. The (build) kogito-apps check has failed. Please check the logs. The (build) kogito-examples check has failed. Please check the logs. The (build) kogito-apps check has failed. Please check the logs. Jenkins rerun kogito-examples tests Jenkins rerun kogito-apps tests The (build) kogito-examples check has failed. Please check the logs. Jenkins rerun kogito-examples tests
2025-04-01T06:39:17.284993
2023-04-03T15:23:17
1652287705
{ "authors": [ "baldimir", "danielezonca", "gitgabrio" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7554", "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/2919" }
gharchive/pull-request
[kie-issues-156] Use specific conversion method for numbers Fixes https://github.com/kiegroup/kie-issues/issues/156. When the convert method is executed on IBM JDK, the order of iteration of methods is different. For Integer, the original implementation picked a wrong method (Integer.getInteger()), which caused a wrong behaviour and test fail. This PR uses valueOf() method specifically for numbers. Adds more tests for various number types. How to replicate CI configuration locally? Build Chain tool does "simple" maven build(s), the builds are just Maven commands, but because the repositories relates and depends on each other and any change in API or class method could affect several of those repositories there is a need to use build-chain tool to handle cross repository builds and be sure that we always use latest version of the code for each repository. build-chain tool is a build tool which can be used on command line locally or in Github Actions workflow(s), in case you need to change multiple repositories and send multiple dependent pull requests related with a change you can easily reproduce the same build by executing it on Github hosted environment or locally in your development environment. See local execution details to get more information about it. How to retest this PR or trigger a specific build: for pull request checks Please add comment: Jenkins retest this for a specific pull request check Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] tests for quarkus branch checks Run checks against Quarkus current used branch Please add comment: Jenkins run quarkus-branch for a quarkus branch specific check Run checks against Quarkus current used branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-branch for quarkus main checks Run checks against Quarkus main branch Please add comment: Jenkins run quarkus-main for a specific quarkus main check Run checks against Quarkus main branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-main for quarkus lts checks Run checks against Quarkus lts branch Please add comment: Jenkins run quarkus-lts for a specific quarkus lts check Run checks against Quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-lts for native checks Run native checks Please add comment: Jenkins run native for a specific native check Run native checks Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native for native lts checks Run native checks against quarkus lts branch Please add comment: Jenkins run native-lts for a specific native lts check Run native checks against quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native-lts Full Kogito testing (with cloud images and operator BDD testing) Please add comment: Jenkins run BDD This check should be used only if a big change is done as it takes time to run, need resources and one full BDD tests check can be done at a time ... How to backport a pull request to a different branch? In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch). NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another. Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added. If something goes wrong, the author will be notified and at this point a manual backporting is needed. NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created. @kiegroup/gatekeepers Could you please merge this, squashing commits? Thanks! I see @baldimir asked the review to @fjtirado too. @baldimir Do you want to wait or is it fine for you to proceed? hi @danielezonca Let's wait for @fjtirado @danielezonca ^^
2025-04-01T06:39:17.438942
2024-02-16T06:29:35
2137959334
{ "authors": [ "kieranabrennan", "zanovis" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7555", "repo": "kieranabrennan/every-breath-you-take", "url": "https://github.com/kieranabrennan/every-breath-you-take/issues/21" }
gharchive/issue
ECG Data Nice job on this app! I wanted to ask if you have plans to incorporate the ECG characteristics in the future, as I see the UUIDs are called out in the .py file. Thanks for any info! @zanovis you'll see now I'm connecting to the sensors with a python package I've written that also has support for ECG https://github.com/kieranabrennan/blehrm/blob/master/blehrm/clients/polar_h10.py
2025-04-01T06:39:17.448821
2024-11-03T06:20:25
2631004400
{ "authors": [ "Orenji-Tangerine" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7556", "repo": "kijai/ComfyUI-Florence2", "url": "https://github.com/kijai/ComfyUI-Florence2/issues/92" }
gharchive/issue
Slower execution when compared with Pytorch 2.1.2 + CU 11.8 + xformers 0.0.23.post1 I just updated my environment to 2.5.0 + CU 12.4 + xformers 0.0.28.post2, and it seems that it is slower and if I have a few images to loads, it will stop for a while before processing next image (at Florence2Run node). In the past, processing 10 images took about 40 secs now it needs 100 secs. Are there anything I can do to speed up? Issue gone after updating to Pytorch 2.5.1 + CU 12.4 (xformers 0.0.28.post3)
2025-04-01T06:39:17.451037
2016-04-25T17:47:29
150935779
{ "authors": [ "cgurnik", "mmailhot" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7557", "repo": "kikinteractive/kik-python", "url": "https://github.com/kikinteractive/kik-python/pull/16" }
gharchive/pull-request
Fix verify_signature in Python 3 We were not properly accounting for the fact that the hashing functions in Python 3 take bytestrings. Note that in Python 3, the body of the HTML request must be a bytestring when passed into the verify_signature function, this is consistent with how Flask and other HTTP libraries work. Fixes #14 @mmailhot :+1:
2025-04-01T06:39:17.466525
2023-03-09T13:15:57
1617231083
{ "authors": [ "kiliman", "zhuhaow" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7558", "repo": "kiliman/remix-flat-routes", "url": "https://github.com/kiliman/remix-flat-routes/issues/43" }
gharchive/issue
Deep flat-file convention causing wrong route added If we add something like this - a+ - b+ - index.tsx This would generate route for a+/b+ and a/b at the same time. Odd. The + are removed and treated like you used . a+/b+/index.tsx should map to a.b.index.tsx Can you show me the remix.config? The content of remix.config const { flatRoutes } = require("remix-flat-routes"); /** @type {import('@remix-run/dev').AppConfig} */ module.exports = { devServerBroadcastDelay: 1000, ignoredRouteFiles: ["**/.*"], server: "./server.js", serverBuildPath: "functions/[[path]].js", serverConditions: ["worker"], serverDependenciesToBundle: "all", serverMainFields: ["browser", "module", "main"], serverMinify: true, serverModuleFormat: "esm", serverPlatform: "neutral", future: { unstable_tailwind: true, unstable_postcss: true, v2_routeConvention: true, }, routes: (defineRoutes) => { return flatRoutes("routes", defineRoutes); }, }; My actual routes looks like this <Route path="admin+/users+" file="routes/admin+/users+/index.tsx" /> <Route path="admin" file="routes/admin+/_layout.tsx"> <Route path="users/edit/:id?" file="routes/admin+/users+/edit.($id).tsx" /> <Route path="users/" index file="routes/admin+/users+/index.tsx" /> <Route path="users/new" file="routes/admin+/users+/new.tsx" /> </Route> Ah, you can't use v2_routeConvention and remix-flat-routes at the same time. v2 doesn't support the + convention. I see. I thought the routes function would take over the route generation. Thanks for the explanation. Change that fix the issue
2025-04-01T06:39:17.471053
2023-09-02T10:52:24
1878608041
{ "authors": [ "Dmitryqq", "killkimno" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7559", "repo": "killkimno/MORT", "url": "https://github.com/killkimno/MORT/issues/15" }
gharchive/issue
Index was out of range error Hello Getting this error every time I open the program, clicking the "Apply" button or finishing quick setting. As the result, settings are not applying. Pic 2 is coming after the first one (it's saying application is shutting down) but nothing happens. By the way russian translate is bad and breaking the interface, how can I change it to english? hello It seems like the app settings are messed up. Do the following to reset and try again. Additional -> Reset config change language Additional -> Advanced settings -> App language Resetting config helped, thanks. Its weird, because it was my 1st launch and already with broken config. And that "Config file" submenu in russian looks like file setup/configuration, and the buttons are "Setting", "Save settings" and "reload". It seems that a problem occurred during the initialization process. Thank you for reporting the worng text! I'll fix it
2025-04-01T06:39:17.473768
2024-09-03T11:52:41
2502675703
{ "authors": [ "kiloutyg" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7560", "repo": "kiloutyg/efnc", "url": "https://github.com/kiloutyg/efnc/issues/100" }
gharchive/issue
Updating stuff for deployment outside initial plant Redo : install script update script correct bugs at startup Update globals variable and their use in email and other stuff.
2025-04-01T06:39:17.529691
2022-01-21T13:47:54
1110490075
{ "authors": [ "kingsaint", "phlobo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7561", "repo": "kingsaint/BioMedical-EL", "url": "https://github.com/kingsaint/BioMedical-EL/issues/7" }
gharchive/issue
Documentation of setup for BioBERT and neleval First of all, thanks for sharing this repo! However, I have a bit of trouble reproducing the results, maybe you can help. Running the training script python run_e2e_span.py --data_dir data/BC5CDR/processed_data --model_type bert --output_dir o utput --do_train --n_gpu 1 --model_name_or_path ./biobert_v1.1_pubmed yields different errors. If I use the BioBERT checkpoint from here https://github.com/dmis-lab/biobert: I get the error: OSError: file ./biobert_v1.1_pubmed/config.json not found If, however, I use this checkpoint https://huggingface.co/dmis-lab/biobert-v1.1/tree/main I'm getting OSError: Error no file named ['pytorch_model-1000000.bin', 'tf_model.h5', 'model.ckpt.index'] found in directory ./biobert-v1.1/ or from_tf set to False neleval The Readme states that evaluation is run with ./neleval/BC5CDR-AllSpan/gold.csv. How do I get this file? Is it some produced as part of running the e2e_span script? Thanks in advance! Hi, Thanks for your interest in our paper. If you are using https://github.com/dmis-lab/biobert, please rename the bert_config.json file in ./biobert_v1.1_pubmed to config.json. Alternatively, you can specify the path to the config file using the --config_name argument in the training script. If you are using https://huggingface.co/dmis-lab/biobert-v1.1/tree/main, please rename pytorch_model.bin to pytorch_model-1000000.bin. gold.csv will be created when you do inference on the trained model. Use the --do_eval argument to do so. Then copy the gold.csv file to a directory named BC5CDR-AllSpan. I hope this is helpful. Thank you so much! Regarding 1. I tried the second option (using the HuggingFace model) and actually progressed a bit further. However, now I am getting an error 01/25/2022 12:12:17 - INFO - __main__ - Loading features from cached file data/BC5CDR/processed_data/cached_train_biobert-v1.1 Traceback (most recent call last): File "run_e2e_span.py", line 1071, in <module> main() File "run_e2e_span.py", line 1021, in main global_step, tr_loss = train(args, model, tokenizer) File "run_e2e_span.py", line 76, in train train_dataset, _, _ = load_and_cache_examples(args, tokenizer) File "run_e2e_span.py", line 725, in load_and_cache_examples all_mention_start_indices = torch.tensor([f.mention_start_indices for f in features], dtype=torch.long) ValueError: expected sequence of length 16 at dim 1 (got 9) Do you have any hints what might be the problem? use --overwrite_output_dir, --overwrite_cache, and --use_hard_and_random_negative during training. During inference use --use_all_candidates
2025-04-01T06:39:17.531539
2024-03-20T15:56:49
2197903816
{ "authors": [ "kingyue737", "xj63" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7562", "repo": "kingyue737/vitify-next", "url": "https://github.com/kingyue737/vitify-next/pull/1" }
gharchive/pull-request
fix(vite.config): mdi icon convert camelCase to kebab-case When the previous solution encounters mdiBattery80, it will be converted to mdi-battery-8-0. When using mdi-battery-80, an error that the icon cannot be found will be prompted. This solution first converts uppercase letters and then matches one or more numbers. There are some weird names here, such as mdiBattery10Bluetooth. : ) Thank you very much!
2025-04-01T06:39:17.541160
2024-02-19T17:20:53
2142851072
{ "authors": [ "dr-frmr", "hosted-fornet", "nick1udwig" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7563", "repo": "kinode-dao/kinode", "url": "https://github.com/kinode-dao/kinode/issues/259" }
gharchive/issue
terminal: pasting causes cursor to go to end of line E.g. imagine I have a line foo baz and I have my cursor on the space and bar in the clipboard. Pasting will properly modify the line, but will move my cursor to the end (i.e. to the z in baz), rather than, as expected, the end of the paste (the space in bar ). FYI @tadad @dr-frmr (not sure who owns this part of terminal). fixed in develop!
2025-04-01T06:39:17.544550
2020-08-03T18:10:48
672253572
{ "authors": [ "mauriciovasquezbernal" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7564", "repo": "kinvolk/inspektor-gadget", "url": "https://github.com/kinvolk/inspektor-gadget/pull/126" }
gharchive/pull-request
Some improvements for network policy advisor Align "--namespace" and "--namespaces" flags Add --output option for report comman @alban I think these changes are still relevant.
2025-04-01T06:39:17.546756
2017-09-05T10:53:57
255237321
{ "authors": [ "dongsupark", "jonboulle" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7565", "repo": "kinvolk/kube-spawn", "url": "https://github.com/kinvolk/kube-spawn/issues/130" }
gharchive/issue
Support overlayfs on btrfs after moby's new release Currently we don't support btrfs as underlying device's filesystem, mainly because moby doesn't allow it. Linux kernel 4.7 or newer already supports overlayfs on btrfs, but moby started to support since https://github.com/moby/moby/pull/33281 (2017-05-19). The commit has not been included in any release so far. Even the most recent release 17.05.0-ce doesn't have it. So we need to wait for a while until moby released a new version with it, as well as Container Linux includes the new version. Then we can allow overlayfs on btrfs. /cc @iaguis fixed by #193 Right. Thanks!
2025-04-01T06:39:17.562037
2017-07-21T10:03:59
244623298
{ "authors": [ "kirilkirkov", "vikaskumarmca" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:7566", "repo": "kirilkirkov/Shopping-Cart-Solution-CodeIgniter", "url": "https://github.com/kirilkirkov/Shopping-Cart-Solution-CodeIgniter/issues/15" }
gharchive/issue
index.php hi thank for sharing good source but i am running this code if am facing index.php problem each and every open page from admin. http://localhost/shopping/admin/publish Problem http://localhost/shopping/index.php/admin/publish right @vikaskumarmca This is not a problem. You can load admin/publish with index.php before and without :).