added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:39:15.431469
| 2020-05-13T14:23:44
|
617492882
|
{
"authors": [
"brianxjx",
"mchenggit"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7439",
"repo": "kabanero-io/kabanero-pipelines",
"url": "https://github.com/kabanero-io/kabanero-pipelines/issues/374"
}
|
gharchive/issue
|
Replacing-pipelineresources-with-tasks
PipelineResources didn't make the cut for beta and have been replaced by a combination of Tekton Catalog Tasks and Workspaces.
v1alpha1 => v1beta1
The new odo based pipelines are already using task and workspace. There is no plan to change existing Appsody based pipeline until we know when PipelineResource is removed from Tekton.
|
2025-04-01T06:39:15.469361
| 2024-10-11T11:58:53
|
2581211326
|
{
"authors": [
"azatsafin",
"odorT"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7440",
"repo": "kafbat/helm-charts",
"url": "https://github.com/kafbat/helm-charts/issues/29"
}
|
gharchive/issue
|
Error: UPGRADE FAILED: parse error at (kafka-ui/templates/deployment.yaml:57): unclosed action
Issue submitter TODO list
[X] I've looked up my issue in FAQ
[X] I've searched for an already existing issues here (legacy) and here
[X] I've tried installing latest charts and the issue still persists there
[X] I'm running a supported version of the application & chart which is listed here
Describe the bug (actual behavior)
it seems new version of kafbat helm chart can't be upgraded anymore
Expected behavior
installation without errors
Your installation details
https://github.com/kafbat/kafka-ui/commit/2956664
kafka-ui-1.4.2
yamlApplicationConfig:
kafka:
clusters:
- name: test
bootstrapServers: test.test:9092
schemaRegistry: http://test1.test1:8081
4. -
Steps to reproduce
just applying this values fails
Screenshots
No response
Logs
No response
Additional context
No response
https://github.com/kafbat/helm-charts/blob/ab4948ba71b99576b7096a120597989a6c869a53/charts/kafka-ui/templates/deployment.yaml#L57
this change broke master
any updates on this?
any updates on this?
Could you kindly share the complete values file and the specific error message you encountered?
command: helm upgrade --install kafka-ui kafbat-ui/kafka-ui -f values.yaml -n kafka-ui
output:
Release "kafka-ui-deleteme" does not exist. Installing it now.
Error: parse error at (kafka-ui/templates/deployment.yaml:57): unclosed action
helm version: version.BuildInfo{Version:"v3.5.4", GitCommit:"1b5edb69df3d3a08df77c9902dc17af864ff05d1", GitTreeState:"clean", GoVersion:"go1.15.11"}
values.yaml:
replicaCount: 1
route:
enabled: true
env:
- name: filtering.groovy.enabled
value: "true"
- name: DYNAMIC_CONFIG_ENABLED
value: "true"
resources:
requests:
cpu: '1'
memory: '1Gi'
yamlApplicationConfig:
management:
health:
ldap:
enabled: false
volumes:
- name: config-data
persistentVolumeClaim:
claimName: config-pv-claim
volumeMounts:
- mountPath: "/etc/kafkaui"
name: config-data
helm version: version.BuildInfo{Version:"v3.5.4", GitCommit:"1b5edb69df3d3a08df77c9902dc17af864ff05d1", GitTreeState:"clean", GoVersion:"go1.15.11"}
Please upgrade helm to the latest version and try again.
yes, it worked ! thanks a lot
|
2025-04-01T06:39:15.475262
| 2024-05-24T20:15:05
|
2316190045
|
{
"authors": [
"YpNo",
"chicknlil",
"kaffetorsk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7441",
"repo": "kaffetorsk/arlo-streamer",
"url": "https://github.com/kaffetorsk/arlo-streamer/issues/22"
}
|
gharchive/issue
|
Pyaarlo
Please update this to the most current version of pyaarlo. The one you are using doesn't have the most recent fixes, nor does it support the newest devices.
@YpNo it seems like our fearless leader @kaffetorsk has jumped ship and that you are the only one answering questions, trying to update the code, etc. Are you able to fork this project to be able to keep on top of things?
Much obliged for your consideration.
Ahah, I have been thinking about it but we should find a real python developper and/or a video stream "expert" to help. I can do some things but I could reach some limits.
I'll keep you inform
Hi,
Okay okay, I think I will fork the project :)
@chicknill @bbo76 @RaidMax @xitation I'll keep you inform when the fork is ready. I don't know if I would treat all your issues or suggestions but I'll do my best.
If you know a real Python developpers with asynchronous operation skill and video handling to help me, it would be great !
The fork is ready if you want to try : https://github.com/YpNo/arlo-camera-streamer
It is the same version with little changes. It should run as today.
But it has been built with the latest version of pyaarlo ;)
Feel free to (re)open issue/discussion about your needs.
Welcome back @kaffetorsk !
I will close my fork if youโre really return back. Feel free to find contributors to help you maintain this great app. I can be one of them if you let me build new releases (with new version of pyaarlo first).
Let us know about the future of this project.
Regards.
Thank you, I will take better care of this repo going forward and appreciate the effort you put in.
As to the future of the project I think it is best to keep the scope to it's original intent (which is enabling arlo cameras for 3rd party NVRs (such as frigate)) and keep it maintained.
I also realize the need for maintenance is higher than I expected, mostly due to pyaarlo beeing based on a reverse-engineered API, suspect to change at any given moment :-)
I'll look into adding contributors
Dropped release v0.7.5 btw, pyaarlo is latest bleeding edge (my fork with a yet to be merged pr)
|
2025-04-01T06:39:15.494823
| 2024-02-06T17:32:12
|
2121341608
|
{
"authors": [
"Itxaka",
"jimmykarily",
"mudler"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7442",
"repo": "kairos-io/kairos",
"url": "https://github.com/kairos-io/kairos/issues/2217"
}
|
gharchive/issue
|
UKI: kcrypt unlock-all doesn't unlock TPM-bound partitions
Currently unlocking partitions encrypted with TPM manually by calling kcrypt unlock-all doesn't work.
Workaround exists, and documented in https://kairos.io/docs/installation/trustedboot/#mount-partitions-after-install
To reproduce:
In the Kairos config, try to run kcrypt unlock-all in an after-install stage (e.g. to write some files to the disk)
Install Kairos in UKI mode
See installation failing.
Possible solution:
Introduce a new stage/hook (e.g. "after-decrypt") to allow people to run code right after decrypting the disks. This stage will also make sure the disk is encrypted again when the stage is done.
Also, there is a --tpm flag in kcrypt unlock-all command that might workaround the issue.
kcrypt unlock-all could work if we add a --tpm flag so it knows it needs to go over the tpm unlock workflow instead of the usual one.
Introduced in version 0.9.0: https://github.com/kairos-io/kcrypt/compare/v0.7.0...v0.9.0 (cut in December 18th: https://github.com/kairos-io/kcrypt/releases/tag/v0.9.0)
|
2025-04-01T06:39:15.495992
| 2024-04-11T09:00:21
|
2237245668
|
{
"authors": [
"mauromorales"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7443",
"repo": "kairos-io/kairos",
"url": "https://github.com/kairos-io/kairos/pull/2463"
}
|
gharchive/pull-request
|
Symlink any /boot/Image* to /boot/vmlinuz
The jetson produces /boot/Image which wasn't caught by the previous mechanism
fixes #2461
tested manually, this tests dont' run on pr so merging
|
2025-04-01T06:39:15.506055
| 2021-07-25T11:05:46
|
952250620
|
{
"authors": [
"kaiwalyakoparkar",
"vind3v17"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7444",
"repo": "kaiwalyakoparkar/classroom-monitor-bot",
"url": "https://github.com/kaiwalyakoparkar/classroom-monitor-bot/pull/23"
}
|
gharchive/pull-request
|
fix: ๐Updated Discord Link
Fixes Issue (#22 )
This PR fixes the following issues : Updated Discord Link In Greeting Action.
hey @vind3v17, I see a different link added there. Kindly add this link (I mentioned this in the issue as well) https://discord.gg/K9kxUXvfND
|
2025-04-01T06:39:15.538150
| 2021-12-17T08:56:22
|
1083027564
|
{
"authors": [
"cnouguier"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7445",
"repo": "kalisio/kapture",
"url": "https://github.com/kalisio/kapture/issues/7"
}
|
gharchive/issue
|
Waiting for animation
When capturing a GeoJSON file we need to wait fot the animation. For now we have defined a wait duration according the activity (map or globe). It could be great to tell Kano to swith off any animation when zooming.
For now, we prodive a waitDelay before taking the snapshot. It could be defined using a WAIT_DELAY environment variable or could be provided on the query
|
2025-04-01T06:39:15.723062
| 2022-12-20T03:09:05
|
1503917791
|
{
"authors": [
"duongbinh214",
"kameleo-team"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7446",
"repo": "kameleo-io/local-api-client-csharp",
"url": "https://github.com/kameleo-io/local-api-client-csharp/issues/2"
}
|
gharchive/issue
|
'No connection could be made because the target machine actively refused it. (localhost:5050)'
My code :
Error:
var baseProfileList = await client.SearchBaseProfilesAsync(deviceType: "desktop", browserProduct: "chrome");
pls help me!!!!!!
Are you sure that Kameleo.CLI.exe is running. If yes, please also make sure it runs on 5050 port.
Please see this article to see how to start Kameleo.CLI.exe
|
2025-04-01T06:39:15.727501
| 2017-03-24T09:35:52
|
216716915
|
{
"authors": [
"kamilsk"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7447",
"repo": "kamilsk/dzone",
"url": "https://github.com/kamilsk/dzone/issues/34"
}
|
gharchive/issue
|
integrate easyjson
https://github.com/mailru/easyjson
as a result of https://github.com/kamilsk/dzone/issues/31
blocked by https://github.com/kamilsk/shared/issues/144
|
2025-04-01T06:39:15.731751
| 2015-01-28T18:45:30
|
55797577
|
{
"authors": [
"kamisama",
"shadyb"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7448",
"repo": "kamisama/Fresque",
"url": "https://github.com/kamisama/Fresque/issues/45"
}
|
gharchive/issue
|
Specifying worker names with stop command
I am going to implement a feature that allows users to specify the worker to kill in command line. This is useful if you are using some automation/CI tools (jenkins, ansible, puppet etc). There are going to be two additional options to the stop command: --worker and --count.
The --worker option allows you to specify a queue name to kill.
The --count option works in tandem with the --worker option and allows you to specify how many to kill.
I'm not sure if you accept feature additions to your codebase, but if you want, I can make a pull request, implement it and then merge back into your repo. That way others can benefit.
Let me know what you think
I don't quite get what you're trying to do. The stop command already print a list of workers, and you can choose the worker you want to stop.
I think there's naming issue, but are you trying to stop a worker, by the name of the polled queue ? In that case, a --queue option would make more sense.
Sorry, I mean't --queue (been working very long hours as of late so am tripping up all over the place). It does print a list of workers, but then it requires stdin input and that's a problem if you are using automation tools.
Take my configuration for example, we have a jenkins project which allows you to specify queues to start. I make a selection and it plugs the variables into a bash script and runs it.
|
2025-04-01T06:39:15.735156
| 2019-10-14T14:00:10
|
506672585
|
{
"authors": [
"kamleshchandnani",
"rohandaxini"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7449",
"repo": "kamleshchandnani/awesome-interview-process",
"url": "https://github.com/kamleshchandnani/awesome-interview-process/pull/10"
}
|
gharchive/pull-request
|
Fix typo and improve wording in README
Hello @kamleshchandnani
Great initiative ๐
Just a minor suggestion if you don't mind ๐
I have fixed following in the README
Typo
Improved the header. You may refer hemingwayapp for the same.
You may review the fixes here.
PS: We will also raise a PR soon to add Kiprosh to the list.
Hey @rohandaxini ,
Thank you so much for taking out time and contributing to this project ๐
|
2025-04-01T06:39:15.740871
| 2024-04-24T07:09:01
|
2260542420
|
{
"authors": [
"filgoBot"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7450",
"repo": "kamp-us/monorepo",
"url": "https://github.com/kamp-us/monorepo/issues/863"
}
|
gharchive/issue
|
Curriculum update needed on how_does_the_web_work.md
The Odin's file, how_does_the_web_work.md is updated. Please update the Kampus' file, checkout file here how_does_the_web_work.md
Latest commits:
How Does the Web Work?: Update descriptive link text (#27681) (#27689) (additions: 7, deletions: 7) on Mar 26 2024, 14:43 UTC
New commits have been made to the Odin's file. Please update the Kampus' file.
Latest commits:
Surround html tag with lines to follow lint (#28132) (additions: 6, deletions: 4) on Jun 8 2024, 02:29 UTC
|
2025-04-01T06:39:15.744164
| 2016-09-08T11:46:57
|
175733035
|
{
"authors": [
"kamsar",
"waaromikniet"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7451",
"repo": "kamsar/Unicorn",
"url": "https://github.com/kamsar/Unicorn/issues/170"
}
|
gharchive/issue
|
Getting Server cannot append header after HTTP headers have been sent
Hi,
I am getting Server cannot append header after HTTP headers have been sent when I try to sync the roles with unicorn. Any ideas?
Regards
Danny
Duplicate of #155. This is fixed in the current prerelease version already.
I am assuming that you're using Sitecore 8.1U3 or 8.2. You can work around this by setting the 'X-Frame-Options' header yourself prior to Sitecore trying to set it on request end.
|
2025-04-01T06:39:15.818956
| 2016-06-16T08:08:15
|
160601638
|
{
"authors": [
"chenbs",
"emersion",
"karalabe"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7452",
"repo": "karalabe/xgo",
"url": "https://github.com/karalabe/xgo/issues/55"
}
|
gharchive/issue
|
windows: missing TaskDialog
Fixed in a recent mingw-w64
# github.com/andlabs/ui
../../andlabs/ui/libui_windows_amd64.a(stddialogs.cpp.obj): In function `msgbox':
/home/simon/projects/libui/windows/stddialogs.cpp:113: undefined reference to `__imp_TaskDialog'
collect2: error: ld returned 1 exit status
how to resolve ?
This seems to only be part of mingw 5+, which has not yet been released.
|
2025-04-01T06:39:15.839083
| 2017-12-22T14:06:08
|
284171299
|
{
"authors": [
"Shaileshz204",
"avpavlov",
"ptrthomas"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7453",
"repo": "karatelabs/karate",
"url": "https://github.com/karatelabs/karate/issues/270"
}
|
gharchive/issue
|
[feature request] Support for #notpresent keyword
Use case: I'm writing tests for Spring Boot Repository. It supports "projections" so user can request different detalization, e.g.
"default" projection: Fields A,B,C
"more-data" projection: Fields A,B,C and array D (it is loaded from separate table)
"complete-data" projection: Fields A,B,C and arrays D,E,F,G (each is loaded from separate table)
I cannot check contains/!contains with single JSON so right now my tests look like
Scenario "default"
<request>
then match response contains
"""
{
A : '#string',
B : '#string' ,
C : '#string'
}
"""
and match response !contains
"""
{
D : '#array',
E : '#array',
F : '#array',
G : '#array'
}
"""
Scenario "more-data"
<request>
then match response contains
"""
{
A : '#string',
B : '#string' ,
C : '#string',
D : '#array'
}
"""
and match response !contains
"""
{
E : '#array',
F : '#array',
G : '#array'
}
"""
Scenario "complete-data"
<request>
then match response contains
"""
{
A : '#string',
B : '#string' ,
C : '#string',
D : '#array',
E : '#array',
F : '#array',
G : '#array'
}
"""
With new keyword added my scenarios will be cleaner and easier to read
Scenario "default"
<request>
then match response contains
"""
{
A : '#string',
B : '#string' ,
C : '#string'
D : '#undefined',
E : '#undefined',
F : '#undefined,
G : '#undefined'
}
"""
Scenario "more-data"
<request>
then match response contains
"""
{
A : '#string',
B : '#string' ,
C : '#string',
D : '#array',
E : '#undefined',
F : '#undefined,
G : '#undefined'
}
"""
Scenario "complete-data"
<request>
then match response contains
"""
{
A : '#string',
B : '#string' ,
C : '#string',
D : '#array',
E : '#array',
F : '#array',
G : '#array'
}
"""
@avpavlov actually there is an #ignore marker already, can you confirm that it is what works for you: https://github.com/intuit/karate#fuzzy-matching
#ignore excludes field from verification. I do not want to exclude, I want to ensure this field is not included.
In other words, #undefined is like local !contains for single field
would #notnull work ?
I meant #null
Thank you, both work - #null and ##null.
However, I would say working #null is a bug in this case, because if I expect some key with null value then missing key should fail scenario.
##null is right fit in this case - any chances you expand "contains / !contains" readme sections to propose it as way to verify key is not presented in JSON?
However, I would say working #null is a bug in this case, because if I expect some key with null value then missing key should fail scenario.
Yes. Or how about this, to check for a null just use null itself. So this should fail (as of now it passes, but I'm proposing to change this):
* def foo = { }
* match foo == { a: null }
In my experience, most teams assume a null value and the key missing to be the same. Typically people set the Json marshaller config to 'strip nulls' for example, to reduce payload bloat.
Can you help by suggesting what change you'd like to see in the contains readme section, I'll be happy to add.
(as of now it passes, but I'm proposing to change this):
Sounds good
most teams assume a null value and the key missing to be the same.
That's true. That's why #undefined could help ;)
Can you help by suggesting what change you'd like to see in the contains readme section,
Let me think
@avpavlov cool, I'm beginning to agree. see I'm quite reasonable :P
how about #notpresent - which could be more clear ? undefined has a certain meaning in JS also ?
notpresent sounds better than undefined. I tried to invent something like na or notavailable or missed but all of these were not enough clear so finally I borrowed keyword from JS
great ! looking at it now. are you able to build from source or do you prefer a release.
Created PR for '##null' in README https://github.com/intuit/karate/pull/271
I can build from sources
@avpavlov yes, is now in the develop branch. I actually decided that both #null and the null value will expect the JSON key to be present. Just felt that this is consistent and reduces confusion. Teams may need to use #ignore or #notpresent which is more clear.
I agree, Karate tests output/protocol not how this output could be interpreted by consumer
Just tested project with develop branch - it works! Thank you!
Is there any way in the Github to subscribe to the notifications about releases?
@avpavlov not sure, but you can watch the project on GiHub. there's a twitter account if you are in to that kind of thing: https://twitter.com/KarateDSL
@avpavlov well. after thinking about it, it made sense to implement #present as well ! thanks for triggering this, I think Karate has become a little better as a result. So now here is what is possible. I recommend that folks don't use != not-equals because it gets really confusing, but hey.
* def foo = { }
* match foo != { a: '#present' }
* match foo == { a: '#notpresent' }
* match foo == { a: '#ignore' }
* match foo == { a: '##null' }
* match foo != { a: '#null' }
* match foo != { a: '#notnull' }
* match foo == { a: '##notnull' }
* match foo != { a: null }
* def foo = { a: null }
* match foo == { a: '#null' }
* match foo == { a: '##null' }
* match foo != { a: '#notnull' }
* match foo != { a: '##notnull' }
* match foo == { a: '#present' }
* match foo == { a: '#ignore' }
* match foo != { a: '#notpresent' }
* def foo = { a: 1 }
* match foo == { a: '#notnull' }
* match foo == { a: '##notnull' }
* match foo != { a: '#null' }
* match foo != { a: '##null' }
* match foo == { a: '#present' }
* match foo == { a: '#ignore' }
* match foo != { a: '#notpresent' }
@ptrthomas When I am validating schema using Karate one of the field as per schema is defined as String but its returning null but it returns string value also in some cases so when its returning null its failing my test as its string. How to handle this. I want it to pass if its null or string when its string in schema.
@Shaileshz204 use stack overflow for questions like this please: https://stackoverflow.com/a/71522605/143475
|
2025-04-01T06:39:15.841799
| 2017-09-13T08:18:10
|
257294817
|
{
"authors": [
"brunetto",
"kardianos"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7454",
"repo": "kardianos/govendor",
"url": "https://github.com/kardianos/govendor/issues/359"
}
|
gharchive/issue
|
Add tests to vendoring only for specific packages
Hi,
I'm sorry if it is trivial but I was not able to find any hint.
Is there a way to add tests only for specific packages in the vendor folder?
I can ignore the tests for all the packages with "ignore" in vendor.json file or, without it, all the tests for all the packages are added.
Nope. There is no way to do a package specific ignore requirement.
Ok, thanks... and is there a way to avoid having tests ignored in vendor.json after govendor init?
You can always remove that after a govendor init. It is just what I consider a "sane default".
Yeah it is a sane default that unfortunately does not apply to my employer default pipeline. I'm trying to find the best way to deal with it. Thanks a lot.
|
2025-04-01T06:39:15.901544
| 2024-08-17T10:40:16
|
2471471473
|
{
"authors": [
"dlmw"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7455",
"repo": "karlomikus/bar-assistant",
"url": "https://github.com/karlomikus/bar-assistant/issues/312"
}
|
gharchive/issue
|
Recipe images not retrieved
[x] I have read the FAQ.
Describe the bug
I have just created my Bar Assistant instance. Everything works fine but the images don't appear. However, thumbnails work fine.
After inspecting the HTML, I see that the URI of the picture is "https://bar.mydomain.com/uploads/cocktails/1/1934-cosmo-1_79lTcl.jpg", but it should be "https://bar.dlmw.ch/bar/uploads/cocktails/1/1934-cosmo-1_79lTcl.jpg". My reverse proxy is Caddy and the Caddyfile looks somewhat like this:
bar.mydomain.com {
handle_path /search/* {
reverse_proxy meilisearch:7700
}
handle_path /bar/* {
reverse_proxy barassistant:3000
}
handle_path /* {
reverse_proxy saltrim:8080
}
}
To Reproduce
Create a Caddyfile containing the snippet above
Versions:
Docker: 24.0.7
LibreWolf: 129.0-1
I managed to solve it by modifying the Caddyfile like so:
bar.dlmw.ch {
# here
@uploads {
path_regexp uploads ^/uploads/(.*)
}
rewrite @uploads /bar/uploads/{re.uploads.1}
handle_path /search/* {
reverse_proxy meilisearch:7700
}
handle_path /bar/* {
reverse_proxy barassistant:3000
}
handle_path /* {
reverse_proxy saltrim:8080
}
}
|
2025-04-01T06:39:15.905812
| 2015-05-12T21:15:28
|
75741338
|
{
"authors": [
"MetaMemoryT",
"maksimr",
"zzo"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7456",
"repo": "karma-runner/karma",
"url": "https://github.com/karma-runner/karma/issues/1402"
}
|
gharchive/issue
|
npm install WARN message
with npm install of karma, npm warns:
npm WARN engine<EMAIL_ADDRESS>wanted: {"node":"~0.8 || ~0.10"} (current: {"node":"0.12.0","npm":"2.7.2"})
#merge
Will be fixed in new version karma
|
2025-04-01T06:39:15.912869
| 2016-06-27T14:01:39
|
162459166
|
{
"authors": [
"cellog",
"dignifiedquire",
"johnjbarton",
"trusktr",
"wesleycho"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7457",
"repo": "karma-runner/karma",
"url": "https://github.com/karma-runner/karma/issues/2211"
}
|
gharchive/issue
|
when karma hangs, how do I debug it?
In my project, https://github.com/cellog/react-selection , I have encountered a bug when running karma remotely on saucelabs. It is caused by code coverage. In short, karma connects to the browser, loads, and then hangs indefinitely. When I remove code coverage, the test runs in about 16 seconds. This happens only remotely, when I run coverage on my local machine, it works great.
How can I debug the hang on saucelabs? I'm new to karma. When using karma remotely with saucelabs, is it loading karma from my machine? If yes, which file can I put debug code into to see if I can figure out where it is failing? Or is there another way?
Thanks much.
Oh and I should mention the code coverage is accomplished by a babel plugin, so there is nothing inside karma, it just serving the transpiled files directly to saucelabs
setting logLevel to DEBUG should give you more details, after that regular debugging techniques for node loke console.log in code and attaching a debugger
it turns out that any concurrency larger than 1 causes the entire thing to fail with disconnects to every browser. Is this expected behavior?
On saucelabs that is probably caused by rate limiting from their side.
fyi, the problem is in the karma-saucelabs-launcher, which is unmaintained and doesn't work any more. After 50 tests, it just dies. I switched to browserstack and it works perfectly. However, there are some side effects of running more than 1 browser test in the same karma process. I have no idea how to isolate them, but the ONLY way I can get karma to work and generate code coverage is to run each browser in its own karma process. You can see what I mean in https://github.com/cellog/react-selection
Major pain.
Fortunately, I can do local development with karma very quickly, and then push on commit to find those chance browser differences eventually. Test runs now take about 15 minutes each, so it's really a pain, but at least it works.
So to summarize: there are several bugs that I can't fix or even track down. The first is that karma and sauce labs are no longer friends, and so I can't use them together at all. The second is that karma has some kind of strange shared stuff interfering with each other whenever I run more than 1 browser in the same karma process, even if concurrency is set to 1. That's a huge one, and might be worth investigating further, since it means the sandboxing is leaking. Let me know if you want me to try things to debug it.
FWIW, I have noticed that the Angular team has set up karma with Sauce Labs to use Sauce Labs via shell script in separate processes - if I had to guess, it is to have environmental separation for purer test environments, but maybe they came across some of the same issues with multiple browsers.
Separate issues should probably be opened if there are specific issues found though, that way we can track them better. Going to close this issue, but feel free to open issues for actionable problems you find with karma.
attaching a debugger
How exactly? I tried
ndb ./node_modules/.bin/karma ... but nothing runs, and the opened devtools window is empty, no source files, nothing paused.
The correct answer is to move to jest and forget about it
Unfortunately I've found ndb fail sometimes. I use node --inspect-brk but I also don't use the .bin file directly.
|
2025-04-01T06:39:15.917482
| 2021-02-08T11:52:14
|
803491200
|
{
"authors": [
"AppVeyorBot",
"devoto13",
"jimbojw",
"maksimr",
"xel23"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7458",
"repo": "karma-runner/karma",
"url": "https://github.com/karma-runner/karma/pull/3651"
}
|
gharchive/pull-request
|
fix(config): check extension before ts-node register
Call require('ts-node').register() after checking configFilePath has .ts extension
Fixes #3329
ัั @devoto13
:white_check_mark: Build karma 2927 completed (commit https://github.com/karma-runner/karma/commit/4157266d7a by @xel23)
@googlebot I signed it!
@xel23 Thanks for the PR!
I guess we can land it as a workaround to solve the most painful appearance of this issue, but I don't think it "fixes" the issue. We really should implement a systematic solution as outlined in https://github.com/karma-runner/karma/issues/3329#issuecomment-772262377.
Technically this is a breaking change. People, who might have relied on type-checking their karma.conf.js with allowJS: true will no longer have the type-checking. As this does not prevent them from running Karma, I am willing to do so to resolve the issue for the majority of users. @johnjbarton What do you think?
Ping
Thank you so much for fixing this! ๐ฅณ ๐
|
2025-04-01T06:39:15.953840
| 2023-06-13T02:53:36
|
1753934976
|
{
"authors": [
"TheUltDev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7459",
"repo": "kat-tax/vslite",
"url": "https://github.com/kat-tax/vslite/pull/21"
}
|
gharchive/pull-request
|
Feat/collaboration
This adds Figma -> RN plugin syncing as well as generic editor syncing:
https://feat-collaboration.vslite.pages.dev/#/strait_domestic_heat_dean_patronage
Right now you need to run your own local websocket server:
HOST=localhost PORT=1234 npx y-websocket
All of this is work in progress, but the PoC is working...
Merged initial sync support. Will follow up on this in another PR.
|
2025-04-01T06:39:15.955203
| 2019-05-20T10:52:45
|
446050339
|
{
"authors": [
"jodh-intel"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7460",
"repo": "kata-containers/documentation",
"url": "https://github.com/kata-containers/documentation/pull/482"
}
|
gharchive/pull-request
|
docs: Added missing doc link
Added a link to the existing how-to-use-virtio-fs-with-kata.md.
Fixes #481.
Signed-off-by: James O. D. Hunt<EMAIL_ADDRESS>
/test
|
2025-04-01T06:39:15.980828
| 2019-05-10T09:35:01
|
442627585
|
{
"authors": [
"egernst",
"grahamwhaley",
"jodh-intel",
"zhiminghufighting"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7461",
"repo": "kata-containers/runtime",
"url": "https://github.com/kata-containers/runtime/issues/1661"
}
|
gharchive/issue
|
qemu-lite failed to enable vga when launching Kata container
Description of problem
When I reconfigure Qemu-lite launch parameter from โ-vga noneโ to โ-vga stdโ to enable vga in Kata runtime, it reports PCI unavailable error as below.
Is there any dependence to enable vga? I ask some people and there is no dependence in kvm +qemu to launch an centos VM.
Expected result
Launch Kata container successfully.
Actual result
May 10 06:31:21 localhost.localdomain kata-runtime[28005]: time="2019-05-10T06:31:21.574044938-04:00" level=info msg="launching /usr/bin/qemu-lite-system-x86_64 with: [-name sandbox-c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b -uuid 3e6695d3-d12f-4bfe-b1e5-2369c0b96df2 -machine pc,accel=kvm,kernel_irqchip,nvdimm -cpu host -qmp unix:/run/vc/vm/c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b/qmp.sock,server,nowait -m 2048M,slots=10,maxmem=8752M -device pci-bridge,bus=pci.0,id=pci-bridge-0,chassis_nr=1,shpc=on,addr=2,romfile= -device virtio-serial-pci,disable-modern=false,id=serial0,romfile= -device virtconsole,chardev=charconsole0,id=console0 -chardev socket,id=charconsole0,path=/run/vc/vm/c522a66fa441bb1eb5
be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b/console.sock,server,nowait -device nvdimm,id=nv0,memdev=mem0 -object memory-backend-file,id=mem0,mem-path=/usr/share/kata-containers/kata-containers-image_clearlinux_1.5.0_agent_a581aebf473.img,size=536870912 -device virtio-scsi-pci,id=scsi0,disable-modern=false,romfile= -object rng-random,id=rng0,filename=/dev/urandom -device virtio-rng,rng=rng0,romfile= -device virtserialport,chardev=charch0,id=channel0,name=agent.channel.0 -chardev socket,id=charch0,path=/run/vc/vm/c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2
f51d077a5dba7732987b/kata.sock,server,nowait -device virtio-9p-pci,disable-modern=false,fsdev=extra-9p-kataShared,mount_tag=kataShared,romfile= -fsdev local,id=extra-9p-kataShared,path=/ru
n/kata-containers/shared/sandboxes/c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b,security_model=none -netdev tap,id=network-0,vhost=on,vhostfds=3,fds=4 -device driver=virtio-net-pci,netdev=network-0,mac=02:42:ac:12:00:02,disable-modern=false,mq=on,vectors=4,romfile= -global kvm-pit.lost_tick_policy=discard -vga std -no-user-config -nodefaults -nographic -
daemonize -kernel /usr/share/kata-containers/vmlinuz-<IP_ADDRESS>-143.1.container -append tsc=reliable no_timer_check rcupdate.rcu_expedited=1 i8042.direct=1 i8042.dumbkbd=1 i8042.nopnp=1 i8042.noaux=1 noreplace-smp reboot=k console=hvc0 console=hvc1 iommu=off cryptomgr.notests net.ifnames=0 pci=lastbus=0 root=/dev/pmem0p1 rootflags=dax,data=ordered,errors=remount-ro rw rootf
stype=ext4 debug systemd.show_status=true systemd.log_level=debug panic=1 nr_cpus=8 init=/usr/lib/systemd/systemd systemd.unit=kata-containers.target systemd.mask=systemd-networkd.service systemd.mask=systemd-networkd.socket -smp 1,cores=1,threads=1,sockets=1,maxcpus=8]" arch=amd64 clicreate container=c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b command=
create name=kata-runtime pid=28005 source=virtcontainers subsystem=qmp
level=error msg="Unable to launch /usr/bin/qemu-lite-system-x86_64: exit status 1" arc
h=amd64 clicreate container=c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d077a5dba7732987b command=create name=kata-runtime pid=28005 source=virtcontainers subsystem=qmp
May 10 06:31:21 localhost.localdomain kata-runtime[28005]: time="2019-05-10T06:31:21.633469001-04:00" level=error msg="qemu-lite-system-x86_64: -device pci-bridge,bus=pci.0,id=pci-bridge-0,chassis_nr=1,shpc=on,addr=2,romfile=: PCI: slot 2 function 0 not available for pci-bridge, in use by VGA\n" arch=amd64 clicreate container=c522a66fa441bb1eb5be1b2fcba79b7829e8922c09d2f51d
077a5dba7732987b command=create name=kata-runtime pid=28005 source=virtcontainers subsystem=qmp
Hi @zhiminghufighting - what are you trying to achieve? Were you hoping that the container would then gain access to the hosts VGA controller, or? It might help us understand your goal and how to achieve it :-)
For reference, in case it helps, there is a guide on how to pass GPU in/out of Kata at https://github.com/kata-containers/documentation/blob/master/use-cases/GPU-passthrough-and-Kata.md
And there is a link to a container that can help get X11 working in a container (including a kata container) at https://github.com/kata-containers/documentation/wiki/UserGuide#x11-containers
@zhiminghufighting - qemu-lite is extremely minimal (by design), so doesn't have graphical support enabled. We disable all graphics at build time - see:
https://github.com/kata-containers/packaging/blob/master/scripts/configure-hypervisor.sh#L206
@grahamwhaley The background is here: i want to launch android based container image in Kata which is used to support cloud gaming. As you know, lots of gaming is running on billion mobile devices with Android OS in China. there is a big request to enable android container in Kata. After i launch Kata container with android based image with our OTC android in container team and found there is an android key process --surface finger can't be started normally caused by wrong framer buffer size. But if i need to get the framer buffer size, i need to enable vga in qemu-lite to support right framer buffer to start software render. Here android container needs a software render and framer buffer & virtual VGA is a mandatory option. The android container doesn't care about if there is a real VGA or GPU in host because it depends on software render library --opengl in android container image.
In above scenario, there is no need of hardware render in android container, so it doesn't depend on GPU GVT-g or GVT-d. I know and check all the related doc of enabling GPUs for kata in your link.
Thanks for your quick response.
Is there any other options for me to use vga?
@jodh-intel please see the detail background and root cause of this requirement in above answer to Graham's question.
I think you already answer the root cause of why the error log is reported by qemu-lite.
Can i enable gpu support by rebuild qemu-lite? If yes, is there any guide doc?
And if there is any other ways? I think even i enable gpu support to solve this issue, it will cause qemu-lite being heavy and increase the memory footprint and resource consumption.
thanks for your quick response!
By the way, there would be a huge potential user scenario for Kata beside cloud is mentioned by me in above comment : gaming in cloud and cloud gaming based on android image container in chine.
Billions of android mobile device and thousands of gaming based on android OS.
@zhiminghufighting - qemu-lite is a highly optimised version of qemu designed for "standard" sorts of container use-cases so doesn't include graphical support.
To just prove the concept, you could of course simply change your config to use the distro-packaged version of qemu which should contain graphics support:
path = "/usr/bin/qemu-system-x86_64"
Although you could rebuild qemu-lite to include graphic support, I don't think that is the best approach. Depending on the architecture, either NEMU or qemu-vanilla would be better I think.
However, both those options are also built without graphical support so you would have to rebuild them with some of the graphical options we disable using https://github.com/kata-containers/packaging/blob/master/scripts/configure-hypervisor.sh.
However, you may need to make changes to the guest kernel configuration as that too is as minimal as possible for "normal" sorts of workloads.
We don't have documentation for this sort of scenario currently.
Next steps
This is an interesting scenario but since it is going to require changes specifically for graphical support and since those may adversely affect memory density and boot speed, we'd need to study the idea carefully. We would also need additional tests, extra testing infrastructure and documentation.
As such, I suggest you raise an RFC issue using the main project repo (which we tend to use to discuss large features that potentially affect lots of different repos):
https://github.com/kata-containers/kata-containers/issues/new
Then, send a mail to the mailing list referring to the issue and ideally present your idea at the Architecture Committee meeting so we can get more input from the entire community.
@jodh-intel Good suggestion!I will collect the detail background and estimate the potential business impact for android image in kata container firstly and then prepare some material for the this user scenario. thanks for your inputs!
I will try the ways provided by you to make sure the feasibility of android in Kata container in next days.
Thanks @zhiminghufighting.
@zhiminghufighting - I realize this is pretty stale now, but was pretty curious about the Android in Kata use case.
Have you made more progress here, or have an end-to-end demo in this space?
|
2025-04-01T06:39:15.997901
| 2021-07-27T01:29:13
|
953416232
|
{
"authors": [
"KFlash",
"aladdin-add"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7462",
"repo": "kataw/kataw",
"url": "https://github.com/kataw/kataw/pull/162"
}
|
gharchive/pull-request
|
chore: add linting as public API
as requested by @aladdin-add this PR implements linting as public API. Linting can now be done like this
import { lintModule, aladdin } from 'kataw';
lintModule('eval', /* reporter */ aladdin, { noEval: true});
or
import { lintScript, aladdin } from 'kataw';
lintScript('eval', /* reporter */ aladdin, { noEval: true});
The 'reporter' argument gives the option to add any reporters and end-users can make their own reporter.
I tried to figure out how Babel does this, but I failed in my research.
Note This PR is blocked by #160
This is how it looks like on the command line with the aladdin reporter
@aladdin-add I'm using the aladdin reporter by default here so this can't be merged before you have merged #160.
In the future we will add a 3rd arg into this so the end-user can choose which reporter to use.
You also requested for custom rules, right? They are coming soon. ESLint is horrible slow you know ;)
lintModule('eval', /* reporter */ aladdin, { noEval: true});
the 2nd param can be an linterOptions:
{
reporter: ...,
globals: {...},
fix: true,
}
@aladdin-add Good idea! But what is Globals in this context? And can you fix the conflict?
@aladdin-add I changed it into what you suggested, but now we got an performance issue ...
Every time we invoke the reporter we need to do 'options.reporter(...)'. It's unnecessary property access.
Can you get rid of this unnecessary property access?
|
2025-04-01T06:39:16.024614
| 2020-07-18T14:12:47
|
660173422
|
{
"authors": [
"razttt",
"rt-2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7463",
"repo": "katursis/Pawn.RakNet",
"url": "https://github.com/katursis/Pawn.RakNet/issues/36"
}
|
gharchive/issue
|
Server crashing on player connect
Same as https://github.com/urShadow/Pawn.RakNet/issues/35
I can see the server, but as soon as I connect, the server crashes.
There is NOTHING in the log at the moment of the crash (Not even the 'incoming connection' message).
I simply removed the plugin in server.cfg and in the script and everything worked fine instantly.
I will have to make more tests when I will have a couple of hours to debug this I will try to see what other includes/plugins might be interfering.
content of serverlog.txt :
----------
Loaded log file: "server_log.txt".
----------
SA-MP Dedicated Server
----------------------
v0.3.7-R2, (C)2005-2015 SA-MP Team
[2020-07-18 09:40:19]
[2020-07-18 09:40:19] Server Plugins
[2020-07-18 09:40:19] --------------
[2020-07-18 09:40:19] Loading plugin: pawnraknet
[2020-07-18 09:40:19] [Pawn.RakNet]
| Pawn.RakNet 1.4.1 | 2016 - 2020
|--------------------------------
| Author and maintainer: urShadow
| Compiled: Jun 18 2020 at 14:19:11
|--------------------------------------------------------------
| Forum thread: https://forum.sa-mp.com/showthread.php?t=640306
|--------------------------------------------------------------
| Repository: https://github.com/urShadow/Pawn.RakNet
|--------------------------------------------------------------
| Wiki: https://github.com/urShadow/Pawn.RakNet/wiki
[2020-07-18 09:40:19] Loaded.
[2020-07-18 09:40:19] Loading plugin: crashdetect
[2020-07-18 09:40:19] CrashDetect plugin 4.19
[2020-07-18 09:40:19] Loaded.
[2020-07-18 09:40:19] Loading plugin: mysql
[2020-07-18 09:40:19] >> plugin.mysql: R39-6 successfully loaded.
[2020-07-18 09:40:19] Loaded.
[2020-07-18 09:40:19] Loading plugin: streamer
[2020-07-18 09:40:19]
*** Streamer Plugin v2.9.3 by Incognito loaded ***
[2020-07-18 09:40:19] Loaded.
[2020-07-18 09:40:19] Loading plugin: FileFunctions
[2020-07-18 09:40:19] Loaded.
[2020-07-18 09:40:19] Loading plugin: FCNPC
[2020-07-18 09:40:19]
[2020-07-18 09:40:19] -------------------------------------------------
[2020-07-18 09:40:19] FCNPC - Fully Controllable NPC v1.8.2
[2020-07-18 09:40:19] Windows SA-MP 0.3.7 R2
[2020-07-18 09:40:19] Jan 8 2018 at 01:14:48
[2020-07-18 09:40:19]
[2020-07-18 09:40:19] Author: OrMisicL (2013 - 2015)
[2020-07-18 09:40:19] Continued by: ziggi (2016 - present)
[2020-07-18 09:40:19] Contributors: kurta999, Neutralneu
[2020-07-18 09:40:19] -------------------------------------------------
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] Loading...
[2020-07-18 09:40:20] Loaded.
[2020-07-18 09:40:20] Loading plugin: PathFinder
[2020-07-18 09:40:20] =========================================
[2020-07-18 09:40:20] PathFinder Plugin 1.0 MT
[2020-07-18 09:40:20] by Pamdex
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] Using MapAndreas Plugin 1.2.1
[2020-07-18 09:40:20] Waiting for Init...
[2020-07-18 09:40:20] =========================================
[2020-07-18 09:40:20] Loaded.
[2020-07-18 09:40:20] Loading plugin: MapAndreas
[2020-07-18 09:40:20] Loaded.
[2020-07-18 09:40:20] Loading plugin: SKY
[2020-07-18 09:40:20] Loaded.
[2020-07-18 09:40:20] Loading plugin: YSF
[2020-07-18 09:40:20] ARRAY_ConsoleCommands: 4e43d8
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] ===============================
[2020-07-18 09:40:20] YSF - kurta999's version R19 loaded
[2020-07-18 09:40:20] (c) 2008 Alex "Y_Less" Cole - (c) 2010 - 2016 kurta999
[2020-07-18 09:40:20] Server version: 0.3.7 R2-1
[2020-07-18 09:40:20] Operating System: Windows
[2020-07-18 09:40:20] Built on: Mar 11 2017 at 10:32:34
[2020-07-18 09:40:20] ===============================
[2020-07-18 09:40:20] Loaded.
[2020-07-18 09:40:20] Loading plugin: sscanf
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] ===============================
[2020-07-18 09:40:20] sscanf plugin loaded.
[2020-07-18 09:40:20] Version: 2.8.2
[2020-07-18 09:40:20] (c) 2012 Alex "Y_Less" Cole
[2020-07-18 09:40:20] ===============================
[2020-07-18 09:40:20] Loaded.
[2020-07-18 09:40:20] Loaded 11 plugins.
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] Ban list
[2020-07-18 09:40:20] --------
[2020-07-18 09:40:20] Loaded: samp.ban
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] Filterscripts
[2020-07-18 09:40:20] ---------------
[2020-07-18 09:40:20] Loading filterscript 'antiddos.amx'...
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] | YSI version 4.00.0001 |
[2020-07-18 09:40:20] | By Alex "Y_Less" Cole |
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] [TESTING]: SERVER_LOG_FIRSTCHAR:22;
[2020-07-18 09:40:20] Loaded!
[2020-07-18 09:40:20] Loading filterscript 'fac_test.amx'...
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] | YSI version 4.00.0001 |
[2020-07-18 09:40:20] | By Alex "Y_Less" Cole |
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] -black screens initialization...
[2020-07-18 09:40:20] -players variables
[2020-07-18 09:40:20] -black screens initialization completed.
[2020-07-18 09:40:20] -Loading Testing Faction...
[2020-07-18 09:40:20] [MYSQL]: Connection to `saarp` succesful!
[2020-07-18 09:40:20] -Loading Objects...
[2020-07-18 09:40:20] -Loading Pickups...
[2020-07-18 09:40:20] -Testing Faction loaded correctly...
[2020-07-18 09:40:20] Loading filterscript 'vehicles.amx'...
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] | YSI version 4.00.0001 |
[2020-07-18 09:40:20] | By Alex "Y_Less" Cole |
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] -black screens initialization...
[2020-07-18 09:40:20] -players variables
[2020-07-18 09:40:20] -black screens initialization completed.
[2020-07-18 09:40:20]
-----------------------------------------
[2020-07-18 09:40:20] Stefan/Kevin974 - Speedometer | rt-2 - Fuel/Engine system
[2020-07-18 09:40:20] -----------------------------------------
[2020-07-18 09:40:20] [VEHICLES]: Setting vehicle Interior positions
[2020-07-18 09:40:20] [VEHICLES]: Setting vehicle Interior external doors positions
[2020-07-18 09:40:20] [VEHICLES]: SA Driving Assoc
[2020-07-18 09:40:20] [VEHICLES]: Spawning cars from database
[2020-07-18 09:40:20] Loading filterscript 'doors.amx'...
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] | YSI version 4.00.0001 |
[2020-07-18 09:40:20] | By Alex "Y_Less" Cole |
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] -black screens initialization...
[2020-07-18 09:40:20] -players variables
[2020-07-18 09:40:20] -black screens initialization completed.
[2020-07-18 09:40:20] Loading City Planning faction and adresses...
[2020-07-18 09:40:20] -Connecting to database
[2020-07-18 09:40:20] -Acquiring vehicles coords
[2020-07-18 09:40:20] -Initializing doors vars
[2020-07-18 09:40:20] -Other variables and timers
[2020-07-18 09:40:20] Loading filterscript 'public.amx'...
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20]
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] | YSI version 4.00.0001 |
[2020-07-18 09:40:20] | By Alex "Y_Less" Cole |
[2020-07-18 09:40:20] | |
[2020-07-18 09:40:20] =======================================
[2020-07-18 09:40:20]
[2020-07-18 09:40:21] -black screens initialization...
[2020-07-18 09:40:21] -players variables
[2020-07-18 09:40:21] -black screens initialization completed.
[2020-07-18 09:40:21] # # # # # # # # # # # # # # # # # # # # # #
[2020-07-18 09:40:21] # # # # The city have a PT systems. # # # #
[2020-07-18 09:40:21] # # # # # # # # # # # # # # # # # # # # # #
[2020-07-18 09:40:21] # # Initializing "PT" System
[2020-07-18 09:40:21] # # Initializing other plugins
[2020-07-18 09:40:21] # # # Initializing FCNPC
[2020-07-18 09:40:21] # # # Initializing SAPT System
[2020-07-18 09:40:21] # # # Loading SAPT Routes
[2020-07-18 09:40:21] Loading filterscript 'testing.amx'...
[2020-07-18 09:40:21] Loaded 6 filterscripts.
[2020-07-18 09:40:22] Filterscript '../scriptfiles/callbackfix.amx' loaded.
[2020-07-18 09:40:22]
[2020-07-18 09:40:22]
[2020-07-18 09:40:22]
[2020-07-18 09:40:22] =======================================
[2020-07-18 09:40:22] | |
[2020-07-18 09:40:22] | YSI version 4.00.0001 |
[2020-07-18 09:40:22] | By Alex "Y_Less" Cole |
[2020-07-18 09:40:22] | |
[2020-07-18 09:40:22] =======================================
[2020-07-18 09:40:22]
[2020-07-18 09:40:25] -AC(new) initialization...
[2020-07-18 09:40:25] -players variables
[2020-07-18 09:40:25] -vehicles variables
[2020-07-18 09:40:25] -AC(new) initialization completed.
[2020-07-18 09:40:25] -black screens initialization...
[2020-07-18 09:40:25] -players variables
[2020-07-18 09:40:25] -black screens initialization completed.
[2020-07-18 09:40:25] -inventories initialization...
[2020-07-18 09:40:25] -general variables
[2020-07-18 09:40:25] -players variables
[2020-07-18 09:40:25] -inventories initialization completed.
[2020-07-18 09:40:26] -Logged Off Players Variables initializations...
[2020-07-18 09:40:26] -melee interactions initialization...
[2020-07-18 09:40:26] -players variables
[2020-07-18 09:40:26] -all players variables
[2020-07-18 09:40:26] -melee interactions initialization completed.
[2020-07-18 09:40:26] -Mailboxs initializations...
[2020-07-18 09:40:26] -SASD faction initializing
[2020-07-18 09:40:26] -Pickup(s)
[2020-07-18 09:40:26] -Setting routes
[2020-07-18 09:40:26] -SAPO faction initializing
[2020-07-18 09:40:26] -Pickup(s)
[2020-07-18 09:40:26] -Setting routes
[2020-07-18 09:40:26] -SATEL faction initializing
[2020-07-18 09:40:26] -Pickup(s)
[2020-07-18 09:40:26] -Setting routes
[2020-07-18 09:40:26] -SPRU faction initializing
[2020-07-18 09:40:26] -Pickup(s)
[2020-07-18 09:40:26] -Setting routes
[2020-07-18 09:40:26] -Missions initializations...
[2020-07-18 09:40:26] -Initializing player status.
[2020-07-18 09:40:26] -Initializing timer(s).
[2020-07-18 09:40:26] -Initializing text strings.
[2020-07-18 09:40:26] -Gang Wars initializations...
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Gang War system.
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Areas.
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Gangs Skins.
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Gangs Cribs.
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Gangs Pickups.
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Territories Wars Timer.
[2020-07-18 09:40:26] [GANGWAR]: -Initializing Other Gangs Vars.
[2020-07-18 09:40:26] -Interactive Menus initializations...
[2020-07-18 09:40:26] -Shops initializations...
[2020-07-18 09:40:26] -Sending query
[2020-07-18 09:40:26] -tutorial initialization...
[2020-07-18 09:40:26] -players variables
[2020-07-18 09:40:26] -office pickup
[2020-07-18 09:40:26] -slides variables
[2020-07-18 09:40:26] -Transport system initializations...
[2020-07-18 09:40:26] -Impex Automatic Sales initialization;
[2020-07-18 09:40:26] -stockid is 53 after load
[2020-07-18 09:40:26] -creating container objects;
[2020-07-18 09:40:26] -creating pickup;
[2020-07-18 09:40:26] -creating container variables;
[2020-07-18 09:40:26] -spawning cargos;
[2020-07-18 09:40:26] -digestion initialization...
[2020-07-18 09:40:26] -players variables
[2020-07-18 09:40:26] -digestion initialization completed.
[2020-07-18 09:40:26] -machines initialization...
[2020-07-18 09:40:26] -players variables
[2020-07-18 09:40:26] -machines variables
[2020-07-18 09:40:26] -machines recipes variables
[2020-07-18 09:40:26] -machine initialization completed.
[2020-07-18 09:40:26] -fire initialization...
[2020-07-18 09:40:26] -vehicles variables
[2020-07-18 09:40:26] -previous fires
[2020-07-18 09:40:26] -fire initialization completed.
[2020-07-18 09:40:26] -underground races initialization...
[2020-07-18 09:40:26] -general variables
[2020-07-18 09:40:26] [TESTING]: urace_racevar_reinit() called.
[2020-07-18 09:40:26] -racers variables
[2020-07-18 09:40:26] -players variables
[2020-07-18 09:40:26] -underground races initialization completed.
[2020-07-18 09:40:26] -fishing initialization...
[2020-07-18 09:40:26] -catch types
[2020-07-18 09:40:26] -players variables
[2020-07-18 09:40:26] -fishing initialization completed.
[2020-07-18 09:40:26] -elevators initialization...
[2020-07-18 09:40:26] -loading elevators...
[2020-07-18 09:40:26] -elevators initialization completed.
[2020-07-18 09:40:26] -Scripted Casinos initializations...
[2020-07-18 09:40:26]
----------------------------------
[2020-07-18 09:40:26]
[2020-07-18 09:40:26] STREAMER_OBJECT_SD =<PHONE_NUMBER>
[2020-07-18 09:40:26] STREAMER_OBJECT_DD = 0
[2020-07-18 09:40:26] NB OF OBJ LOADED: 5126/8000
[2020-07-18 09:40:26] ----------------------------------
[2020-07-18 09:40:26] -removed buildings initialization...
[2020-07-18 09:40:26] -vars initializing...
[2020-07-18 09:40:26] -building listing...
[2020-07-18 09:40:26] -testing configuration:
[2020-07-18 09:40:26] -679 building removed;
[2020-07-18 09:40:26] -modelid:923 is removed 3 times:
[2020-07-18 09:40:26] -modelid:1216 is removed 14 times:
[2020-07-18 09:40:26] -modelid:'traffic light'(1283) is removed 239 times:
[2020-07-18 09:40:26] -modelid:1284 is removed 17 times:
[2020-07-18 09:40:26] -modelid:1315 is removed 62 times:
[2020-07-18 09:40:26] -modelid:1350 is removed 20 times:
[2020-07-18 09:40:26] -modelid:1373 is removed 8 times:
[2020-07-18 09:40:26] -modelid:1374 is removed 8 times:
[2020-07-18 09:40:26] -modelid:'interior boxes'(1421) is removed 2 times:
[2020-07-18 09:40:26] -modelid:'interior boxes'(1431) is removed 2 times:
[2020-07-18 09:40:26] -modelid:1440 is removed 2 times:
[2020-07-18 09:40:26] -modelid:1441 is removed 2 times:
[2020-07-18 09:40:26] -modelid:2647 is removed 7 times:
[2020-07-18 09:40:26] -modelid:2663 is removed 6 times:
[2020-07-18 09:40:26] -modelid:2672 is removed 4 times:
[2020-07-18 09:40:26] -modelid:2673 is removed 2 times:
[2020-07-18 09:40:26] -modelid:2674 is removed 2 times:
[2020-07-18 09:40:26] -modelid:2675 is removed 3 times:
[2020-07-18 09:40:26] -modelid:2676 is removed 3 times:
[2020-07-18 09:40:26] -modelid:2677 is removed 3 times:
[2020-07-18 09:40:26] -modelid:3377 is removed 8 times:
[2020-07-18 09:40:26] -modelid:3378 is removed 8 times:
[2020-07-18 09:40:26] -modelid:3474 is removed 2 times:
[2020-07-18 09:40:26] -modelid:3516 is removed 4 times:
[2020-07-18 09:40:26] -modelid:3567 is removed 5 times:
[2020-07-18 09:40:26] -modelid:3569 is removed 5 times:
[2020-07-18 09:40:26] -modelid:'cargos'(3574) is removed 32 times:
[2020-07-18 09:40:26] -modelid:'abandonned car'(3593) is removed 11 times:
[2020-07-18 09:40:26] -modelid:'abandonned car'(3594) is removed 10 times:
[2020-07-18 09:40:26] -modelid:3621 is removed 5 times:
[2020-07-18 09:40:26] -modelid:3625 is removed 7 times:
[2020-07-18 09:40:26] -modelid:3664 is removed 4 times:
[2020-07-18 09:40:26] -modelid:3665 is removed 3 times:
[2020-07-18 09:40:26] -modelid:3688 is removed 5 times:
[2020-07-18 09:40:26] -modelid:3744 is removed 32 times:
[2020-07-18 09:40:26] -modelid:3747 is removed 5 times:
[2020-07-18 09:40:26] -modelid:3769 is removed 6 times:
[2020-07-18 09:40:26] -modelid:3780 is removed 3 times:
[2020-07-18 09:40:26] -Administration initializations...
[2020-07-18 09:40:26] -SAARP_fac_cityp_tcmd initializations...
[2020-07-18 09:40:26]
[2020-07-18 09:40:26] --------------------------------------
[2020-07-18 09:40:26] Anticheat Nex-AC loaded!
[2020-07-18 09:40:26] Anticheat version: 1.9.53
[2020-07-18 09:40:26] Author: Nexius
[2020-07-18 09:40:26] --------------------------------------
[2020-07-18 09:40:26] -Initializing principal game mode.
[2020-07-18 09:40:26] -Mysql connection.
[2020-07-18 09:40:26] -Setting Variables.
[2020-07-18 09:40:26] -Preparing streamer plugin...
[2020-07-18 09:40:26] -tick rate: 50;
[2020-07-18 09:40:26] -max pickups: -1;
[2020-07-18 09:40:26] -cell distance:<PHONE_NUMBER>;
[2020-07-18 09:40:26] -cell size: 300.000000;
[2020-07-18 09:40:26] -Preparing Main Textdraws...
[2020-07-18 09:40:26] -Spawning map icons...
[2020-07-18 09:40:26] -Plants initializations...
[2020-07-18 09:40:26] -Races initializations...
[2020-07-18 09:40:26] -Blood Stains initializations...
[2020-07-18 09:40:26] -Cheats initializations...
[2020-07-18 09:40:26] -Noob Path Helper initializations...
[2020-07-18 09:40:26] -Death Reasons initializations...
[2020-07-18 09:40:26] -Objects initializations...
[2020-07-18 09:40:26] -Acquiring object types
[2020-07-18 09:40:26] -Spawning ground objects models
[2020-07-18 09:40:26] -Spawning posts
[2020-07-18 09:40:26] -Vehicles interior initializations...
[2020-07-18 09:40:26] -Acquiring vehicles coords
[2020-07-18 09:40:26] -Spawning vehicle interior pickups
[2020-07-18 09:40:26] -Dialog Menus initializations...
[2020-07-18 09:40:26] -LS Stadiums initializations...
[2020-07-18 09:40:26] -Loading Police Forces faction
[2020-07-18 09:40:26] -Loading objects
[2020-07-18 09:40:26] -Loading pickups
[2020-07-18 09:40:26] [TESTING]: LSPD_Weapons_Pickup = 89
[2020-07-18 09:40:26] [TESTING]: SAMA_Armour_Pickup = 94
[2020-07-18 09:40:26] -Taxi faction initializing
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Loading SF Airport Management faction
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Loading LS International Airport faction
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Loading SF Military Police faction
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Loading LS Military Police faction.
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Loading SA Anti Terrorists faction.
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Spawning objects
[2020-07-18 09:40:26] -Mush faction initializing
[2020-07-18 09:40:26] -Spawning objects
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Mush faction initializing
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -City Planning faction initializing
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Spawning objects
[2020-07-18 09:40:26] -Gouv faction initializing (SADA/SAAA/SABA/SACB)
[2020-07-18 09:40:26] -Spawning objects
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] -Xoomer faction initializing
[2020-07-18 09:40:26] -Spawning pickups
[2020-07-18 09:40:26] [TESTING]: LSPD_Weapons_Pickup = 89
[2020-07-18 09:40:26] -IG Player Variables initializations...
[2020-07-18 09:40:26] -Vehicle Variables initializations...
[2020-07-18 09:40:26] -Other Variables initializations...
[2020-07-18 09:40:26] -Principal game mode initializing completed.
[2020-07-18 09:40:26]
Game mode ready!
[2020-07-18 09:40:26] Number of vehicle models: 0
[2020-07-18 09:40:26] [SERVER]: Vehicles are loading...
[2020-07-18 09:40:26] [MYSQL]: Receiving response on "SpawnVehicles", 861 rows.
[2020-07-18 09:40:36] [VEHICLES]: Vehicle spawned, 861 vehicles.
[2020-07-18 09:40:36] Spawning errors:
-Vehicule id:1240(m:588) does not have a vehicle interior view associated!;
-Vehicule id:1250(m:558) does not have a vehicle interior view associated!;
[2020-07-18 09:40:36] [MYSQL]: Receiving response on "getZonesFromDatabase", 43 rows.
[2020-07-18 09:40:36] [MYSQL]: Receiving response on "SpawnGazstations", 33 rows.
[2020-07-18 09:40:36] [DOORS]: Doors are loading...
[2020-07-18 09:40:36] [MYSQL]: Receiving response on "LoadDoors", 1486 rows.
[2020-07-18 09:40:46] [DOORS]: loaded 16/20 map icons.
[2020-07-18 09:40:46] [MYSQL]: Receiving response on "LoggedOff_InitAllAtStart_res", 857 rows.
[2020-07-18 09:40:46] -All Player Variables initializations...
[2020-07-18 09:40:51] [MYSQL]: Receiving response on "getAreasFromDatabase", 43 rows.
[2020-07-18 09:40:51] [MYSQL]: Receiving response on "ReceiveShopsList", 157 rows.
[2020-07-18 09:40:52] [MYSQL]: Receiving response on "GetGroundCargos", 193 rows.
[2020-07-18 09:40:52] [MYSQL]: Receiving response on "Machines_LoadFromDB", 16 rows.
[2020-07-18 09:40:53] [MYSQL]: Receiving response on "elevators_loadFromDatabase", 2 rows.
[2020-07-18 09:40:53] [MYSQL]: Receiving response on "LoadSAAGFields", 11 rows.
[2020-07-18 09:40:53] [MYSQL]: Receiving response on "LoadSAAGPlants", 73 rows.
[2020-07-18 09:40:53] [MYSQL]: Receiving response on "ReceiveGovOffices", 69 rows.
[2020-07-18 09:40:53] [MYSQL]: Receiving response on "ReceiveObjectTypes", 66 rows.
[2020-07-18 09:40:53] [MYSQL]: Receiving response on "SpawnGroundObjects", 4770 rows.
[2020-07-18 09:41:03] [MYSQL]: Receiving response on "SpawnPosts", 4 rows.
[2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(0)", 24 rows.
[2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(1)", 61 rows.
[2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(2)", 20 rows.
[2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(3)", 189 rows.
[2020-07-18 09:41:04] ## SASD ROUTE TOO LARGE ## ## will truncate r:1,t:fac_sasd_jy_lsscres from 189 to 65 ##
[2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(4)", 10 rows.
[2020-07-18 09:41:04] [MYSQL]: Receiving response on "GetSasdRouteFromDb(5)", 26 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(6)", 2 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(7)", 150 rows.
[2020-07-18 09:41:05] ## SASD ROUTE TOO LARGE ## ## will truncate r:2,t:fac_sasd_jy_lsnebiz from 150 to 65 ##
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(8)", 0 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(9)", 31 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(10)", 19 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(11)", 12 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "GetSasdRouteFromDb(12)", 9 rows.
[2020-07-18 09:41:05] [TESTING]: Finishing route, thisroute_max_leg:14, thisroute_total_leg:16
[2020-07-18 09:41:05] [TESTING]: Finishing route, thisroute_max_leg:5, thisroute_total_leg:7
[2020-07-18 09:41:05] [TESTING]: Finishing route, thisroute_max_leg:5, thisroute_total_leg:7
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "elevator_floor_loadFromDatabase", 21 rows.
[2020-07-18 09:41:05] [MYSQL]: Receiving response on "elevator_floor_loadFromDatabase", 11 rows.
[2020-07-18 09:41:05] [SERVER]: The time is 9:41
[2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:247, toggle:1) called.
[2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:318, toggle:1) called.
[2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:357, toggle:1) called.
[2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:418, toggle:1) called.
[2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:461, toggle:1) called.
[2020-07-18 09:41:10] fire_toggleVehFire(vehicleid:559, toggle:1) called.
[2020-07-18 09:41:28] [SERVER]: The time is 9:41
[2020-07-18 09:41:51] [SERVER]: The time is 9:41
[2020-07-18 09:42:13] [SERVER]: The time is 9:42
[2020-07-18 09:42:36] [SERVER]: Executing 'serverStart1'.
[2020-07-18 09:42:36]
[2020-07-18 09:42:36]
[2020-07-18 09:42:36]
[2020-07-18 09:42:36] =======================================
[2020-07-18 09:42:36] | |
[2020-07-18 09:42:36] | YSI version 4.00.0001 |
[2020-07-18 09:42:36] | By Alex "Y_Less" Cole |
[2020-07-18 09:42:36] | |
[2020-07-18 09:42:36] =======================================
[2020-07-18 09:42:36]
[2020-07-18 09:42:36] -black screens initialization...
[2020-07-18 09:42:36] -players variables
[2020-07-18 09:42:36] -black screens initialization completed.
[2020-07-18 09:42:36] รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ
[2020-07-18 09:42:36] รยฐ รยฐ รยฐ รยฐ The state is being infected รยฐ รยฐ รยฐ รยฐ
[2020-07-18 09:42:36] รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ รยฐ
[2020-07-18 09:42:36] รยฐ รยฐ Initializing "Infected" System
[2020-07-18 09:42:36] รยฐ รยฐ Initializing other plugins
[2020-07-18 09:42:36] รยฐ รยฐ รยฐ Initializing Map Andreas
[2020-07-18 09:42:37] PathFinder Plugin -> Creating New Thread
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ Initializing FCNPC
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ Initializing PathFinder
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ Generating spawn loacations.
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ initializing infected
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ initializing reserves
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ initializing timers
[2020-07-18 09:42:37] รยฐ รยฐ รยฐ initializing SQL connection
[2020-07-18 09:42:37] รยฐ รยฐ Script initialized
[2020-07-18 09:42:37] Filterscript 'infected.amx' loaded.
[2020-07-18 09:42:37]
----------------------------------
[2020-07-18 09:42:37] GOUVERNEMENT OBJECT SCRIPT
[2020-07-18 09:42:37] nz = 12.557999, stream is
[2020-07-18 09:42:37] NB OF OBJ LOADED: 4338/X
[2020-07-18 09:42:37] ----------------------------------
[2020-07-18 09:42:37] Filterscript 'mobjects.amx' loaded.
[2020-07-18 09:42:37]
[2020-07-18 09:42:37]
[2020-07-18 09:42:37]
[2020-07-18 09:42:37] =======================================
[2020-07-18 09:42:37] | |
[2020-07-18 09:42:37] | YSI version 4.00.0001 |
[2020-07-18 09:42:37] | By Alex "Y_Less" Cole |
[2020-07-18 09:42:37] | |
[2020-07-18 09:42:37] =======================================
[2020-07-18 09:42:37]
[2020-07-18 09:42:37] -black screens initialization...
[2020-07-18 09:42:37] -players variables
[2020-07-18 09:42:37] -black screens initialization completed.
[2020-07-18 09:42:37] -LSPD Tickets initialization...
[2020-07-18 09:42:37] -Loading Testing1 Faction...
[2020-07-18 09:42:37] -Loading Testing1337 Faction...
[2020-07-18 09:42:37] [MYSQL]: Connection to `saarp` succesful!
[2020-07-18 09:42:37] -Loading Objects...
[2020-07-18 09:42:37] -Loading Areas...
[2020-07-18 09:42:37] -Loading Vehicles vars...
[2020-07-18 09:42:37] -Testing1 Faction loaded correctly...
[2020-07-18 09:42:37] Filterscript 'fac_test1.amx' loaded.
[2020-07-18 09:42:37]
[2020-07-18 09:42:37]
[2020-07-18 09:42:37]
[2020-07-18 09:42:37] =======================================
[2020-07-18 09:42:37] | |
[2020-07-18 09:42:37] | YSI version 4.00.0001 |
[2020-07-18 09:42:37] | By Alex "Y_Less" Cole |
[2020-07-18 09:42:37] | |
[2020-07-18 09:42:37] =======================================
[2020-07-18 09:42:37]
[2020-07-18 09:42:38] -black screens initialization...
[2020-07-18 09:42:38] -players variables
[2020-07-18 09:42:38] -black screens initialization completed.
[2020-07-18 09:42:38] -setting BIOS environments
[2020-07-18 09:42:38] -computers variables
[2020-07-18 09:42:38] -players variables
[2020-07-18 09:42:38] -computer initialization...
[2020-07-18 09:42:38] -computer initialization completed.
[2020-07-18 09:42:38] -tele-comunication system initializations...
[2020-07-18 09:42:38] -Phone Objects adjustments
[2020-07-18 09:42:38] -Phone Variables
[2020-07-18 09:42:38] -Starting timers
[2020-07-18 09:42:38] -Loading Test2 Script...
[2020-07-18 09:42:38] -Loading Map Andreas...
[2020-07-18 09:42:38] [MYSQL]: Connection to `saarp` succesful!
[2020-07-18 09:42:38] -Test2 Script loaded correctly...
[2020-07-18 09:42:38] Filterscript 'fac_test2.amx' loaded.
[2020-07-18 09:42:38] [MYSQL]: Receiving response on "cptr_LoadFromDatabase", 3 rows.
[2020-07-18 09:42:38] [MYSQL]: Receiving response on "GetDbPhones", 46 rows.
[2020-07-18 09:42:38] [SERVER]: The time is 9:42
[2020-07-18 09:42:41] [SERVER]: Executing 'serverStart2'.
[2020-07-18 09:42:41] Server password has been removed.
[2020-07-18 09:42:41] Filterscript 'testing.amx' unloaded.
[2020-07-18 09:42:44] [INFECTED]: Infected must be spawned.
[2020-07-18 09:42:44] [npc:join] Infected has joined the server (203:<IP_ADDRESS>)
[2020-07-18 09:42:44] [INFECTED]: Connecting player is infected.
[2020-07-18 09:42:52] [INFECTED]: Infected must be spawned.
[2020-07-18 09:42:52] [npc:join] Infected has joined the server (202:<IP_ADDRESS>)
[2020-07-18 09:42:52] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:01] [SERVER]: The time is 9:43
[2020-07-18 09:43:01] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:01] [npc:join] Infected has joined the server (201:<IP_ADDRESS>)
[2020-07-18 09:43:01] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:09] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:09] [npc:join] Infected has joined the server (200:<IP_ADDRESS>)
[2020-07-18 09:43:09] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:18] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:18] [npc:join] Infected has joined the server (199:<IP_ADDRESS>)
[2020-07-18 09:43:18] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:24] [SERVER]: The time is 9:43
[2020-07-18 09:43:26] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:26] [npc:join] Infected has joined the server (198:<IP_ADDRESS>)
[2020-07-18 09:43:26] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:35] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:35] [npc:join] Infected has joined the server (197:<IP_ADDRESS>)
[2020-07-18 09:43:35] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:43] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:43] [npc:join] Infected has joined the server (196:<IP_ADDRESS>)
[2020-07-18 09:43:43] [INFECTED]: Connecting player is infected.
[2020-07-18 09:43:44] [TESTING]: RefreshDoorMapIcons() called (Initializer, interval=90).
[2020-07-18 09:43:46] [SERVER]: The time is 9:43
[2020-07-18 09:43:51] [INFECTED]: Infected must be spawned.
[2020-07-18 09:43:51] [npc:join] Infected has joined the server (195:<IP_ADDRESS>)
[2020-07-18 09:43:51] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:00] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:00] [npc:join] Infected has joined the server (194:<IP_ADDRESS>)
[2020-07-18 09:44:00] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:08] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:08] [npc:join] Infected has joined the server (193:<IP_ADDRESS>)
[2020-07-18 09:44:08] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:09] [SERVER]: The time is 9:44
[2020-07-18 09:44:17] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:17] [npc:join] Infected has joined the server (192:<IP_ADDRESS>)
[2020-07-18 09:44:17] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:25] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:25] [npc:join] Infected has joined the server (191:<IP_ADDRESS>)
[2020-07-18 09:44:25] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:31] [SERVER]: The time is 9:44
[2020-07-18 09:44:33] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:33] [npc:join] Infected has joined the server (190:<IP_ADDRESS>)
[2020-07-18 09:44:33] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:41] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:41] [npc:join] Infected has joined the server (189:<IP_ADDRESS>)
[2020-07-18 09:44:42] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:50] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:50] [npc:join] Infected has joined the server (188:<IP_ADDRESS>)
[2020-07-18 09:44:50] [INFECTED]: Connecting player is infected.
[2020-07-18 09:44:54] [SERVER]: The time is 9:44
[2020-07-18 09:44:58] [INFECTED]: Infected must be spawned.
[2020-07-18 09:44:58] [npc:join] Infected has joined the server (187:<IP_ADDRESS>)
[2020-07-18 09:44:58] [INFECTED]: Connecting player is infected.
[2020-07-18 09:45:06] [INFECTED]: Infected must be spawned.
[2020-07-18 09:45:06] [npc:join] Infected has joined the server (186:<IP_ADDRESS>)
[2020-07-18 09:45:06] [INFECTED]: Connecting player is infected.
[2020-07-18 09:45:15] [INFECTED]: Infected must be spawned.
[2020-07-18 09:45:15] [npc:join] Infected has joined the server (185:<IP_ADDRESS>)
[2020-07-18 09:45:15] [INFECTED]: Connecting player is infected.
[2020-07-18 09:45:16] [SERVER]: The time is 9:45
[2020-07-18 09:45:23] [INFECTED]: Infected must be spawned.
[2020-07-18 09:45:23] [npc:join] Infected has joined the server (184:<IP_ADDRESS>)
Thank you
rt-2
I fixed this problem by experimenting with the include orders. Here is what I done:
I moved the include from line 26, now to almost the bottom of my includes.
Note that I could not place it before "fixes.inc", otherwise, "fixes.inc" would give me an ALS error.
Thank you,
rt-2
I fixed this problem by experimenting with the include orders. Here is what I done:
I moved the include from line 26, now to almost the bottom of my includes. Note that I could not place it before "fixes.inc", otherwise, "fixes.inc" would give me an ALS error.
Thank you, rt-2
this fix dont work for me, do u try another?
|
2025-04-01T06:39:16.029184
| 2020-04-21T13:43:11
|
604009534
|
{
"authors": [
"aaronbriel",
"kaushaltrivedi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7464",
"repo": "kaushaltrivedi/fast-bert",
"url": "https://github.com/kaushaltrivedi/fast-bert/pull/205"
}
|
gharchive/pull-request
|
added call to convert posix path to string to fix TypeError in save_pโฆ
โฆretrained call. Fixes Issue #200
Merged. Thanks.
|
2025-04-01T06:39:16.031088
| 2024-12-20T13:23:35
|
2752738033
|
{
"authors": [
"mwestphal",
"theusst"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7465",
"repo": "kaust-vislab/MidSurfer",
"url": "https://github.com/kaust-vislab/MidSurfer/issues/2"
}
|
gharchive/issue
|
Ship a binary version of the plugin
Since midsurfer only depends on ParaView, creating (and shipping) a binary version of this plugin should be possible by using the tools developed by Kitware:
https://github.com/Kitware/paraview-ci-example
Let me know what you think
Dear Mathieu,
Thank you for the suggestion! This is definitely on my list. The paper is currently under review, we will decide on how to proceed once we receive the reviews. I will keep this issue open for now.
Thomas
|
2025-04-01T06:39:16.033537
| 2019-05-28T17:34:22
|
449392464
|
{
"authors": [
"CyrilDebon",
"Soviut"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7466",
"repo": "kavalcante/vue-truncate-collapsed",
"url": "https://github.com/kavalcante/vue-truncate-collapsed/pull/14"
}
|
gharchive/pull-request
|
FEATURE property to control truncation state
This allows the truncation to be expanded or collapsed via a property.
Contracted:
<truncate :truncated="true">...</truncate>
Expanded:
<truncate :truncated="false">...</truncate>
Bound:
<truncate :truncated="isExpanded">...</truncate>
data() {
isExpanded: true
}
NOTE: I'm unsure if the property should be named to something truthy by default, like expanded so that true is expanded and false is collapsed
Could someone can merge this ? I need it haha
Or did @Soviut did u found an alternative solution ?
|
2025-04-01T06:39:16.043818
| 2024-03-17T13:28:32
|
2190672499
|
{
"authors": [
"ndepomereu",
"palexdev"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7467",
"repo": "kawansoft/SympleGit-Java",
"url": "https://github.com/kawansoft/SympleGit-Java/issues/1"
}
|
gharchive/issue
|
Some feedback...
Hello, I just wanted to leave some feedback on this library.
Context
I'm developing a tool which needs Git functionalities. In particular, but not limited to, it has to clone a huge amount of repositories from different sources.
I started by using JGit but it's so bad it's almost ridiculous. Very often, clone operations are super slow and would end with a EOF exception. Of course, it does not happen with the native command.
Switching to SympleGit
Let's say that the README is quite catchy:
However, JGit's API comes with a learning curve and lacks direct, one-to-one support for CLI actions. Therefore, SympleGit is likely to be a more straightforward option for simple Git integration in many Java projects, particularly those utilizing basic Git functionalities. Let's delve into the details!
In my honest opinion, it cannot be more false. JGit is super easy to use, at least in my use case, every git command is a class in JGit. In SympleGit the clone command class does not exist.
The great thing about SympleGit is that it uses the native command to perform operations, which is so much faster and performant. The bad thing, again in my opinion, is that the API is not so well thought, and it's lacking.
The idea of creating custom git commands with executeGitCommand(...) surely is good to cover all cases even those that are not implemented yet, but still having them implemented as classes would make everything much easier to use.
It's so confusing to do this:
SympleGit sympleGit = SympleGit.custom()
.setDirectory(repoDirectoryPath)
.build();
Why do I have to give the directory here instead of giving it to the command directly? Like this for JGit:
CloneCommand cmd = new CloneCommand()
.setDirectory(destDir.resolve(path).toFile())
.setURI(url)
.setRemote(remote)
.setBranch(branch);
// In this case, destDir is the base dir in which I want to store all the projects
// path is the where I want to clone the repository, the last part of the path will be the name of the directory
I find it a bit more intuitive
3) I read the README multiple times, but I still don't know how I can track the progress of a command. Probably because, yet again, it's not very intuitive. By nature, working with Process and ProcessBuilder in Java is a cumbersome task, processing the output of a process properly is hard. And for this very reason, a library that uses such APIs should make it as easy as possible for the end user to use it.
The first thing that comes to my mind when I want to track the progress of an external process is something like this:
// I get the why of the check...
if (! gitCommander.isResponseOk()) {
System.out.println("An Error Occured: " + gitCommander.getProcessError());
return;
}
// Then...
while (command.output ...) {
// print output
}
JGit here is doing a far better job here, I can modify the above command like this:
CloneCommand cmd = new CloneCommand()
.setProgressMonitor(new TextProgressMonitor()) // And boom, I automatically have output to the console
.setDirectory(destDir.resolve(path).toFile())
.setURI(url)
.setRemote(remote)
.setBranch(branch);
// Not only that, I can even make custom monitor implementations
Conclusion
I believe SympleGit could become a very good alternative to JGit, but it definitely needs to grow, improve and expand the API.
In the meantime, I think I'll implement a custom solution that uses the native command just like SympleGit because it's simply much much better than JGit which crashes all the time.
Hi,
Sorry for very late reply & thanks for the comments.
I was a bid dubitative, about why it's confusing to do that:
SympleGit sympleGit = SympleGit.custom()
.setDirectory(repoDirectoryPath)
.build();
Could you please elaborate a little? I'm not sure of the importance of the issue.
About the progress monitor, I clearly understand the need, but can you tel when it's important to set up monitor?
Maybe my projects are not big enough, I never had the need...
Of course, we will do it cleanly, extendable like you wish.
Best,
N.
|
2025-04-01T06:39:16.048738
| 2018-05-18T05:07:51
|
324271876
|
{
"authors": [
"kayex",
"peter279k"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7468",
"repo": "kayex/http-codes",
"url": "https://github.com/kayex/http-codes/pull/3"
}
|
gharchive/pull-request
|
Test enhancement
Changed log
Set the multiple PHPUnit version for the different PHP versions.
Add the white filter list in phpunit.xml setting.
Thank you for the pull request! ๐
Do you mind explaining the reason for these changes? Especially the added PHP version constraint, since it breaks compatibility with PHP 5. I know PHP 5 is reaching its end of life very soon, but I don't see a reason to break compatibility unless it's required.
I also think the phpunit version string should be "^6.2 || ^7.0" rather than "^6.2|^7.0", at least according to the Composer docs.
As you say, I set the required PHP version at least 7.0+ because the PHP 5 is EOL.
How about setting the >=5.6 in composer.json require key?
The PHPUnit version is "^6.2 || ^7.0" is the correct defined version description.
|
2025-04-01T06:39:16.060860
| 2016-08-26T22:33:03
|
173556237
|
{
"authors": [
"ShibiHe",
"eggie5"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7470",
"repo": "kayzhu/LSHash",
"url": "https://github.com/kayzhu/LSHash/issues/15"
}
|
gharchive/issue
|
projection type
The code is using np.random.randn() times input vector.
In the LSH paper survey, we are using either (Gaussian Distribution * input + bias)/W or (Uniform Distribution * input). I was wondering if we should change the distribution to uniform in the code?
I have seen that Gaussian Random Projection is on way to implement the random projections. What is this LSH paper survey?
|
2025-04-01T06:39:16.064261
| 2022-11-30T09:23:04
|
1469309694
|
{
"authors": [
"kazurayam"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7471",
"repo": "kazurayam/inspectus",
"url": "https://github.com/kazurayam/inspectus/issues/40"
}
|
gharchive/issue
|
Use materialstore-0.13.0-SNAPSHOT or higher
The materialstore-0.13.0-SNAPSHOT resolved the issue
The pom.xml has no
Now most of the external dependencies for the inspectus project can be resolved by Gradle automatically.
done at 0.6.0
|
2025-04-01T06:39:16.066148
| 2018-04-29T22:51:08
|
318757159
|
{
"authors": [
"alexanderreiff"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7472",
"repo": "kbacha/stellar-sdk",
"url": "https://github.com/kbacha/stellar-sdk/pull/201"
}
|
gharchive/pull-request
|
Add operations for account to CLI
Adds a command to request all operations relating to a given account
to the CLI application.
Fixes #163
Is there a GIF that reflects how this work made you feel?
@correlator @kbacha can you guys take another look at this? Thanks!
@kbacha can you take a look at this refactor? Took me a bit of Rust-y try and error to figure out the indent levelโaware nesting, but in the end, the solution I got was pretty simple.
|
2025-04-01T06:39:16.066985
| 2024-09-04T13:26:21
|
2505377078
|
{
"authors": [
"TomAugspurger",
"minrk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7473",
"repo": "kbatch-dev/helm-chart",
"url": "https://github.com/kbatch-dev/helm-chart/pull/6"
}
|
gharchive/pull-request
|
add extraPodLabels, hub access label by default
ensures network access to the hub with the default hub networkPolicy in the jupyterhub helm chart
Thanks!
|
2025-04-01T06:39:16.117525
| 2022-04-26T08:15:06
|
1215578937
|
{
"authors": [
"MichalMed",
"ahmadjana",
"blcham",
"ledsoft"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7474",
"repo": "kbss-cvut/kbss-website",
"url": "https://github.com/kbss-cvut/kbss-website/pull/13"
}
|
gharchive/pull-request
|
Patch 1
this branch for our open mic sessions:
It should contains the schedule and contents of the seminars
@blcham
Does it make sense to have the open mic schedule publicly visible, when the sessions are internal to KBSS? Or is there a plan to make the sessions public?
Does it make sense to have the open mic schedule publicly visible, when the sessions are internal to KBSS? Or is there a plan to make the sessions public?
Yes, this was the idea, see #7. I would like to publish some of the presentations and possibly invite somebody to join our presentation online.
Good thing is to have some static page with programme and link it every time like three days before.
|
2025-04-01T06:39:16.120976
| 2018-05-09T14:59:34
|
321606485
|
{
"authors": [
"kc9eye"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7475",
"repo": "kc9eye/UData",
"url": "https://github.com/kc9eye/UData/issues/5"
}
|
gharchive/issue
|
Create Installer Form
Create an installer form to collect information on settings that writes the config file and tests the database, creates objects, and initializes the admin. It should also test whether or not it can send mail, and write to the disk for file uploads.
this is not going to happen by me
|
2025-04-01T06:39:16.130758
| 2023-07-03T15:38:34
|
2097506867
|
{
"authors": [
"Peefy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7476",
"repo": "kcl-lang/tree-sitter-kcl",
"url": "https://github.com/kcl-lang/tree-sitter-kcl/issues/6"
}
|
gharchive/issue
|
[Feature] Tree-sitter grammar
Discussed in https://github.com/kcl-lang/kcl/discussions/595
Originally posted by matoous July 3, 2023
Hi team!
Are there any plans to offer Tree-sitter grammer for KCL? I would be happy to contribute but wonder whether the repository should be maintained under the kcl-lang organization.
Hi there @octonawish-akcodes If you are willing, you can work in this repo. Thank you! โค๏ธ
|
2025-04-01T06:39:16.143584
| 2022-10-27T14:45:36
|
1425760800
|
{
"authors": [
"hardys",
"ncdc",
"p0lyn0mial"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7477",
"repo": "kcp-dev/kcp",
"url": "https://github.com/kcp-dev/kcp/issues/2274"
}
|
gharchive/issue
|
bug:
Describe the bug
In e2e-sharded runs we're seeing an error like:
PROXY W1025 18:20:39.100256 31146 reflector.go:324] k8s.io/client-go@v0.0.0-20221025160842-38c73163e766/tools/cache/reflector.go:167: failed to list *v1alpha1.ClusterWorkspace: Unauthorized
PROXY E1025 18:20:39.101457 31146 reflector.go:138] k8s.io/client-go@v0.0.0-20221025160842-38c73163e766/tools/cache/reflector.go:167: Failed to watch *v1alpha1.ClusterWorkspace: failed to list *v1alpha1.ClusterWorkspace: Unauthorized
After debugging locally, it appears to be because we're using the root shard-admin token all shards when creating the clients used by the clusterworkspace informer
We can see here in some local debug log output that the shard-admin token hitting kcp-1 is actually the token from kcp-0:
$ grep "invalid bearer token" kcp-1.log | head -n1
E1027 14:04:04.549596 929593 authentication.go:63] "Unable to authenticate the request SHDEBUG" err="invalid bearer token" req=&{Method:GET URL:/apis/tenancy.kcp.dev/v1alpha1/clusterworkspaces:5678dc3ffb5f4156924efc5c9a4732f12e73fadca3592faa5c762f91fd0fc3f5?limit=500&resourceVersion=0 Proto:HTTP/2.0 ProtoMajor:2 ProtoMinor:0 Header:map[Accept:[application/json, */*] Accept-Encoding:[gzip] Authorization:[Bearer 4ae09225-6a3b-4413-a5a7-d105b45a4466] User-Agent:[kcp-front-proxy/v1.24.3+kcp (linux/amd64) kubernetes/574fe23]] Body:0xc0071db1a0 GetBody:<nil> ContentLength:0 TransferEncoding:[] Close:false Host:<IP_ADDRESS>:6445 Form:map[] PostForm:map[] MultipartForm:<nil> Trailer:map[] RemoteAddr:<IP_ADDRESS>:37122 RequestURI:/clusters/%2A/apis/tenancy.kcp.dev/v1alpha1/clusterworkspaces:5678dc3ffb5f4156924efc5c9a4732f12e73fadca3592faa5c762f91fd0fc3f5?limit=500&resourceVersion=0 TLS:0xc00e689080 Cancel:<nil> Response:<nil> ctx:0xc00fe05d70}
$ grep -B2 4ae09225-6a3b-4413-a5a7-d105b45a4466 ../.kcp-0/admin.kubeconfig
- name: shard-admin
user:
token: 4ae09225-6a3b-4413-a5a7-d105b45a4466
Steps To Reproduce
Check e2e-sharded logs, or run make test-e2e-sharded locally
Expected Behaviour
IIUC we need the proxy informer to be able to list/watch clusterworkspace objects on all shards, so we have to use an authentication method which isn't limited to the root shard.
Some ideas were discussed on slack and it seems like switching to cert based auth may be the best option - this should mean we can authenticate with any shard, and avoid modifying the proxy config when a shard gets added/removed.
Additional Context
No response
/assign
sgtm, thanks for picking it up.
So looking at the history here, I see that the current per-shard client code landed in https://github.com/kcp-dev/kcp/pull/1203 - so it would be helpful to get input from @sttts before I start changing it :)
If we switch this to use client certs, then I think we need to handle the case where each shard has a different cert, and the method to enable that appears to be the mapping file, however in the sharded-test-server case, we seem to only create mappings to the root shard:
$ grep "server:" .kcp-0/admin.kubeconfig
server: https://<IP_ADDRESS>:6444
server: https://<IP_ADDRESS>:6444/clusters/root
$ grep "server:" .kcp-1/admin.kubeconfig
server: https://<IP_ADDRESS>:6445
server: https://<IP_ADDRESS>:6445/clusters/root
$ cat .kcp-front-proxy/mapping.yaml
- backend: https://localhost:6444
backend_server_ca: .kcp/serving-ca.crt
path: /services/
proxy_client_cert: .kcp-front-proxy/requestheader.crt
proxy_client_key: .kcp-front-proxy/requestheader.key
- backend: https://localhost:6444
backend_server_ca: .kcp/serving-ca.crt
path: /clusters/
proxy_client_cert: .kcp-front-proxy/requestheader.crt
proxy_client_key: .kcp-front-proxy/requestheader.key
So I think for multi-shard deployments we need to specify each shard as a backend (using the IP, not localhost), then we can look up the necessary client cert/key using the shard BaseURL, does that sound reasonable?
If we switch this to use client certs, then I think we need to handle the case where each shard has a different cert,
I think that should be okay, I think that the proxy could specify a well-known ServerName during bootstrapping a TLS connection to a shard. That would give us a certificate that would much a well-known CA. Does it make sense?
If we switch this to use client certs, then I think we need to handle the case where each shard has a different cert,
I think that should be okay, I think that the proxy could specify a well-known ServerName during bootstrapping a TLS connection to a shard. That would give us a certificate that would much a well-known CA. Does it make sense?
Thanks for the feedback @p0lyn0mial - I get that the ServerName has to match the CN/SAN in the server cert, but I'm unclear if the proxy has to do anything special here - the user provides the server/client certs, and the ServerName will be derived from the ClusterWorkspaceShard spec in the indexer, so the hostname (or potentially IP for CI/dev testing) just has to match?
I think leveraging the client certs from the mapping file is definitely viable, and probably optimal if we want to ensure the informer uses per-shard credentials.
I can't help wondering if we should just pass an admin kubeconfig to the proxy though which would be potentially much simpler, any thoughts on that? The main disadvantage AFAICS is that we'll need to rename/replace the --root-kubeconfig CLI flag
/cc @ncdc who mentioned he has some work in progress related to this
Summary of some discussion with @ncdc and @p0lyn0mial
@ncdc has some related work in progress which we can possibly revive then use the same client-cert pattern to solve this (convert the informer client to use a proxy->shard client cert instead of the shard-admin token)
We can't easily use an admin kubeconfig to switch to client certs, since in the sharded case the client cert terminates at the proxy, then we use request-header at the shard (so, we can't use the same kubeconfig for the proxy->shard auth)
Accidentally marked as fixed by 2297
/reopen
/assign @p0lyn0mial
This should be resolved via https://github.com/kcp-dev/kcp/pull/2382
|
2025-04-01T06:39:16.163683
| 2024-04-24T19:17:43
|
2262005918
|
{
"authors": [
"souravroy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7478",
"repo": "kdhrubo/db2rest",
"url": "https://github.com/kdhrubo/db2rest/issues/502"
}
|
gharchive/issue
|
MongoDB - Order Tests
Need to mark tests with @Order() to avoid intermittent test failure.
PR https://github.com/kdhrubo/db2rest/pull/503 is merged. Closing the ticket.
|
2025-04-01T06:39:16.227083
| 2018-10-10T08:04:32
|
368540274
|
{
"authors": [
"Actimel",
"tomasfejfar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7479",
"repo": "keboola/php-component",
"url": "https://github.com/keboola/php-component/issues/47"
}
|
gharchive/issue
|
Add support for state file
Comes up from https://github.com/keboola/db-extractor-mysql/pull/81#discussion_r223825923
Would be nice to have getter for state file. Besides there could be also method for writing into outcomes state.
Uลพ tam byla issue https://github.com/keboola/php-component/issues/46 ;)
|
2025-04-01T06:39:16.229555
| 2020-11-25T17:07:51
|
750987183
|
{
"authors": [
"odinuv",
"pivnicek"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7480",
"repo": "keboola/storage-api-php-client-branch-wrapper",
"url": "https://github.com/keboola/storage-api-php-client-branch-wrapper/pull/5"
}
|
gharchive/pull-request
|
Update composer.json for latest sapi client
Required for https://github.com/keboola/input-mapping/pull/68
Pending sapi-client release https://github.com/keboola/storage-api-php-client/releases/tag/untagged-3d86260daac18b393369
@odinuv what should we call this, 0.0.3?
hmm, test failed on "branch not found". is there a possible concurrency issue?
they don't run in parallel, it's quite suspicous
feel free to make it 1.0.0 :) it's already in production
|
2025-04-01T06:39:16.246661
| 2024-08-07T09:41:52
|
2453042770
|
{
"authors": [
"JorTurFer",
"LY-today",
"enaguo",
"guanqinglin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7481",
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/issues/6045"
}
|
gharchive/issue
|
create ScaledObject when triggers type is prometheus metricType is Value err
Report
when triggers type is prometheus๏ผ metricType is Value๏ผcreate ScaledObject ๏ผkeda-operator is err
2024-08-07T09:33:45Z ERROR Reconciler error {"controller": "scaledobject", "controllerGroup": "keda.sh", "controllerKind": "ScaledObject", "scaledObject": {"name":"my-helm-demo","namespace":"crane-system"}, "namespace": "crane-system", "name": "my-helm-demo", "reconcileID": "f750f79f-5a1b-4fc9-88d2-b136a11b65ff", "error": "HorizontalPodAutoscaler.autoscaling \"keda-hpa-my-helm-demo\" is invalid: spec.metrics[0].external.target.type: Invalid value: \"value\": must be either Utilization, Value, or AverageValue"}
apiVersion: keda.sh/v1alpha1
kind: ScaledObject
metadata:
name: my-helm-demo
namespace: crane-system
spec:
scaleTargetRef:
apiVersion: apps/v1
kind: Deployment
name: my-helm-demo
pollingInterval: 15
minReplicaCount: 2
maxReplicaCount: 50
advanced:
horizontalPodAutoscalerConfig:
behavior:
scaleDown:
stabilizationWindowSeconds: 0
policies:
- type: Percent
value: 100
periodSeconds: 15
scaleUp:
stabilizationWindowSeconds: 0
policies:
- type: Percent
value: 500
periodSeconds: 15
triggers:
- type: prometheus
metricType: Value
metadata:
serverAddress: xxx
metricName: portrait_pod_cpu_predict_compression_all_num
query: portrait_pod_cpu_predict_compression_all_num{}
threshold: "50"
Expected Behavior
hpa resources can be created normally
Actual Behavior
hpa not create
Name: my-helm-demo
Namespace: crane-system
Labels: scaledobject.keda.sh/name=my-helm-demo
Annotations: <none>
API Version: keda.sh/v1alpha1
Kind: ScaledObject
Metadata:
Creation Timestamp: 2024-08-07T09:33:24Z
Finalizers:
finalizer.keda.sh
Generation: 1
Resource Version: 24908
UID: 0d11f0d0-216e-4696-be18-51a49cad0bd3
Spec:
Advanced:
Horizontal Pod Autoscaler Config:
Behavior:
Scale Down:
Policies:
Period Seconds: 15
Type: Percent
Value: 100
Stabilization Window Seconds: 0
Scale Up:
Policies:
Period Seconds: 15
Type: Percent
Value: 500
Stabilization Window Seconds: 0
Max Replica Count: 50
Min Replica Count: 2
Polling Interval: 15
Scale Target Ref:
API Version: apps/v1
Kind: Deployment
Name: my-helm-demo
Triggers:
Metadata:
Metric Name: portrait_pod_cpu_predict_compression_all_num
Query: portrait_pod_cpu_predict_compression_all_num{}
Server Address: xxxx
Threshold: 50
Metric Type: Value
Type: prometheus
Status:
Conditions:
Message: Failed to ensure HPA is correctly created for ScaledObject
Reason: ScaledObjectCheckFailed
Status: False
Type: Ready
Message: ScaledObject check failed
Reason: UnkownState
Status: Unknown
Type: Active
Status: Unknown
Type: Fallback
External Metric Names:
s0-prometheus-portrait_pod_cpu_predict_compression_all_num
Original Replica Count: 8
Scale Target GVKR:
Group: apps
Kind: Deployment
Resource: deployments
Version: v1
Scale Target Kind: apps/v1.Deployment
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning ScaledObjectCheckFailed 59s (x17 over 6m27s) keda-operator Failed to ensure HPA is correctly created for ScaledObject
Steps to Reproduce the Problem
Logs from KEDA operator
2024-08-07T09:33:45Z ERROR Reconciler error {"controller": "scaledobject", "controllerGroup": "keda.sh", "controllerKind": "ScaledObject", "scaledObject": {"name":"my-helm-demo","namespace":"crane-system"}, "namespace": "crane-system", "name": "my-helm-demo", "reconcileID": "f750f79f-5a1b-4fc9-88d2-b136a11b65ff", "error": "HorizontalPodAutoscaler.autoscaling \"keda-hpa-my-helm-demo\" is invalid: spec.metrics[0].external.target.type: Invalid value: \"value\": must be either Utilization, Value, or AverageValue"}
KEDA Version
< 2.11.0
Kubernetes Version
< 1.28
Platform
None
Scaler Details
prometheus
Anything else?
$ kubectl version
Client Version: v1.29.2
Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3
Server Version: v1.21.1
WARNING: version difference between client (1.29) and server (1.21) exceeds the supported minor version skew of +/-1
kubectl apply --server-side -f https://github.com/kedacore/keda/releases/download/v2.8.0/keda-2.8.0.yaml
Hello,
KEDA v2.8 is quite old (almost 2 years). I don't remember any issue related but just in case, have you tried using v2.8.2 -> https://github.com/kedacore/keda/releases/tag/v2.8.2?
Sadly, that version is out of support, and we won't ship any new releases for it. I know that recent versions don't face with that issue, I'd suggest upgrading to a recent version.
Hello, KEDA v2.8 is quite old (almost 2 years). I don't remember any issue related but just in case, have you tried using v2.8.2 -> https://github.com/kedacore/keda/releases/tag/v2.8.2?
Sadly, that version is out of support, and we won't ship any new releases for it. I know that recent versions don't face with that issue, I'd suggest upgrading to a recent version.
The problem was indeed solved by switching to v2.8.2, but isnโt this the recommended version officially provided by keda?
https://keda.sh/docs/2.15/operate/cluster/#kubernetes-compatibility
The problem was indeed solved by switching to v2.8.2
Nice! Happy to read it ๐
I guess that you are using old k8s cluster, so probably v2.8 is the greatest version that you can use (if you are running k8s < 1.23 KEDA > 2.9 won't work). The point is that we don't ship fixes for old versions, we could ship a fix for the previous version but 2.8 is 7 versions far from v2.15 and although we will try to support with misconfigurations, if there is a real bug unresolved, we won't ship any fix for old versions.
The problem was indeed solved by switching to v2.8.2
Nice! Happy to read it ๐
I guess that you are using old k8s cluster, so probably v2.8 is the greatest version that you can use (if you are running k8s < 1.23 KEDA > 2.9 won't work). The point is that we don't ship fixes for old versions, we could ship a fix for the previous version but 2.8 is 7 versions far from v2.15 and although we will try to support with misconfigurations, if there is a real bug unresolved, we won't ship any fix for old versions.
@JorTurFer
OK, I understand what you mean. There is another situation for you, that is, our latest version of keda is found to be incompatible on the cluster of version 1.21, and panic will occur. The reason is that the v2 version of hpa cannot be detected.
I also encountered the same problem which caused the function not to work properly๏ผ
Yes, I said, KEDA >= v2.9 requires k8s >= 1.23. This is because k8s introduced a breaking change when they released HPA v2 and removed HPA v1beta1. v2.8 uses v1beta1 and v2.9 uses v2. There isn't any fix or action to do about this, v2.8 uses v1beta1 and v2.9 uses v2.
@JorTurFer
OK, thank you for your answer and wish you a happy life
@JorTurFer Excuse me, when I only configure cron, will the replicas of the workload become minReplicaCount if it is not in the cron period?
@JorTurFer Excuse me, when I only configure cron, will the replicas of the workload become minReplicaCount if it is not in the cron period?
@JorTurFer The test is like this. When I manually modify replicas during non-cron periods, it will eventually become minReplicaCount, which is not very friendly to the business. Based on version 1.8.2, is there any way to allow the business to still have the right to make decisions in this situation?
Excuse me, when I only configure cron, will the replicas of the workload become minReplicaCount if it is not in the cron period?
Yes, outside the cron period and assuming the cron as the only scaler, it'll return 0 so it'll scale to minReplicaCount
@JorTurFer่ฏท้ฎไธไธ๏ผๅฝๆๅช้
็ฝฎcron็ๆถๅ๏ผๅฆๆไธๅจcronๅจๆๅ
๏ผๅทฅไฝ่ด่ฝฝ็replicasไผๅๆminReplicaCountๅ๏ผ
@JorTurFerๆต่ฏๆฏ่ฟๆ ท็๏ผๅฝๆๅจ้cronๆ้ดๆๅจไฟฎๆนreplicasๆถ๏ผๆๅไผๅๆminReplicaCount๏ผๅฏนไธๅกไธๅคชๅๅฅฝ๏ผๅบไบ1.8.2็ๆฌ๏ผๆๆฒกๆๅๆณ่ฎฉไธๅกๅจ่ฟ็งๆ
ๅตไธไพ็ถๆฅๆๅณ็ญๆ๏ผ
I also encountered the same problem๏ผHow to solve it
|
2025-04-01T06:39:16.254467
| 2015-07-16T19:43:41
|
95511708
|
{
"authors": [
"aendrew",
"dustinlarimer",
"josephwegner",
"standaniels",
"timfalls"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7482",
"repo": "keen/dashboards",
"url": "https://github.com/keen/dashboards/issues/79"
}
|
gharchive/issue
|
Bower package is empty
I don't know why, but bower install keen-dashboards results in a license, a README, and nothing else.
Which is a shame, really; you could easily install the necessary CSS and whathaveyou by specifying it as such in the main stanza, which would mean folks like me who use Yeoman could use your layouts far more easily.
Might write a PR.
Thanks for writing this up @aendrew. Admittedly, we haven't spent much time on more involved ways of distributing Keen Dashboards (bower, yeoman, etc.). That would be super cool if you wrote up a PR!
@aendrew just wanted to echo @josephwegner's encouragement toward the PR. We really appreciate your thoughts and any potential contribution!
I don't know anything about bower, but I just came across this SO topic and noticed that there is only a bower.json file in the master branch of this project.
@standaniels Ah! That would make sense why there's nothing when pulling from bower. If the gh-pages branch was merged with master, that would probably resolve this.
Resolved- thanks, everyone!
|
2025-04-01T06:39:16.258400
| 2017-03-23T17:47:06
|
216516094
|
{
"authors": [
"amounierlltech",
"matm"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7483",
"repo": "keeneyetech/ecli",
"url": "https://github.com/keeneyetech/ecli/issues/42"
}
|
gharchive/issue
|
Can't use --config to define login file location
I use release 0.10 for windows 64.
I try to login using --config flag, but always have error 1 in return.
I created a .ecli.json somewhere as defined in login standard output, but json format is not exact in this message. missing " after password, and additionnal coma after url. But fixing format didn't allow me to connect.
I type : ecli login platform --config "D:\myfolder.ecli.json"
What should be --config argument syntax, are windows paths accepted ? is it folder of file path ?
Looks like your JSON config file is malformed maybe. It must look like
{"platform": {"login": "yourlogin", "password": "yourpassword", "url": "https://yoururl/api/v2"}}
Just create it in your local directory and call it by filename:
ecli login platform --config myfile.json
You can remove the dot . at the beginning on the file name and use any name like myfile.json.
It's working with this syntax, thx for your support.
|
2025-04-01T06:39:16.259592
| 2020-12-18T14:04:38
|
770910665
|
{
"authors": [
"r-czajkowski"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7484",
"repo": "keep-network/keep-core",
"url": "https://github.com/keep-network/keep-core/pull/2224"
}
|
gharchive/pull-request
|
Liquidity rewards fetching data
Ref: #2216
This PR adds support for fetching and displaying the rewards data from the LPRewards contracts. All data are stored in the redux store.
@michalsmiarowski I'm ready for the second round!
|
2025-04-01T06:39:16.262899
| 2021-07-26T13:17:51
|
952895281
|
{
"authors": [
"michalinacienciala"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7485",
"repo": "keep-network/npm-version-bump",
"url": "https://github.com/keep-network/npm-version-bump/pull/2"
}
|
gharchive/pull-request
|
Stabilize failing test
The test checking if npm-version-bump action "updates version for
already published environment that don't match initial preid" was
unstable - was refering to actively worked on npm packages versions,
making it necessary to update the action every time new package with
that base version is released on ropsten.
In this commit we change the reference to different version, which is no
longer actively worked on.
@nkuba, I wasn't sure what to pick as a base branch (main or v2) for this PR. I see that main is a couple of commits behind v2 and there's no open PR for merging v2 to main.
Should I keep the base branch as v2 and after the merge of this PR should I tag the v2 branch code with v2 tag?
|
2025-04-01T06:39:16.342303
| 2024-12-11T06:53:18
|
2731986306
|
{
"authors": [
"augustin-v",
"feltroidprime"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7486",
"repo": "keep-starknet-strange/garaga",
"url": "https://github.com/keep-starknet-strange/garaga/issues/270"
}
|
gharchive/issue
|
feat: Centralize RiscZero constants definition by creating Rust bindings for py/ts.
Currently, the same constants (CONTROL ROOT & CONTROL_ID) are defined in multiple locations.
In python : https://github.com/keep-starknet-strange/garaga/blob/073843663fd253ad33a697a6aae285704df45b21/hydra/garaga/starknet/groth16_contract_generator/parsing_utils.py#L20-L24
In rust : https://github.com/keep-starknet-strange/garaga/blob/073843663fd253ad33a697a6aae285704df45b21/tools/garaga_rs/src/calldata/full_proof_with_hints/groth16.rs#L297-L311
In typescript : https://github.com/keep-starknet-strange/garaga/blob/073843663fd253ad33a697a6aae285704df45b21/tools/npm/garaga_ts/src/node/starknet/groth16ContractGenerator/parsingUtils.ts#L7-L9
Keep them in rust only and create a python / wasm binding so that updating them is easier.
Note : the tools/garaga_rs package has already multiple examples of python / wasm bindings.
Hello, I would love to work on this. I'm really excited about Garaga with the latest Noir compatibility update and would be honored to be a part of it.
I have just completed implementing a L1 gas prices sampling oracle to Katana in Rust so I believe to be able to tackle this issue. I'll make sure to communicate if I encounter any problems.
@feltroidprime Thank you, on it!
|
2025-04-01T06:39:16.344986
| 2023-06-13T16:02:41
|
1755228327
|
{
"authors": [
"MdTeach",
"abdelhamidbakhta",
"greged93"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7487",
"repo": "keep-starknet-strange/madara",
"url": "https://github.com/keep-starknet-strange/madara/issues/626"
}
|
gharchive/issue
|
dev: improve testing of to_invoke_tx
Improve the testing of the rpc-core/src/utils.rs file by adding unit tests for to_invoke_tx. The tests are expected to cover edge cases in order to be accepted.
@greged93 I would love to take this!
@greged93 I would love to take this!
Assigned to you
@MdTeach need to wait on #597 to be merged
Hey @MdTeach, #597 was merged and moved the logic to the file primitives/starknet/src/transation/types.rs and the implementation changed to impl TryFrom<BroadcastedInvokeTransaction> for InvokeTransaction. Would be great if you are still able to provide tests for this.
sure @greged93
|
2025-04-01T06:39:16.347869
| 2023-06-21T16:23:39
|
1767970229
|
{
"authors": [
"lambda-0x"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7488",
"repo": "keep-starknet-strange/madara",
"url": "https://github.com/keep-starknet-strange/madara/pull/721"
}
|
gharchive/pull-request
|
dev: use Felt252Wrapper instead of [u8; 32]
Pull Request type
Please add the labels corresponding to the type of changes your PR introduces:
Refactoring (no functional changes, no API changes)
What is the current behavior?
Resolves: #716
Does this introduce a breaking change?
Not sure if this counts as breaking change.
Other information
(OT: we should add clippy to husky hooks as well)
|
2025-04-01T06:39:16.351213
| 2024-09-11T10:48:40
|
2519390257
|
{
"authors": [
"0xChqrles",
"ikemHood",
"manlikeHB",
"mubarak23",
"ugur-eren"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7489",
"repo": "keep-starknet-strange/zkramp",
"url": "https://github.com/keep-starknet-strange/zkramp/issues/65"
}
|
gharchive/issue
|
[CONTRACT] setup liquidity addition
Complete this issue according to this article: How to manage liquidity in a P2P ramping app
Please add integration tests
@0xChqrles i
Hi @0xChqrles, an i work on this?
Hi @0xChqrles Can i work on this issue if it's available?
Hey @0xChqrles I would like to work on this so I can work on #67 which is a dependant.
@0xChqrles, can I work on this?
@0xChqrles am available to work on this ,
With previous experience working on this codebase,
I have read the article
I can implement this with 3 days
ETH: 72hrs
@0xChqrles am available to work on this
|
2025-04-01T06:39:16.394902
| 2019-05-21T01:50:27
|
446380147
|
{
"authors": [
"keiohta"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7490",
"repo": "keiohta/tf2rl",
"url": "https://github.com/keiohta/tf2rl/issues/21"
}
|
gharchive/issue
|
Reproduce DQN paper results for Atari
Reproduce following paper results (at least 1 game, because @keiohta does not have enough computation resource)
Human-level control through deep reinforcement learning
Preprocessing
Working directly with raw Atari 2600 frames, which are 210 x 160 pixel images with a 128-colour palette, can be demanding in terms of computation and memory requirements.
We apply a basic preprocessing step aimed at reducing the input dimensionality and dealing with some artefacts of the Atari 2600 emulator.
First, to encode a single frame we take the maximum value for each pixel colour value over the frame being encoded and the previous frame. This was necessary to remove flickering that is present in games where some objects appear only in even frames while other objects appear only in odd frames, an artefact caused by the limited number of sprites Atari 2600 can display at once.
Second, we then extract the Y channel, also known as luminance, from the RGB frame and rescale it to
84 x 84. The function w from algorithm 1 described below applies this preprocessing to the m most recent frames and stacks them to produce the input to the Q-function, in which m = 4, although the algorithm isrobust to different values of m (for example, 3 or 5).
|
2025-04-01T06:39:16.425009
| 2023-07-05T17:48:18
|
1790015887
|
{
"authors": [
"keltia"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7491",
"repo": "keltia/fetiche-rs",
"url": "https://github.com/keltia/fetiche-rs/issues/20"
}
|
gharchive/issue
|
Statistics/metrics ought to be gathered at Engine level.
Right now, sources::access::opensky has some stats gathering code. This should be expanded into an Engine-wide system.
tracing might be a good crate to use.
More likely as a separate Actor within fetiched.
|
2025-04-01T06:39:16.433226
| 2018-05-13T12:23:23
|
322589437
|
{
"authors": [
"kemokemo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7492",
"repo": "kemokemo/gckdir",
"url": "https://github.com/kemokemo/gckdir/issues/18"
}
|
gharchive/issue
|
failed to create the hash value of large files
https://github.com/kemokemo/gckdir/blob/1fb158b19ff1160d2a23244216b076f4c52e3ae1/lib/hash.go#L96-L103
The ioutil.ReadFile() is easy to use, but this read all of the files in memory.
If you have any large files (ex. several gigabytes), the function to create hash value will be very slowly.
In some cases, gckdir application crashes. ๐ข
Let's fix it!
The official godoc of the "crypto/sha256" package is most valuable.
https://godoc.org/crypto/sha256#ex-New--File
|
2025-04-01T06:39:16.440307
| 2024-10-01T19:24:17
|
2560016431
|
{
"authors": [
"CXWorks",
"kenba"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7493",
"repo": "kenba/opencl3",
"url": "https://github.com/kenba/opencl3/issues/69"
}
|
gharchive/issue
|
Unsound issue in SvmVec
Hi, thanks for your time to read this issue. Our static analysis tool found there might be an unsound issue in your set_len implementation of buffer:
https://github.com/kenba/opencl3/blob/1d7d74da2ff76c93c7418dd016386a8b907e2bd3/src/svm.rs#L300-L311
As mentioned in the comments, this may introduce uninitilaized memory and reading it is considered as undefined behavior in Rust. As a reference, in std library, all the set_len method are marked as unsafe:
https://github.com/rust-lang/rust/blob/63a0bdd5622eaf6b9524702f055bb4525acfc9f2/library/alloc/src/vec/mod.rs#L1849-L1853
Thanks again for your time.
Change incorporated in version 0.10.0.
|
2025-04-01T06:39:16.487044
| 2016-02-08T23:11:45
|
132283591
|
{
"authors": [
"kennethreitz",
"victor1969"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7494",
"repo": "kennethreitz/requests",
"url": "https://github.com/kennethreitz/requests/issues/3000"
}
|
gharchive/issue
|
Problem with Content-Lenght is missing
I'm playing clonk rage but i can't play online because this happend
Content-Lenght is missing pls help me
PS: i don't speak english so well
@victor1969 can you please share the code you are using so we can understand what you're asking?
where i get the code?
@victor1969 is this an error you saw while playing a game? Why did you decide to post about it here?
i'm just trying to play online Clonk rage but says on the server list
Internet server on league.clonkspot.org
Invalid server response: Content-Length is missing!
@victor1969 unfortunately, that has nothing to do with project. There's probably somewhere else online for you to report this, though!
:(
|
2025-04-01T06:39:16.490377
| 2016-04-21T15:21:26
|
150106470
|
{
"authors": [
"Lukasa",
"kevinburke",
"sigmavirus24"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7495",
"repo": "kennethreitz/requests",
"url": "https://github.com/kennethreitz/requests/pull/3108"
}
|
gharchive/pull-request
|
Flip conditional in session.send()
Previously we checked that the request being sent was an instance of a
PreparedRequest. If a user somehow created a PreparedRequest using a different
Requests library instance, this check makes the request un-sendable.
(This happened recently - unbeknownst to me, my server was running an outdated
version of pip, vulnerable to this issue - pypa/pip#1489, which creates
multiple subdirectories (src/requests, src/requests/requests) when you rerun
pip install --target. So the PreparedRequest was being created in one version
of the library and compared against the other version of the library, and
throwing this exception, even though they were both PreparedRequest instances!)
It would probably be preferable to check the object's behavior (instead of
its type), but a PreparedRequest has a lot of behavior, and it wouldn't be
really feasible or allow us to provide a helpful error message to check all
of it here. Instead flip the conditional to guard against the user sending an
unprepared Request, which should still give us most of the benefits of the
better error message.
Fixes #3102
I'm happy with this! Go for it @kennethreitz, merge if you'd like to. =D
Not if I merge it first ;)
|
2025-04-01T06:39:16.502134
| 2017-10-24T00:23:57
|
267858266
|
{
"authors": [
"coveralls",
"obi1kenobi"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7496",
"repo": "kensho-technologies/graphql-compiler",
"url": "https://github.com/kensho-technologies/graphql-compiler/pull/49"
}
|
gharchive/pull-request
|
Update filter handler functions to use FilterOperationInfo objects.
Since we now have a lot of different filter operators, it has become clear that the old filter function signature can use some streamlining. This PR adds a new FilterOperationInfo object that holds all non-contextual data that is relevant to a filtering operation, such as the directive object itself, or the type and field name being filtered on.
This new abstraction:
simplifies the function signature of filtering functions to just 3 arguments, while getting rid of all unused variables; and
simplifies the reasoning for which filters are valid in which situations.
Coverage decreased (-0.5%) to 92.553% when pulling f50fc3a51842cc330bf735048237f23bfecf0b8a on filter_operation_info into da6e03c424abc17cc242341674823ea8c21c060d on edge_degree_filtering.
|
2025-04-01T06:39:16.511739
| 2015-06-19T13:25:02
|
89564228
|
{
"authors": [
"jdart",
"kentcdodds"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7497",
"repo": "kentcdodds/kcd-angular",
"url": "https://github.com/kentcdodds/kcd-angular/pull/3"
}
|
gharchive/pull-request
|
bower was removing all the good stuff
constants, directives, filters and services were not available after bower installing.
I've merged it, but to be perfectly honest, I don't recommend you use this module. I will not be supporting it. I recommend you copy and paste and modify the code.
|
2025-04-01T06:39:16.516208
| 2016-09-07T10:55:58
|
175473095
|
{
"authors": [
"craigcosmo",
"leggomuhgreggo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7498",
"repo": "kenwheeler/slick",
"url": "https://github.com/kenwheeler/slick/issues/2508"
}
|
gharchive/issue
|
how to center item and some
this is how I want it to be displayed
this is how I got so far
What I want to do is show one center image, and half of the right and left images,
Problem: In my solution, I was able to show one center image, but the right and left image only show little, I want them to show more
Here is my code:
http://jsfiddle.net/fp1kcah6/
Check out centerPadding
|
2025-04-01T06:39:16.561042
| 2020-11-25T11:15:19
|
750728349
|
{
"authors": [
"ankitjain28may",
"bacherfl",
"codecov-io",
"johannes-b"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7499",
"repo": "keptn/keptn",
"url": "https://github.com/keptn/keptn/pull/2733"
}
|
gharchive/pull-request
|
Keptn Smart Auth: Autofetch keptn endpoint and api-token
Signed-off-by: ankitjain28may<EMAIL_ADDRESS>
[ ] Can you please add the message "CLI is not authenticated" in case of ClusterIP / NotePort. (to inform the user about the current status)
Codecov Report
Merging #2733 (18ba47f) into master (9bd9c79) will decrease coverage by 13.64%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #2733 +/- ##
===========================================
- Coverage 36.26% 22.61% -13.65%
===========================================
Files 161 50 -111
Lines 9288 1110 -8178
Branches 197 197
===========================================
- Hits 3368 251 -3117
+ Misses 5381 832 -4549
+ Partials 539 27 -512
Flag
Coverage ฮ
moduleA
22.61% <รธ> (-0.27%)
:arrow_down:
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage ฮ
bridge/client/app/_services/data.service.ts
0.98% <0.00%> (-0.12%)
:arrow_down:
...lient/app/project-board/project-board.component.ts
2.83% <0.00%> (-0.03%)
:arrow_down:
bridge/client/app/_models/service.ts
0.00% <0.00%> (รธ)
helm-service/controller/action_handler.go
platform-support/openshift-route-service/main.go
api/ws/hub.go
...ce/pkg/handler/approval_triggered_event_handler.go
helm-service/pkg/helm/helm_helper.go
cli/cmd/update.go
.../pkg/configurationchanger/configuration_changer.go
... and 100 more
Continue to review full report at Codecov.
Legend - Click here to learn more
ฮ = absolute <relative> (impact), รธ = not affected, ? = missing data
Powered by Codecov. Last update ba0b2ae...18ba47f. Read the comment docs.
@bacherfl The unit tests are failing because it needs this PR to be merged - https://github.com/keptn/kubernetes-utils/pull/13
@bacherfl The unit tests are failing because it needs this PR to be merged - keptn/kubernetes-utils#13
Ok, got it. Just had one comment about the other PR - once that has been adressed we can merge that one
@bacherfl The unit tests are failing because it needs this PR to be merged - keptn/kubernetes-utils#13
Ok, got it. Just had one comment about the other PR - once that has been adressed we can merge that one
@ankitjain28may I have just merged the kubernetes-utils PR. Can you please update the kubernetes-utils dependency to the latest status from the master branch? When that is done we can merge this PR as well
@bacherfl Have updated the dependency to latest master, Please review :)
|
2025-04-01T06:39:16.569686
| 2021-11-20T14:03:57
|
1059152027
|
{
"authors": [
"fchollet",
"fmerizzi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7500",
"repo": "keras-team/keras-io",
"url": "https://github.com/keras-team/keras-io/pull/716"
}
|
gharchive/pull-request
|
fixing issue in GCN example
The small issue prevented the code from running in Colab environment, because of inconsistencies in the dimension of the output layer.
Thanks for the PR. This is fixed by #717 717
|
2025-04-01T06:39:16.589182
| 2023-10-06T19:27:58
|
1930825996
|
{
"authors": [
"fchollet",
"qlzh727",
"sampathweb"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7501",
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/issues/18567"
}
|
gharchive/issue
|
TensorFlow GPU - Fix keras/layers/merging/merging_test.py
Fix failing test - keras/layers/merging/merging_test.py::MergingLayersTest::test_sparse_dot_2d Fatal Python error: Aborted and update TODO in https://github.com/keras-team/keras/blob/master/keras/kokoro/github/ubuntu/gpu/build.sh#L39
https://source.cloud.google.com/results/invocations/9df9ee7e-5666-4644-abd2-01a10771faeb/targets/keras%2Fgithub%2Fubuntu%2Fgpu%2Ftensorflow%2Fpresubmit/log
keras/layers/merging/merging_test.py::MergingLayersTest::test_sparse_dot_2d Fatal Python error: Aborted
Current thread 0x00007f51610f0740 (most recent call first):
File "/tmpfs/venv/lib/python3.9/site-packages/tensorflow/python/ops/linalg/sparse/gen_sparse_csr_matrix_ops.py", line 1114 in sparse_matrix_sparse_mat_mul
File "/tmpfs/src/github/keras/keras/backend/tensorflow/numpy.py", line 119 in sparse_sparse_matmul
File "/tmpfs/src/github/keras/keras/backend/tensorflow/numpy.py", line 156 in matmul
File "/tmpfs/src/github/keras/keras/ops/numpy.py", line 3431 in matmul
File "/tmpfs/src/github/keras/keras/layers/merging/dot.py", line 171 in batch_dot
File "/tmpfs/src/github/keras/keras/layers/merging/dot.py", line 320 in _merge_function
File "/tmpfs/src/github/keras/keras/layers/merging/base_merge.py", line 189 in call
File "/tmpfs/src/github/keras/keras/ops/operation.py", line 47 in __call__
File "/tmpfs/src/github/keras/keras/utils/traceback_utils.py", line 114 in error_handler
File "/tmpfs/src/github/keras/keras/layers/layer.py", line 810 in __call__
File "/tmpfs/src/github/keras/keras/utils/traceback_utils.py", line 114 in error_handler
File "/tmpfs/src/github/keras/keras/testing/test_case.py", line 380 in run_layer_test
File "/tmpfs/src/github/keras/keras/layers/merging/merging_test.py", line 240 in test_sparse
File "/tmpfs/venv/lib/python3.9/site-packages/absl/testing/parameterized.py", line 319 in bound_param_test
File "/usr/lib/python3.9/unittest/case.py", line 550 in _callTestMethod
File "/usr/lib/python3.9/unittest/case.py", line 592 in run
File "/usr/lib/python3.9/unittest/case.py", line 651 in __call__
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/unittest.py", line 333 in runtest
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 262 in <lambda>
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 341 in from_call
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 261 in call_runtest_hook
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 222 in call_and_report
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 133 in runtestprotocol
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/runner.py", line 114 in pytest_runtest_protocol
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 350 in pytest_runtestloop
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 325 in _main
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 271 in wrap_session
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/main.py", line 318 in pytest_cmdline_main
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_callers.py", line 77 in _multicall
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_manager.py", line 115 in _hookexec
File "/tmpfs/venv/lib/python3.9/site-packages/pluggy/_hooks.py", line 493 in __call__
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/config/__init__.py", line 169 in main
File "/tmpfs/venv/lib/python3.9/site-packages/_pytest/config/__init__.py", line 192 in console_main
File "/tmpfs/venv/bin/pytest", line 8 in <module>
github/keras/keras/kokoro/github/ubuntu/gpu/build.sh: line 34: 4954 Aborted (core dumped)
The culprit is between 2.15.0.dev20230918 (good) and 2.15.0.dev20230919 (bad).
Tested via: pytest keras/layers/merging/merging_test.py::MergingLayersTest::test_basic_add
Culprit in one of the changes in this range: git log e4a6720f42a..dfcf1d40e46 --oneline
Thanks Ramesh for the repo, we will revisit this during the triage meeting.
On a side note, I didn't find any change on the sparse side between those two date. Will need to dig deep for the root cause.
Here's a small code snippet to reproduce the issue in Colab with Keras Master and TF-Nightly -
!pip uninstall -y keras tensorflow
!pip install tf-nightly[and-cuda]==2.15.0.dev20231009 --extra-index-url https://pypi.nvidia.com
!pip uninstall -y keras-nightly
# Install Keras from Master via `python pip_build.py --install`
input = keras.layers.Input(shape=(2,))
x1 = keras.layers.Dense(4, activation='relu')(input)
x2 = keras.layers.Dense(4, activation='relu')(input)
added = keras.layers.Add()([x1, x2])
out = keras.layers.Dense(1)(added)
model = keras.models.Model(inputs=input, outputs=out)
x = np.random.randn(8, 2)
y = np.random.randn(8, 1)
model.compile(optimizer='sgd', loss='mse')
model.fit(x, y, epochs=1)
If I replace Add with Concatenate also it breaks. This is a high priority error since it breaks very important layer in TF GPU. Also, the same test fails for JAX GPU as well.
@fchollet - If you have any thoughts or suggestions to try let me know.
The example you provided doesn't even use sparse inputs, which is different from the error on top. The error DNN library initialization failed somehow indicate that its a setup issue for GPU.
The example you provided doesn't even use sparse inputs, which is different from the error on top. The error DNN library initialization failed somehow indicate that its a setup issue for GPU.
There are multiple failures in merging_test.py. I tried to run the basic test case with add and that fails. Initially I reported on the sparse test which actually aborts with core dump
keras/layers/merging/merging_test.py::MergingLayersTest::test_sparse_dot_2d Fatal Python error: Aborted
TF Nightly 09/18 works for ALL the tests in merging_test.py. So I think its a common issue due to change in TF on 09/19 between these commits in TF: git log e4a6720f42a..dfcf1d40e46 --oneline
Somehow I wasn't able to produce the on colab with T4 GPU. https://colab.sandbox.google.com/drive/1_hMJieL_6DobTPUbZ6BRZIEVz0YRHhBo#scrollTo=GM2B7qEqNYqk
Maybe I didn't config the GPU properly?
@sampathweb do u have a testable env that I can run with?
Also seems to be failing with JAX-GPU now:
github/keras/keras/kokoro/github/ubuntu/gpu/build.sh: line 57: 4493 Aborted (core dumped) pytest keras --ignore keras/applications --ignore keras/layers/merging/merging_test.py --ignore keras/trainers/data_adapters/py_dataset_adapter_test.py --ignore keras/backend/jax/distribution_lib_test.py --cov=keras
I wil work on this tomorrow. I used Colab v100 as my test env
Seems to be a Cudnn TF compilation issue.
2023-10-17 20:23:09.628643: I external/local_xla/xla/service/service.cc:176] StreamExecutor device (0): Tesla V100-SXM2-16GB, Compute Capability 7.0
2023-10-17 20:23:10.277194: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:447] Loaded runtime CuDNN library: 8.7.0 but source was compiled with: 8.9.4. CuDNN library needs to have matching major version and equal or higher minor version. If using a binary install, upgrade your CuDNN library. If building from sources, make sure the library loaded at runtime is compatible with the version specified during compile configuration.
2023-10-17 20:23:10.278786: W tensorflow/core/framework/op_kernel.cc:1839] OP_REQUIRES failed at xla_ops.cc:574 : FAILED_PRECONDITION: DNN library initialization failed. Look at the errors above for more details.
Tested via pytest keras/layers/merging/merging_test.py::MergingLayersTest::test_basic_add
Don't have a resolution yet. But might be related to this change that's within the range - git log e4a6720f42a..dfcf1d40e46 --oneline
commit 3de44168950a5972ba4cfa7e3c6cbf4cffa67fe6
Author: A. Unique TensorFlower<EMAIL_ADDRESS>Date: Mon Sep 18 13:50:11 2023 -0700
Upgrade to LLVM 17, CUDA 12.2, and CuDNN 8.9.4
This is updating TF's default toolchain to LLVM 17, as well as
CUDA and cuDNN to the latest releases.
PiperOrigin-RevId: 566403707
|
2025-04-01T06:39:16.607783
| 2017-10-18T08:58:28
|
266412679
|
{
"authors": [
"lfoppiano",
"stzellerhoff"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7502",
"repo": "kermitt2/grobid",
"url": "https://github.com/kermitt2/grobid/issues/249"
}
|
gharchive/issue
|
Build error duplicate entry
Installing version 0.4.4 fails reporting this error:
Failed to execute goal org.apache.maven.plugins:maven-shade-plugin:3.0.0:shade (default) on project grobid-core: Error creating shaded jar: duplicate entry: META-INF/services/shadedgrobid.org.apache.lucene.codecs.PostingsFormat
How can I avoid duplicate entries?
Stephan
hi @stzellerhoff could you paste here the whole log?
Thanks
Luca
|
2025-04-01T06:39:16.652767
| 2024-06-26T03:17:52
|
2374096242
|
{
"authors": [
"aarontrowbridge",
"andgoldschmidt"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7503",
"repo": "kestrelquantum/QuantumCollocation.jl",
"url": "https://github.com/kestrelquantum/QuantumCollocation.jl/issues/132"
}
|
gharchive/issue
|
[Bug]: namespace includes common variable names
What happened?
In _integrators.jl,
export state
export controls
export timestep
export comps
export dim
Version
stable release
What does this bug affect?
[ ] quantum system construction
[ ] problem setup
[ ] problem solution
[ ] problem performance
[ ] solution analysis
[ ] plotting
[ ] documentation
[ ] tests
[ ] other (please specify below)
Other information
No response
these should all be prefixed with get_ probably
closed by #149
|
2025-04-01T06:39:16.681900
| 2024-02-07T01:46:04
|
2121997626
|
{
"authors": [
"kevincobain2000",
"vunhatchuong"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7504",
"repo": "kevincobain2000/gobrew",
"url": "https://github.com/kevincobain2000/gobrew/issues/174"
}
|
gharchive/issue
|
Bug: Windows 10 Symlink required privilege
Problem
Windows 10 symlink required privilege:
Looks like it hasn't been resolved: https://github.com/golang/go/issues/22874
Solution
You can either require privilege or use cmd through os.exec.Command:
if runtime.GOOS == "windows" {
if err = exec.Command("cmd", "mklink", "/d", link , destination).Run(); err == nil {
return nil
}
}
Thanks for reporting.
Pull req too please?
Ok I'll try
|
2025-04-01T06:39:16.688436
| 2022-03-20T08:43:10
|
1174476354
|
{
"authors": [
"b0o",
"kevinhwang91"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7505",
"repo": "kevinhwang91/nvim-bqf",
"url": "https://github.com/kevinhwang91/nvim-bqf/issues/60"
}
|
gharchive/issue
|
Support previewing vim-fugitive :Gclog entries
Feature description
Currently, when the quickfix window is populated by vim-fugitive's :Gclog command, vim-bqf's preview window is blank unless the highlighted entry has been manually opened first. I'm guessing it has to do with the special file URIs fugitive uses, which look like fugitive:///home/maddy/.config/nvim/.git//f0e7c59ec59a25bfe928b555dd8387242d2b810f.
Describe the solution you'd like
It would be nice if nvim-bqf could properly preview these vim-fugitive quickfix items.
Additional context
If the highlighted entry hasn't been manually opened yet:
Once the highlighted entry has been opened:
nvim-bqf will never support this feature because of the performance issue. You can use should_preview_cb hack what you want. You can also use the below setting to experience how slow it is.
require('bqf').setup({
should_preview_cb = function(bufnr)
local bufname = vim.api.nvim_buf_get_name(bufnr)
if bufname:match('^fugitive://') then
vim.api.nvim_buf_call(bufnr, function()
cmd(('noa do fugitive BufReadCmd %s'):format(bufname))
end)
end
return true
end
})
For me, I use https://github.com/rbong/vim-flog which is a git log wrapper and has the ability to limit the commits. More importantly, it is an extension based on fugitive.
Cool, thank you for the tips, and for the wonderful plugin!
I took your hack and hacked it even further - this loads the fugitive buffer asynchronously:
require('bqf').setup {
preview = {
should_preview_cb = function(bufnr)
local bufname = vim.api.nvim_buf_get_name(bufnr)
if bufname:match '^fugitive://' then
local pvs = require 'bqf.preview.session'
local extmark = require 'bqf.preview.extmark'
local ts = require 'bqf.preview.treesitter'
vim.schedule(function()
vim.api.nvim_buf_call(bufnr, function()
local fbufnr = pvs.float_bufnr()
vim.cmd(('do fugitive BufReadCmd %s'):format(bufname))
pvs.floatbuf_reset()
ts.disable_active(fbufnr)
extmark.clear_highlight(fbufnr)
vim.cmd(('silent lua require"bqf.utils".transfer_buf(%d, %d)'):format(bufnr, fbufnr))
end)
end)
end
return true
end,
},
}
I copied some code from bqf.preview.handler.open() to do this, I'm sure there's a better way but this works well enough for me.
Please update the code which is convenient to hack.
local bqf_pv_timer
require('bqf').setup {
preview = {
should_preview_cb = function(bufnr, qwinid)
local bufname = vim.api.nvim_buf_get_name(bufnr)
if bufname:match '^fugitive://' and not vim.api.nvim_buf_is_loaded(bufnr) then
if bqf_pv_timer and bqf_pv_timer:get_due_in() > 0 then
bqf_pv_timer:stop()
bqf_pv_timer = nil
end
bqf_pv_timer = vim.defer_fn(function()
vim.api.nvim_buf_call(bufnr, function()
vim.cmd(('do fugitive BufReadCmd %s'):format(bufname))
end)
require('bqf.preview.handler').open(qwinid, nil, true)
end, 60)
end
return true
end
}
}
Awesome, much cleaner. Thank you!
|
2025-04-01T06:39:16.696267
| 2023-07-13T14:15:19
|
1803128402
|
{
"authors": [
"kevinscaria",
"noshaq"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7506",
"repo": "kevinscaria/InstructABSA",
"url": "https://github.com/kevinscaria/InstructABSA/issues/16"
}
|
gharchive/issue
|
Missing create_data_in_joint_task_format method at DatasetLoader
Hi. Thanks for sharing the code of your papers.
I'm trying to reproduce the training for the joint task, but this error appears:
AttributeError: 'DatasetLoader' object has no attribute 'create_data_in_joint_task_format'
There is any way to solve it with the other methods in the class?
Thanks!
Hi,
There has been some API changes. I have not updated the run_model.py script yet. But for joint task, it should be:
create_data_in_aspe_format()
From here on joint task will be called as ASPE (Aspect Sentiment Pair Extraction), since new tasks are being added.
Great! Thanks for your answer.
|
2025-04-01T06:39:16.780386
| 2017-06-07T12:21:49
|
234190595
|
{
"authors": [
"bharathvaj1995",
"kevva"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7531",
"repo": "kevva/brightness",
"url": "https://github.com/kevva/brightness/issues/13"
}
|
gharchive/issue
|
(node:8987) UnhandledPromiseRejectionWarning: Unhandled promise rejectio n (rejection id: 1): Error: No backlight device found
This is the error that occurs when executing on linux platform.
Please update the module to the latest node.js with catch()
Error: No backlight device found
Using the CLI or the regular API? Anyway, you can see why it fails in the error message.
Please update the module to the latest node.js with catch()
You should catch errors yourself.
Try running ls /sys/class/backlight. Does it list anything?
Nothing. But i installed backlight using apt-get install xbacklight
What distro are you running? And I assume you can change brightness otherwise?
|
2025-04-01T06:39:16.783143
| 2016-11-02T04:58:14
|
186712651
|
{
"authors": [
"shinnn",
"tunnckoCore"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7532",
"repo": "kevva/decompress",
"url": "https://github.com/kevva/decompress/issues/44"
}
|
gharchive/issue
|
Decouple builtin plugins?
Currently i just uses decompress-unzip, but it needs some more workarounds it because of the dest. I don't need the others. Yea they are pretty small deps, but... would be better. :)
FYI you can directly use a plugin without decompress. Example: https://github.com/shinnn/get-zonetab/blob/ddf68e876dcc93cc891d3c289c45f97ea215d3ad/index.js#L15
Yea, that's what I'm doing now, but i should handle files manually.
utils.unzip()(buf).then((files) => {
// handle / decompress files
})
maxogden/extract-zip seems to fit your use case.
|
2025-04-01T06:39:16.784607
| 2017-03-07T19:52:17
|
212540055
|
{
"authors": [
"maxtaco",
"thibaultdej"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7533",
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/6160"
}
|
gharchive/issue
|
Can't log into the Keybase app
Keybase GUI Version: 1.0.18-20170223121400+af0e640
I tried to log in through the shell after resetting my keys and it returned 'ERROR SecretSyncer: no device found for ID=660e***'. I can log in in the browser, but not on the desktop.
Thanks for the report, this should be fixed in our next release.
|
2025-04-01T06:39:16.821384
| 2017-09-21T16:46:35
|
259562025
|
{
"authors": [
"AngryAnt",
"Avamander",
"Hexstream",
"IpsmLorem",
"ItsExtra",
"Mikaela",
"QuentinBellus",
"TheKangaroo",
"bradmurray",
"cbluth",
"cjb",
"dabura667",
"dlumma",
"domq",
"fortran77",
"getraf",
"grilldos",
"heronhaye",
"hexagile",
"jcausey-astate",
"jfinstrom",
"junderw",
"mahmad-nutrien",
"maxtaco",
"mvdan",
"neunenak",
"rsoltanzadeh",
"srd424",
"supakeen",
"tstrijdhorst",
"w3bb",
"wolfmcnally"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7534",
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/8586"
}
|
gharchive/issue
|
Please Support Multiple Identities or Personas
I have two personas online, my professional persona and my personal/family personas are kept separate online.
In this real-life example, I will never associate my facebook with my github via way of keybase (i wont proof facebook and github on the same keybase account, because that then associates them).
Are my concerns common, is anyone else concerned about the same thing?
Can/Will keybase support multiple personas or identities?
We do support this. You can keybase logout of one account and then log in to the second.
We don't plan to make it much easier than it already is, because most people don't do this.
I also would love the ability to be logged in to multiple accounts simultaneously.
I understand Keybase is โsingle identityโ driven, but I think that idea is misguided.
The equivalent in GnuPG would be to say โERROR: โgenkeys: you can only generate one private key ring per instance of gnupg, please log out of current OS user and log into a new user to manage a second private keypairโ
You are just creating an artificial hurdle that wonโt prevent people from doing it, and will only hurt the people who try jumping over that hurdle and trip and fall and break their face.
People who want to do this will do it, so why not make it simpler for them?
If the answer is โitโs not a priorityโ I totally understand and can put up with the inconvenience of switching between identities.
But if the answer is โmultiple email addresses are bad. multiple personas are bad, everyone should know about all your emails and PGP keys and social accountsโ then I think that is misguided.
The whole โdiscourage multiple accountsโ thing makes sense for Facebook and twitter etc. but not for keybase imo.
The logout/login process is not very user friendly because if people want to run multiple identities they usually want to be able to be logged into all of them at the same time which would mean the clients should be able to handle it. Running multiple instances on win/osx/linux would work, but multiple instances on a phone are problematic.
@bradmurray I agree. Logging in/out is not representative of real-life, for example, any one of my acquaintances should be able to reach me at any time, despite whether they are professional acquaintances or otherwise. Just because i'm "at work" doesnt mean i shouldnt receive notifications from anyone else.
Excellent software but honestly I'm surprised this wasn't a must-have in version 1. Part of managing one's privacy these days is managing one's identities. I know many people who would balk at using Keybase because they can't easily manage multiple identities.
We already support this.
On Wed, Feb 14, 2018 at 1:32 AM Wolf McNally<EMAIL_ADDRESS>wrote:
Excellent software but honestly I'm surprised this wasn't a must-have in
version 1. Part of managing one's privacy these days is managing one's
identities. I know many people who would balk at using Keybase because
they can't easily manage multiple identities.
โ
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/keybase/client/issues/8586#issuecomment-365509735,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AA05_5nmKbxpy6Qo4gSBEbrGYXtP5js0ks5tUn35gaJpZM4Pfm9d
.
@maxtaco I just checked the iPhone app and logging in to multiple accounts simultaneously (the subject of this issue) is not yet supported.
I need to log out and log back in with a different account. During which time I can no longer get notifications for the previous account.
We already support this.
How is this already supported? I run keybase on Android, Windows and OSX and on all of them I have to logout and log back in to change identities.
I'm a new keybase user using it for multiple identities and I too would like to see the UX around being logged into multiple identities simultaneously become better.
This is a sorely missing feature, imo.
I'll add my vote for this one as well, and also point out that the Firefox container solution doesn't help with chat -- chat seems to be only supported with the native app, with the exception of the limited chat enabled by the browser extension. So, the browser can't really fix this issue.
I have a similar use case to others here; I have a need to keep my work-related communication separate from personal accounts. I'm happy to see the new "exploding messages" feature; I hope this one might get some consideration now that that one is done. (Maybe also chat search via the GUI, but that's a different issue.)
Hey, I just started to use keybase and the first thing I checked was if I can create separate accounts for private and business use cases. It would be great to have these two logged in in the app simultaneously.
Jumping on this bandwagon as well. A lot of services nowadays do support multiple identities and/or personas logged in at the same time. It'd be awesome if keybase could do the same for those of us who keep separate online identities for separate areas of work.
N+1'd. An important use case is the ability to make "bot" accounts to be members in a team (e.g. to auto-push DevOps secrets into a Keybase Git repo)
+1 on this. Multiple identities should be supported.
We don't plan to make it much easier than it already is, because most people don't do this.
Many people who already have accounts in different places do have one for work and one for their personal identity. I know many, MANY people in this case, and looking at the list of people in this thread, there are indeed many of us.
Please bring this to Keybase
+1 I also have personal (where I am generally opinionated on things that matter to me) and my company account (fun but filtered)
Conflicted if the "already supported" responses ought to be treated as depressing or hilarious.
Dropping in to +1 here. Keybase looks fascinating but I have had my fingers burnt in the past by letting different areas of my life overlap (new rule: colleagues only get added on social media when one or other of us leaves!) Not doing that again!
Want this as well.
The logout/login process is not very user friendly because if people want to run multiple identities they usually want to be able to be logged into all of them at the same time which would mean the clients should be able to handle it. Running multiple instances on win/osx/linux would work, but multiple instances on a phone are problematic.
If it is not feasible on phone, then don't try to implement this for phone. Do it for the desktop app. It is (obviously) a largely overlooked feature.
+1000
For me, this is a dealbreaker - as much as I love keybase, until multiple personas will be properly supported, it will be just a toy app I'm checking out every now and then.
+1 This is the main reason I'm not using keybase tbh.
It's quite understandable that software developers want to discourage or prevent account creation spam. It can clog up the namespace, for one thing. It allows Sybil attacks in dicsussions, for another, so a person can pretend to be multiple people all arguing the same side (happens on Reddit and Twitter all the time). Making it too easy for a person to have many logins will also make it too easy for the person to take up many times the free allocation of cloud storage per person.
So I definitely will not argue for multiple identities without limit.
However, two is a bare minimum, because many of us find it desirable to keep our personal and professional lives separate. If carrying a single mobile device, you would want to have both identities logged in. Or you might have separate business and personal devices, and in that case, you can much more easily implement security procedures, so certain business-related files are never allowed to touch a personal device.
You might need a third identity to say and do naughty things, not naughty enough to make you a criminal in the civilized world, but naughty enough to damage your personal or professional life if found out and, in some less civilized parts of the world, be stoned to death.
You can probably imagine reasons for additional identities.
I think about five or six would be a reasonable limit.
Definitely a must have !
I just joined a workplace that uses Keybase and would love to have the ability to simultaneously sign into my work and personal accounts.
I'd understand if it wasn't a priority though.
We're close on this, either in this release or the next.
In https://github.com/keybase/client/issues/2104#issuecomment-493079134 it's said that you can currently keybase login -s anotheraccount to switch account
Is there any UI support planned for this?
yes.
however I currently can't get desktop notifications for any of the other logged in accounts. I have to switch and visually check for unread chats.
Please add support for this!
I honestly can't believe someone, never mind someone responsible for a product like keybase, would honestly think "most people don't have multiple identities on the internet"
There's a subculture of people who seem to have melted their selves into the internet like butter on toast. People who do this will naturally interact with lots of others who have done the same to a greater or lesser extent so they come to believe it is normal. If a primary way you interact with other humans is online, then you might eventually arrive at the conclusion that the people posting prolifically online are representative of all humans. Because how else would you ever comes across anyone? It's a weird self perpetuating bubble.
That said I am not sure if the devs/bosses here think this (though I haven't read all the other related issues yet so maybe they do) or if they are merely indifferent or uninterested.
IMHO security and privacy/anonymity are intrinsically linked in every practical way.
Like the others, I would also consider this tool much more useful if I didn't have to sit and think which portion of my life it is most suited for, which is my train of thought at the moment.
Feature: is in current release version.
Feature Request Issue: continues discussion as if it is not there.
Feature:
Feature: is in current release version.
I am new to this whole github thing but why is the issue still marked as "open"? Also #2104 "Feature Request: Logging into multiple keybase accounts locally". In my limited experience, mostly lurking, devs/admins are extremely happy to close these things once they are resolved.
Found this thread (and the many others which link here) trying to determine if it is possible or no. All I see is many people stating it's not possible. I can't see anything obvious on my phone about how to do this. "Add identities" lets me verify on about 800 mastodon servers.
Because these threads get long, and after a while it's no longer clear what feature they were talking about to begin with.
Also, Keybase client has 2.6k issues open... which tells you that this project is not focused much on closing / managing issues and is more focused on coding / reviewing code.
The closed issues mostly tend to be bugs that were fixed, since they are much easier to put a binary value on completion.
I think this issue is still open and people are still talking about this because currently, there is no Graphical User Interface (GUI) support for multiple identities or personas through the Keybase app.
Here's my +1 for GUI support in addition to the Command Line Interface (CLI) implementation.
@junderw
Feature: is in current release version.
The feature definitely doesn't exist in current version in an acceptable manner.
+1. Especially with chat there's times where I'd want to use alternative identities. If Keybase is aiming for businesses to support Keybase, then it would be incredibly useful to have different accounts for work and private use.
Note that these days twitter allows to simultaneously connect to up to 5 accounts/identities at the same time, right there through the normal official GUI. Just click on "More" then the blue "โ" to add more accounts. (I never use that feature, but I'm glad it exists.)
Also note that these days, gmail (or only G Suite?) supports multiple accounts natively:
Kebase technically supports it, but having to re-enter your password
every time is a massive pain.
On Wed., Nov. 6, 2019, 7:19 a.m. Jean-Philippe Paradis, <
<EMAIL_ADDRESS>wrote:
Also note that these days, gmail (or only G Suite?) supports multiple
accounts natively:
[image: Multiple gmail accounts]
https://camo.githubusercontent.com/5646d7d9503d7d5513b0623695dfd33968a03675/68747470733a2f2f7777772e68657873747265616d2e6e65742f746d702f676d61696c2d6d756c7469706c652d6163636f756e74732e706e67
โ
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/keybase/client/issues/8586?email_source=notifications&email_token=AIEGR7W5YENP5SGDKQTYRS3QSKY5DA5CNFSM4D36N5O2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEDGK3GY#issuecomment-550284699,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AIEGR7U5DEVHW4FWYV6FJFLQSKY5DANCNFSM4D36N5OQ
.
Being able to be logged in to multiple accounts simultaneously is distinctly different.
I think this issue is still open and people are still talking about this because currently, there is no Graphical User Interface (GUI) support for multiple identities or personas through the Keybase app.
Here's my +1 for GUI support in addition to the Command Line Interface (CLI) implementation.
We just got our wish.
Multiple identities or personas support through the Keybase app Graphical User Interface (GUI).
That's in addition to the Command Line Interface (CLI) implementation.
Yes!
Sounds great, but as soon as I select "Log in as another user" it immediately logs me out and I have to enter my password again to log back in...
This is now in the app in version 5.0 on all platforms. You should be able to switch users without entering your password every time.
@Hexstream, if you have "Always stay logged in" checked in the Advanced Settings tab, that sounds like a bug. Please send feedback after reproducing and make another issue. Thanks.
You are right, checking "Always stay logged in" solves the problem, but this behavior is not intuitive at all. I created a new issue.
Thanks so much for fixing this issue! I use this feature on a daily basis, and it works great.
I have two suggestions to make it even more awesome:
https://github.com/keybase/client/issues/23850 to show notifications for all accounts configured as "always stay logged in".
https://github.com/keybase/client/issues/23851 to allow switching between "always stay logged in" accounts with just a keyboard shortcut, or a single click, and remember what tab was last open for each account.
|
2025-04-01T06:39:16.823953
| 2018-03-30T17:49:36
|
310127146
|
{
"authors": [
"chrisnojima"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7535",
"repo": "keybase/client",
"url": "https://github.com/keybase/client/pull/11157"
}
|
gharchive/pull-request
|
fix jump to new conversation of kbfs chats
@keybase/react-hackers this fixes the jump to new convo flow
I renamed the payload param and flow didn't catch it because the payload types weren't exact. i don't remember why we undid that (it used to be) but it really should go back i think
|
2025-04-01T06:39:16.826098
| 2019-12-09T22:17:54
|
535371730
|
{
"authors": [
"buoyad",
"maxtaco"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7536",
"repo": "keybase/client",
"url": "https://github.com/keybase/client/pull/21496"
}
|
gharchive/pull-request
|
fix loading team on create
Messy because team get takes a name
Spawn a getDetails on team created because the component will not have the teamname
Store entry teamNameToID on team created so we have it as soon as possible
cc @keybase/y2ksquad
We can easily expose an RPC that takes an ID (if we donโt have one already)
@maxtaco changing the gui to use the team ID here is part of y2k-974
|
2025-04-01T06:39:16.831689
| 2016-10-12T07:47:11
|
182457338
|
{
"authors": [
"oconnor663",
"sijad"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7537",
"repo": "keybase/kbfs",
"url": "https://github.com/keybase/kbfs/pull/419"
}
|
gharchive/pull-request
|
Update kbfstool URL
it has been changed in https://github.com/keybase/kbfs/commit/3cf9045546540bb72435bb48cd53471773a46c76 but URL didn't updated.
can be related to #309
Thanks!
|
2025-04-01T06:39:16.836158
| 2016-10-28T03:44:06
|
185832218
|
{
"authors": [
"strib"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7538",
"repo": "keybase/kbfs",
"url": "https://github.com/keybase/kbfs/pull/473"
}
|
gharchive/pull-request
|
cr: speed up while journaling by avoiding extraneous block uploads and downloads
This PR enables a journaled TLF to use only local information during conflict resolution, and completely avoids creating a conflict branch on the servers. Getting there required quite a few changes (let me know if you'd prefer them broken out into separate PRs):
tlfJournal pauses flushing as soon as it notices a conflict.
conflictResolver figures out which blocks don't need to be uploaded, and passes that to MDOps.ResolveBranch so the journal can ignore them.
We now write the resolution MD to the journal, rather than directly to the server (due to the ResolveBranch semantics).
Since the journal might have uploaded some blocks before discovering a conflict, BlockServer now has an IsFlushed method that the conflictResolver can use to tell which blocks have already been flushed.
blockJournal now has a way to avoid removing the data for flushed blocks from disk, until the next successful MD flush. This helps us avoid re-downloading blocks in the case where the resolution MD itself hits a conflict during flushing, and we have to re-do CR all over again.
(This is possible now because we are writing the resolution MD to the journal, and won't know if there's a conflict until later.)
conflictResolver needed several tweaks to work correctly when the unmerged branch contains a resolutionOp, which wasn't possible before.
RootMetadata.deepCopy needs to also copy the cached block changes -- otherwise when the journal converts a branch and replaces the cached MD entries, it could wind up forgetting the cached block change pointer.
This fixes a bug in master that we just didn't have test coverage for, but now we do.
Ok fixed up, ready for another look please.
Known unrelated CI failures. Merging.
|
2025-04-01T06:39:16.838675
| 2017-04-16T05:12:02
|
221991159
|
{
"authors": [
"akalin-keybase",
"coveralls"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7539",
"repo": "keybase/saltpack",
"url": "https://github.com/keybase/saltpack/pull/28"
}
|
gharchive/pull-request
|
Switch on major versions when reading
Otherwise, we may panic when reading a header
from an unknown minor version.
Move some non-common functions out to better places.
Coverage increased (+0.04%) to 86.244% when pulling 72453ddb0f9d49ab06adb214e119187bb313b94d on akalin/fix-switch into 5ff37ca6c69573e0786615153f260a45b266476b on master.
Hmm. I should probably write a test for this. I'll prep one tomorrow.
|
2025-04-01T06:39:16.900387
| 2023-02-23T11:52:19
|
1596719277
|
{
"authors": [
"THS-on",
"codecov-commenter",
"maugustosilva"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7540",
"repo": "keylime/keylime",
"url": "https://github.com/keylime/keylime/pull/1323"
}
|
gharchive/pull-request
|
roadmap: update for 2023
This updates the roadmap with the major changes planned for Keylime in 2023.
Please have a look and see, if the goals are defined well enough and if I forgot something to add.
Codecov Report
Merging #1323 (cc93e8a) into master (8ddfd12) will decrease coverage by 0.49%.
The diff coverage is n/a.
:mega: This organization is not using Codecovโs GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more
Additional details and impacted files
Flag
Coverage ฮ
packit-e2e
74.15% <รธ> (-0.49%)
:arrow_down:
testsuite
11.07% <รธ> (รธ)
unittests
1.92% <รธ> (รธ)
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage ฮ
keylime/keylime_agent.py
72.55% <0.00%> (-0.58%)
:arrow_down:
Just to make clear the key realization on the parenthetical comment on the 4th bullet on @THS-on answer. It seems clear to me there is the need for separating the "ascertaining the integrity" of the attestation artifacts (i.e., tpm2_checkquote, the replaying of both MB and IMA log and matching of PCRs) and the subsequent validation against policy.
|
2025-04-01T06:39:16.949687
| 2016-05-24T15:17:22
|
156535078
|
{
"authors": [
"morenoh149",
"pll33"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7541",
"repo": "keystonejs/generator-keystone",
"url": "https://github.com/keystonejs/generator-keystone/issues/200"
}
|
gharchive/issue
|
Installation failed with "yo keystone"
Expected behavior
I was following the "Get Started" page, "Commands". I wanted to install the latest version of keystone and startup new project.
Actual behavior
After the keystone-generator installed, tried to do "yo keystone" and the installation has started but it does not generate the project. I got the following error:
screen shot 2016-05-24 at 10 21 49
Thanks for helping me out!
Issue was fixed with #195, the Gruntfile is out of date on the 0.3.16 with npm install
@mxstbr
|
2025-04-01T06:39:16.956905
| 2021-06-25T09:20:40
|
929998322
|
{
"authors": [
"Jytesh",
"Kikobeats"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7542",
"repo": "keyvhq/keyv",
"url": "https://github.com/keyvhq/keyv/pull/14"
}
|
gharchive/pull-request
|
(WIP) build: restructure project
Changes
[x] Setup Dependabot
[x] Use standard instead of xo
[x] Unify Licenses
[x] Setup npm metadata
[x] Setup editorconfig
Should homepage be changed to whatever site we deploy docs to?
|
2025-04-01T06:39:16.963016
| 2024-07-17T10:47:51
|
2413277227
|
{
"authors": [
"chrdebru",
"dachafra",
"frmichel"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7544",
"repo": "kg-construct/rml-cc",
"url": "https://github.com/kg-construct/rml-cc/pull/49"
}
|
gharchive/pull-request
|
Included RML Strategy, fixed some mistakes, and updated sources to respect RML-IO
@anaigmo: I have removed "join" from the comment in the ontology to make rml:Strategy and rml:strategy implicitly specific to RML-CC.
@frmichel Can you check?
Hi @chrdebru, sorry I've been away from the last discussions. I'm just confused about commit https://github.com/kg-construct/rml-cc/pull/49/commits/29216b27393ed1b1ab2a83e8fd94cd09e1b394a1, I though the plan was to move rml:Strategy to core, but it is reintegrated into CC here. Why is that?
@frmichel the explanation is here https://github.com/kg-construct/rml-core/issues/51 (last comment from Ben)
Ok got it, thx @dachafra, sorry for missing that.
no worries! :-)
Just one tiny remark, the rendered.html mentions 10 January 2024 as the spec date. This is on purpose or should be updated?
Apart from that I've good, you can merge the PR whenever you like.
|
2025-04-01T06:39:16.972887
| 2023-01-18T12:50:53
|
2101722153
|
{
"authors": [
"bjdmeest",
"dachafra",
"elsdvlee",
"pmaria"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7545",
"repo": "kg-construct/rml-core",
"url": "https://github.com/kg-construct/rml-core/issues/74"
}
|
gharchive/issue
|
Join specification when logical source is the same
Let's say we have two triple maps that refer to the same logical source (and with same, we really mean same URI, not "same because the descriptions lead to the semantically same logical source").
Sample source (CSV)
id,parent_id
1,2
2,1
Base mapping (YARRRML)
prefixes
ex: http://example.com#
sources:
test: [data.csv]
mappings:
test1:
s: ex:$(id)
po:
p: ex:parent
o:
mapping: test2
test2:
s: ex:$(parent_id)
We have following use cases that are underspecified in de spec
the spec currently says If the logical source of the child triples map and the logical source of the parent triples map of a referencing object map are not identical, then the referencing object map must have at least one join condition.
If a join condition is specified AND the logical source is not the same: common case, execute join condition between each iteration pair
If a join condition is specified AND the logical source is the same: same as above
If no join condition is specified AND the logical source is not the same: do a full join (i.e., take all iterations into account)
example output: ex:1 ex:parent ex:2, ex:1 ex:parent ex:1, ex:2 ex:parent ex:2, ex:2 ex:parent ex:1
If no join condition is specified AND the logical source is the same: don't do a full join, but take the current iteration into account
example output: ex:1 ex:parent ex:2, ex:2 ex:parent ex:1
this last one is the edge case, but allows to 'join per iteration'. Question is: should we make this edge case explicit, or should there be a different way to tackle this edge case?
To continue the discussion of this issue, and considering that there is already a spec written, I would suggest making a PR @elsdvlee so the rest can review it and provide comments!
See https://github.com/kg-construct/rml-core/pull/78
awesome! Please assign @andimou @pmaria @bjdmeest @DylanVanAssche as potential reviewers
My view on defining equality of logical sources:
Object equality in programming languages is used as the basis for many things. For example comparison in different data structures for uniqueness and hashing. (Think dictionaries, sets etc.)
I strongly believe we should be able to leverage this for logical sources. I think source and logical source equality is something that is very useful to have when building RML processors.
Therefor, I would propose to come up with a definition of equality which can be implemented as such.
My proposal would be to define a logical source or source to be equal to another logical source or source if the RML-defined properties of the description of both are equal.
RML-defined: Those properties that are defined by a specification to have behavior that influences the behavior of an RML processor.
These properties MUST be listed for the rml:LogicalSource specification.
These properties MUST be listed for any rml:Source description.
Doing so will allow RML processors to map these descriptions to standard object equality mechanisms in their respective programming languages to best leverage the language's abilities.
|
2025-04-01T06:39:17.016576
| 2017-05-19T12:56:03
|
229966649
|
{
"authors": [
"khorolets"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7548",
"repo": "kiasaki/docker-alpine-postgres",
"url": "https://github.com/kiasaki/docker-alpine-postgres/issues/33"
}
|
gharchive/issue
|
Fatal error with latest update
Hi! Thanks for your image, I'm happy using it!
Unfortunately, with the latest update, I'm not able to run postgres in docker-compose.yml
The error is
FATAL: could not create lock file "/run/postgresql/.s.PGSQL.5432.lock": No such file or directory
Your image part from docker-compose.yml
postgres:
restart: always
image: kiasaki/alpine-postgres:latest
environment:
- POSTGRES_PASSWORD=test
- POSTGRES_DB=test_db
- POSTGRES_USER=test_user
volumes:
- ./persistant/pgdata:/var/lib/postgresql/data
I've switched to 9.5 tag and everything works fine.
Sorry looks like the same as #32
You can close the issue if you decide that it's a duplicate. Thank you!
|
2025-04-01T06:39:17.018193
| 2016-11-15T11:15:14
|
189364456
|
{
"authors": [
"Scirrio",
"Yukun99",
"kicjow"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7549",
"repo": "kicjow/Crazy-Enchantments",
"url": "https://github.com/kicjow/Crazy-Enchantments/issues/110"
}
|
gharchive/issue
|
Telepathy does not work with Fortune enchantment
The telepathy enchantment disables fortune from allowing ore blocks to yield more than 1 of their resource, whereas the fortune enchantment without telepathy would allow ores to yield 2-3.
Is there any possible fix to the telepathy enchantment to allow more than 1 resource to be sent to the inventory when fortune is being used on the same tool?
Ill be fixing it in the next update.
#Badbones69totherescue
This has been fixed and will be in the next update.
|
2025-04-01T06:39:17.027636
| 2019-02-08T20:24:19
|
408310364
|
{
"authors": [
"dannyalright",
"eoji"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7550",
"repo": "kickstarter/android-oss",
"url": "https://github.com/kickstarter/android-oss/pull/471"
}
|
gharchive/pull-request
|
Logged in drawer updates
What โ
making avatar 48 by 48
making name bigger
Story ๐
Trello
before and after ๐
killing it! One tweak: could you please increase the distance from the top of the screen to the header avatar?
Between top-of-header-avatar and system bar: 16dp (or 18 if you want to use our grid)
Between top-of-header-avatar and top-of-screen: 40dp (or 42 if you want to use our grid)
That space is 16dp, my avi has a white background so it's hard to see
Improve navigation and filtering on Android app
|
2025-04-01T06:39:17.035892
| 2021-06-16T17:20:33
|
922880477
|
{
"authors": [
"msadoon",
"nativeksr"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7551",
"repo": "kickstarter/ios-oss",
"url": "https://github.com/kickstarter/ios-oss/pull/1521"
}
|
gharchive/pull-request
|
[Ignore] CI Failure Validation
Just drafting PR to verify if a CI failure is related to changes in Cartfile or general environment issue.
SwiftFormat found issues:
File
Rules
Kickstarter-iOS/Library/OptimizelyResultTypeTests.swift:13:1
warning: (spaceInsideComments) Add leading and/or trailing space inside comments.
Generated by :no_entry_sign: Danger
SwiftFormat found issues:
File
Rules
Kickstarter-iOS/AppDelegate.swift:10:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:11:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:12:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:13:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:14:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:15:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:16:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:17:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:18:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:19:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:20:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:21:1
warning: (sortedImports) Sort import statements alphabetically.
Kickstarter-iOS/AppDelegate.swift:22:1
warning: (sortedImports) Sort import statements alphabetically.
Generated by :no_entry_sign: Danger
No longer needed to check ci validation
|
2025-04-01T06:39:17.038781
| 2018-05-30T19:09:57
|
327861802
|
{
"authors": [
"Scollaco",
"ifbarrera"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7552",
"repo": "kickstarter/ios-oss",
"url": "https://github.com/kickstarter/ios-oss/pull/368"
}
|
gharchive/pull-request
|
Password reveal functionality
What
Show/hide password on login. ๐
See ๐
Native: Password reveal option to make entering passwords easier
|
2025-04-01T06:39:17.130269
| 2022-01-21T06:20:06
|
1110099930
|
{
"authors": [
"cristianonicolai",
"kie-ci",
"nmirasch"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7553",
"repo": "kiegroup/kogito-runtimes",
"url": "https://github.com/kiegroup/kogito-runtimes/pull/1916"
}
|
gharchive/pull-request
|
KOGITO-6606 KOGITO-6446 Produce event listener before starting Data Iโฆ
โฆndex
The (build) kogito-runtimes check has failed. Please check the logs.
The (build) kogito-apps check has failed. Please check the logs.
The (build) kogito-examples check has failed. Please check the logs.
The (build) kogito-apps check has failed. Please check the logs.
Jenkins rerun kogito-examples tests
Jenkins rerun kogito-apps tests
The (build) kogito-examples check has failed. Please check the logs.
Jenkins rerun kogito-examples tests
|
2025-04-01T06:39:17.284993
| 2023-04-03T15:23:17
|
1652287705
|
{
"authors": [
"baldimir",
"danielezonca",
"gitgabrio"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7554",
"repo": "kiegroup/kogito-runtimes",
"url": "https://github.com/kiegroup/kogito-runtimes/pull/2919"
}
|
gharchive/pull-request
|
[kie-issues-156] Use specific conversion method for numbers
Fixes https://github.com/kiegroup/kie-issues/issues/156.
When the convert method is executed on IBM JDK, the order of iteration of methods is different. For Integer, the original implementation picked a wrong method (Integer.getInteger()), which caused a wrong behaviour and test fail.
This PR uses valueOf() method specifically for numbers.
Adds more tests for various number types.
How to replicate CI configuration locally?
Build Chain tool does "simple" maven build(s), the builds are just Maven commands, but because the repositories relates and depends on each other and any change in API or class method could affect several of those repositories there is a need to use build-chain tool to handle cross repository builds and be sure that we always use latest version of the code for each repository.
build-chain tool is a build tool which can be used on command line locally or in Github Actions workflow(s), in case you need to change multiple repositories and send multiple dependent pull requests related with a change you can easily reproduce the same build by executing it on Github hosted environment or locally in your development environment. See local execution details to get more information about it.
How to retest this PR or trigger a specific build:
for pull request checks
Please add comment: Jenkins retest this
for a specific pull request check
Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] tests
for quarkus branch checks
Run checks against Quarkus current used branch
Please add comment: Jenkins run quarkus-branch
for a quarkus branch specific check
Run checks against Quarkus current used branch
Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-branch
for quarkus main checks
Run checks against Quarkus main branch
Please add comment: Jenkins run quarkus-main
for a specific quarkus main check
Run checks against Quarkus main branch
Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-main
for quarkus lts checks
Run checks against Quarkus lts branch
Please add comment: Jenkins run quarkus-lts
for a specific quarkus lts check
Run checks against Quarkus lts branch
Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-lts
for native checks
Run native checks
Please add comment: Jenkins run native
for a specific native check
Run native checks
Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native
for native lts checks
Run native checks against quarkus lts branch
Please add comment: Jenkins run native-lts
for a specific native lts check
Run native checks against quarkus lts branch
Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native-lts
Full Kogito testing (with cloud images and operator BDD testing)
Please add comment: Jenkins run BDD
This check should be used only if a big change is done as it takes time to run, need resources and one full BDD tests check can be done at a time ...
How to backport a pull request to a different branch?
In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch).
NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another.
Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added.
If something goes wrong, the author will be notified and at this point a manual backporting is needed.
NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created.
@kiegroup/gatekeepers
Could you please merge this, squashing commits? Thanks!
I see @baldimir asked the review to @fjtirado too.
@baldimir
Do you want to wait or is it fine for you to proceed?
hi @danielezonca Let's wait for @fjtirado
@danielezonca ^^
|
2025-04-01T06:39:17.438942
| 2024-02-16T06:29:35
|
2137959334
|
{
"authors": [
"kieranabrennan",
"zanovis"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7555",
"repo": "kieranabrennan/every-breath-you-take",
"url": "https://github.com/kieranabrennan/every-breath-you-take/issues/21"
}
|
gharchive/issue
|
ECG Data
Nice job on this app! I wanted to ask if you have plans to incorporate the ECG characteristics in the future, as I see the UUIDs are called out in the .py file. Thanks for any info!
@zanovis you'll see now I'm connecting to the sensors with a python package I've written that also has support for ECG
https://github.com/kieranabrennan/blehrm/blob/master/blehrm/clients/polar_h10.py
|
2025-04-01T06:39:17.448821
| 2024-11-03T06:20:25
|
2631004400
|
{
"authors": [
"Orenji-Tangerine"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7556",
"repo": "kijai/ComfyUI-Florence2",
"url": "https://github.com/kijai/ComfyUI-Florence2/issues/92"
}
|
gharchive/issue
|
Slower execution when compared with Pytorch 2.1.2 + CU 11.8 + xformers 0.0.23.post1
I just updated my environment to 2.5.0 + CU 12.4 + xformers 0.0.28.post2, and it seems that it is slower and if I have a few images to loads, it will stop for a while before processing next image (at Florence2Run node). In the past, processing 10 images took about 40 secs now it needs 100 secs. Are there anything I can do to speed up?
Issue gone after updating to Pytorch 2.5.1 + CU 12.4 (xformers 0.0.28.post3)
|
2025-04-01T06:39:17.451037
| 2016-04-25T17:47:29
|
150935779
|
{
"authors": [
"cgurnik",
"mmailhot"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7557",
"repo": "kikinteractive/kik-python",
"url": "https://github.com/kikinteractive/kik-python/pull/16"
}
|
gharchive/pull-request
|
Fix verify_signature in Python 3
We were not properly accounting for the fact that the hashing functions in Python 3 take bytestrings.
Note that in Python 3, the body of the HTML request must be a bytestring when passed into the verify_signature function, this is consistent with how Flask and other HTTP libraries work.
Fixes #14
@mmailhot :+1:
|
2025-04-01T06:39:17.466525
| 2023-03-09T13:15:57
|
1617231083
|
{
"authors": [
"kiliman",
"zhuhaow"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7558",
"repo": "kiliman/remix-flat-routes",
"url": "https://github.com/kiliman/remix-flat-routes/issues/43"
}
|
gharchive/issue
|
Deep flat-file convention causing wrong route added
If we add something like this
- a+
- b+
- index.tsx
This would generate route for a+/b+ and a/b at the same time.
Odd. The + are removed and treated like you used .
a+/b+/index.tsx should map to a.b.index.tsx
Can you show me the remix.config?
The content of remix.config
const { flatRoutes } = require("remix-flat-routes");
/** @type {import('@remix-run/dev').AppConfig} */
module.exports = {
devServerBroadcastDelay: 1000,
ignoredRouteFiles: ["**/.*"],
server: "./server.js",
serverBuildPath: "functions/[[path]].js",
serverConditions: ["worker"],
serverDependenciesToBundle: "all",
serverMainFields: ["browser", "module", "main"],
serverMinify: true,
serverModuleFormat: "esm",
serverPlatform: "neutral",
future: {
unstable_tailwind: true,
unstable_postcss: true,
v2_routeConvention: true,
},
routes: (defineRoutes) => {
return flatRoutes("routes", defineRoutes);
},
};
My actual routes looks like this
<Route path="admin+/users+" file="routes/admin+/users+/index.tsx" />
<Route path="admin" file="routes/admin+/_layout.tsx">
<Route path="users/edit/:id?" file="routes/admin+/users+/edit.($id).tsx" />
<Route path="users/" index file="routes/admin+/users+/index.tsx" />
<Route path="users/new" file="routes/admin+/users+/new.tsx" />
</Route>
Ah, you can't use v2_routeConvention and remix-flat-routes at the same time.
v2 doesn't support the + convention.
I see. I thought the routes function would take over the route generation.
Thanks for the explanation.
Change that fix the issue
|
2025-04-01T06:39:17.471053
| 2023-09-02T10:52:24
|
1878608041
|
{
"authors": [
"Dmitryqq",
"killkimno"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7559",
"repo": "killkimno/MORT",
"url": "https://github.com/killkimno/MORT/issues/15"
}
|
gharchive/issue
|
Index was out of range error
Hello
Getting this error every time I open the program, clicking the "Apply" button or finishing quick setting. As the result, settings are not applying. Pic 2 is coming after the first one (it's saying application is shutting down) but nothing happens.
By the way russian translate is bad and breaking the interface, how can I change it to english?
hello
It seems like the app settings are messed up.
Do the following to reset and try again.
Additional -> Reset config
change language
Additional -> Advanced settings -> App language
Resetting config helped, thanks. Its weird, because it was my 1st launch and already with broken config.
And that "Config file" submenu in russian looks like file setup/configuration, and the buttons are "Setting", "Save settings" and "reload".
It seems that a problem occurred during the initialization process.
Thank you for reporting the worng text! I'll fix it
|
2025-04-01T06:39:17.473768
| 2024-09-03T11:52:41
|
2502675703
|
{
"authors": [
"kiloutyg"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7560",
"repo": "kiloutyg/efnc",
"url": "https://github.com/kiloutyg/efnc/issues/100"
}
|
gharchive/issue
|
Updating stuff for deployment outside initial plant
Redo :
install script
update script
correct bugs at startup
Update globals variable and their use in email and other stuff.
|
2025-04-01T06:39:17.529691
| 2022-01-21T13:47:54
|
1110490075
|
{
"authors": [
"kingsaint",
"phlobo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7561",
"repo": "kingsaint/BioMedical-EL",
"url": "https://github.com/kingsaint/BioMedical-EL/issues/7"
}
|
gharchive/issue
|
Documentation of setup for BioBERT and neleval
First of all, thanks for sharing this repo!
However, I have a bit of trouble reproducing the results, maybe you can help.
Running the training script
python run_e2e_span.py --data_dir data/BC5CDR/processed_data --model_type bert --output_dir o utput --do_train --n_gpu 1 --model_name_or_path ./biobert_v1.1_pubmed
yields different errors.
If I use the BioBERT checkpoint from here https://github.com/dmis-lab/biobert:
I get the error: OSError: file ./biobert_v1.1_pubmed/config.json not found
If, however, I use this checkpoint https://huggingface.co/dmis-lab/biobert-v1.1/tree/main
I'm getting OSError: Error no file named ['pytorch_model-1000000.bin', 'tf_model.h5', 'model.ckpt.index'] found in directory ./biobert-v1.1/ or from_tf set to False
neleval
The Readme states that evaluation is run with ./neleval/BC5CDR-AllSpan/gold.csv. How do I get this file? Is it some produced as part of running the e2e_span script?
Thanks in advance!
Hi,
Thanks for your interest in our paper.
If you are using https://github.com/dmis-lab/biobert, please rename the bert_config.json file in ./biobert_v1.1_pubmed to config.json. Alternatively, you can specify the path to the config file using the --config_name argument in the training script.
If you are using https://huggingface.co/dmis-lab/biobert-v1.1/tree/main, please rename pytorch_model.bin to pytorch_model-1000000.bin.
gold.csv will be created when you do inference on the trained model. Use the --do_eval argument to do so. Then copy the gold.csv file to a directory named BC5CDR-AllSpan.
I hope this is helpful.
Thank you so much! Regarding 1. I tried the second option (using the HuggingFace model) and actually progressed a bit further. However, now I am getting an error
01/25/2022 12:12:17 - INFO - __main__ - Loading features from cached file data/BC5CDR/processed_data/cached_train_biobert-v1.1
Traceback (most recent call last):
File "run_e2e_span.py", line 1071, in <module>
main()
File "run_e2e_span.py", line 1021, in main
global_step, tr_loss = train(args, model, tokenizer)
File "run_e2e_span.py", line 76, in train
train_dataset, _, _ = load_and_cache_examples(args, tokenizer)
File "run_e2e_span.py", line 725, in load_and_cache_examples
all_mention_start_indices = torch.tensor([f.mention_start_indices for f in features], dtype=torch.long)
ValueError: expected sequence of length 16 at dim 1 (got 9)
Do you have any hints what might be the problem?
use --overwrite_output_dir, --overwrite_cache, and --use_hard_and_random_negative during training. During inference use --use_all_candidates
|
2025-04-01T06:39:17.531539
| 2024-03-20T15:56:49
|
2197903816
|
{
"authors": [
"kingyue737",
"xj63"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7562",
"repo": "kingyue737/vitify-next",
"url": "https://github.com/kingyue737/vitify-next/pull/1"
}
|
gharchive/pull-request
|
fix(vite.config): mdi icon convert camelCase to kebab-case
When the previous solution encounters mdiBattery80, it will be converted to mdi-battery-8-0.
When using mdi-battery-80, an error that the icon cannot be found will be prompted.
This solution first converts uppercase letters and then matches one or more numbers.
There are some weird names here, such as mdiBattery10Bluetooth. : )
Thank you very much!
|
2025-04-01T06:39:17.541160
| 2024-02-19T17:20:53
|
2142851072
|
{
"authors": [
"dr-frmr",
"hosted-fornet",
"nick1udwig"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7563",
"repo": "kinode-dao/kinode",
"url": "https://github.com/kinode-dao/kinode/issues/259"
}
|
gharchive/issue
|
terminal: pasting causes cursor to go to end of line
E.g. imagine I have a line
foo baz
and I have my cursor on the space and bar in the clipboard. Pasting will properly modify the line, but will move my cursor to the end (i.e. to the z in baz), rather than, as expected, the end of the paste (the space in bar ).
FYI @tadad @dr-frmr (not sure who owns this part of terminal).
fixed in develop!
|
2025-04-01T06:39:17.544550
| 2020-08-03T18:10:48
|
672253572
|
{
"authors": [
"mauriciovasquezbernal"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7564",
"repo": "kinvolk/inspektor-gadget",
"url": "https://github.com/kinvolk/inspektor-gadget/pull/126"
}
|
gharchive/pull-request
|
Some improvements for network policy advisor
Align "--namespace" and "--namespaces" flags
Add --output option for report comman
@alban I think these changes are still relevant.
|
2025-04-01T06:39:17.546756
| 2017-09-05T10:53:57
|
255237321
|
{
"authors": [
"dongsupark",
"jonboulle"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7565",
"repo": "kinvolk/kube-spawn",
"url": "https://github.com/kinvolk/kube-spawn/issues/130"
}
|
gharchive/issue
|
Support overlayfs on btrfs after moby's new release
Currently we don't support btrfs as underlying device's filesystem, mainly because moby doesn't allow it. Linux kernel 4.7 or newer already supports overlayfs on btrfs, but moby started to support since https://github.com/moby/moby/pull/33281 (2017-05-19). The commit has not been included in any release so far. Even the most recent release 17.05.0-ce doesn't have it.
So we need to wait for a while until moby released a new version with it, as well as Container Linux includes the new version. Then we can allow overlayfs on btrfs.
/cc @iaguis
fixed by #193
Right. Thanks!
|
2025-04-01T06:39:17.562037
| 2017-07-21T10:03:59
|
244623298
|
{
"authors": [
"kirilkirkov",
"vikaskumarmca"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:7566",
"repo": "kirilkirkov/Shopping-Cart-Solution-CodeIgniter",
"url": "https://github.com/kirilkirkov/Shopping-Cart-Solution-CodeIgniter/issues/15"
}
|
gharchive/issue
|
index.php
hi
thank for sharing good source but i am running this code if am facing index.php problem each and every open page from admin.
http://localhost/shopping/admin/publish Problem
http://localhost/shopping/index.php/admin/publish right
@vikaskumarmca This is not a problem. You can load admin/publish with index.php before and without :).
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.