added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:34:57.142940
2022-02-16T19:14:44
1140451342
{ "authors": [ "nxtlo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9197", "repo": "nxtlo/aiobungie", "url": "https://github.com/nxtlo/aiobungie/pull/144" }
gharchive/pull-request
Fix HTML error responses. Long story short Checks [x] I have run nox tests and all pipelines have passed. [x] I have wriiten tests according to the code i added, modified or deleted. Related issues @GregShiner Please check if this branch fixes the issue and raises the actual error for your needs or not. Installing by pip install git+https://github.com/nxtlo/aiobungie@bug/fix-content-type-errors Stuff that i noticed. Bungie is returning 500 status code when making a request that requires OAuth2 and failed which instantly should return 401. Bungie is returning HTML instead of JSON as an API response. Who knows if this HTML redirect page is always caused by methods that requires authorization?. So as far as i know the issue you're facing can't really be handled internally by the lib since it's kinda a Bungie thing. The HTML will require us to read the bytes of the response, decode it, and check what caused the error parsing this returned HTML page. But even if this can fix your issue, I'm not really sure if other errors can cause the same HTML redirect which's kinda unsafe to write something like this. Any thought regarding this will be helpful! What i would personally do is handle this error personally since we know whats causing it in this context import aiobungie try: await oauth_stuff("token") except aiobungie.HTTPError as e: if "text/html" in e.message: # We know why this error happened. ...
2025-04-01T04:34:57.272681
2021-11-04T22:56:17
1045283171
{ "authors": [ "bufferhe4d", "duguorong009", "mimoo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9199", "repo": "o1-labs/proof-systems", "url": "https://github.com/o1-labs/proof-systems/issues/213" }
gharchive/issue
[kimchi] distinguish ZK_ROWS and PERM_FINAL_ACC We currently use 2 zero-knowledge rows, but specify 3 in the code (under the constant ZK_ROWS). We use this to create enough space for three equations: one constraint for the final accumulator value of the permutation. In the original PLONK protocol this constraint does not exist as the constraint on the initial accumulator value takes care of that (because the accumulator wraps around). With our way of doing zero-knowledge we need to add that last constraint in a new row. two random values, for the zero-knowledge, as intended It'd be nice to reflect that in the code. Also, due to that we currently randomize the last three rows of the witness columns, which is not a big deal as they cannot be used by the circuit, but it would be good to document that :o) random thought: intuitively, it sounds like we could use unused columns in zero gates (or other gates) to randomize the witnesses and save some rows. It doesn't save us that many rows though (2 is really negligible) and it wouldn't actually save us anything because we still need that space for the permutation. Nevermind. Adding more color to this issue: The relevant code is here: https://github.com/o1-labs/proof-systems/blob/master/kimchi/src/prover.rs#L109 //~ 1. Ensure we have room in the witness for the zero-knowledge rows. //~ We currently expect the witness not to be of the same length as the domain, //~ but instead be of the length of the (smaller) circuit. //~ If we cannot add `ZK_ROWS` rows to the columns of the witness before reaching //~ the size of the domain, abort. let length_witness = witness[0].len(); let length_padding = d1_size .checked_sub(length_witness) .ok_or(ProofError::NoRoomForZkInWitness)?; if length_padding < ZK_ROWS as usize { return Err(ProofError::NoRoomForZkInWitness); } //~ 2. Pad the witness columns with Zero gates to make them the same length as the domain. //~ Then, randomize the last `ZK_ROWS` of each columns. for w in &mut witness { if w.len() != length_witness { return Err(ProofError::WitnessCsInconsistent); } // padding w.extend(std::iter::repeat(Fr::<G>::zero()).take(length_padding)); // zk-rows for row in w.iter_mut().rev().take(ZK_ROWS as usize) { *row = Fr::<G>::rand(rng); } } The constant ZK_ROWS is currently set to 3, but it should be set to 2 as we only need to randomize 2 rows (since we provide 2 evaluations in the protocol). That being said, we do not need to pad 3 rows here, as the permutation needs 3 clean rows at the end (1 for the final accumulator value, see the blogpost I linked, and 2 for adding zero-knowledge to the permutation argument) Currently working on this. there's also ZK_ROWS in lookup as well. I think it might be good to rename ZK_ROWS to EVALS or some constant that says how many evaluations we have in the protocol. (Or compute ZK_ROWS as EVALS + 1 just to be overly cautious. @mimoo Please check the PR #974, and give me comments.
2025-04-01T04:34:57.277440
2024-11-05T15:11:39
2635738025
{ "authors": [ "epugh", "wrigleyDan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9200", "repo": "o19s/opensearch-hybrid-search-optimization", "url": "https://github.com/o19s/opensearch-hybrid-search-optimization/issues/3" }
gharchive/issue
Share links to public repo and solicit feedback Share links etc post video. links shared, closing ticket
2025-04-01T04:34:57.279298
2017-08-22T09:21:24
251891805
{ "authors": [ "nuest" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9201", "repo": "o2r-project/o2r-muncher", "url": "https://github.com/o2r-project/o2r-muncher/issues/55" }
gharchive/issue
Test out --squash option when building images Take a loog at how the created image tarball looks - is squashing a plus for reproducibility? https://stackoverflow.com/questions/41764336/how-does-the-new-docker-squash-work see also example in README of https://github.com/goldmann/docker-squash Squashing may improve manual inspection after a long time, because there are not multiple layers to explore. But it also reduces transparency as there is not one layer per Dockerfile instruction.
2025-04-01T04:34:57.284419
2024-10-24T16:22:17
2612055627
{ "authors": [ "Pacatro", "spenserblack" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9202", "repo": "o2sh/onefetch", "url": "https://github.com/o2sh/onefetch/issues/1443" }
gharchive/issue
Language Request: COBOL Language Name πŸ–Š COBOL Logo πŸ“· Upstream support βœ… [X] Does tokei already support the language in question? Frankly, I'm surprised we don't already support it :sweat_smile: If you're planning on making a PR I'd suggest waiting until #1305 drops, which might happen soon. Frankly, I'm surprised we don't already support it :sweat_smile: If you're planning on making a PR I'd suggest waiting until #1305 drops, which might happen soon. Got It! I'll keep an eye on that 😁 Hi again! Just wondering if it’s cool to go ahead with the PR now. Yeah, I think you can go ahead and make a PR. In hindsight, there wasn't really a reason to wait.
2025-04-01T04:34:57.287465
2019-11-14T09:32:33
522731668
{ "authors": [ "AKolakowska47", "o2sh", "spenserblack" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9203", "repo": "o2sh/onefetch", "url": "https://github.com/o2sh/onefetch/issues/163" }
gharchive/issue
Support for Groovy I would be awesome if you can add support for Groovy :) @AKolakowska47 are you okay with this? If so, can you plz close this issue. Thanks @o2sh Your commit seems to be OK, but I'm waitin' for binary to check it ;) When the new one (2.1.1.) will be available? @AKolakowska47 This binary was built from 90bdc4ecb48d71e23871e4266a96e5d381c7d226 Onefetch 2.2.0 is now available.
2025-04-01T04:34:57.289134
2023-02-06T13:35:43
1572584155
{ "authors": [ "davidsilveira", "maartenesser" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9204", "repo": "o2web/graphql-auth", "url": "https://github.com/o2web/graphql-auth/pull/34" }
gharchive/pull-request
Add support to Rails 7 and Ruby 3 set graphql dependency to rails 7 update ruby version to version 3 Hi @davidsilveira, Do you have an estimate on when this PR will be merged? @maartenesser I'm using this code in a project already and haven't had any problems yet. Gemfile: gem 'graphql-auth', git: 'https://github.com/davidsilveira/graphql-auth.git', branch: 'rails-7'
2025-04-01T04:34:57.311071
2014-12-16T17:04:50
52139199
{ "authors": [ "nicolaasmatthijs", "sathomas" ], "license": "ECL-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9205", "repo": "oaeproject/3akai-ux", "url": "https://github.com/oaeproject/3akai-ux/issues/3960" }
gharchive/issue
(Bug Bash) Android Issues Not sure how much (if any) we want to fix, but documented for posterity: Android 4.1.1 on a Huawei Prism II smart phone. With the default web browser The min-width: 400px design doesn't fit on the screen In additions, the .woff format icon font is not supported And the site is basically not navigable With Firefox 34.0 The min-width: 400px design doesn't fit on the screen However, all other functionality seems to work okay. With Chrome 39.0.2171.93 Everything works as expected Moving to the technical backlog as this won't be an immediate priority
2025-04-01T04:34:57.326149
2023-08-10T11:14:42
1844969242
{ "authors": [ "monkey-w1n5t0n", "oakes" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9206", "repo": "oakes/odoyle-rules", "url": "https://github.com/oakes/odoyle-rules/issues/24" }
gharchive/issue
and 10 is "you need a PhD in parallel systems and two decades of industry experience", how much do I have to think before I type to work with an odoyle session from multiple threads - and is it worth it? Are there any obvious gotchas, or any areas that may have some not-so-odvious gotchas and require more thought? More specifically, I'm confused when it comes to thinking about facts being inserted vs rules being fired at different times, or whether it's possible (or even desirable) to selectively fire rules on specific threads. If it is, then do we have to ensure that the rules don't interact in any way, or can that still be managed? I guess what I'm ultimately wondering about is whether rule firing can be selective and incremental, so that in an architecture like this: thread B / \ thread A thread D \ / thread C Thread A can fire only those rules that both B and C require, B and C fire independent rules in parallel, and thread D merges everything and carries on. If it's possible, can the dependency graph be determined automatically, so that thread A knows which rules it has to fire given which rules threads B and C each want to fire themselves? Example use case Let's assume that I'm writing a TODO app with Vim-style modal editing and that I want something like the following async architecture, where each number is a loop running on a separate thread (real or core.async pool): Receives keyboard and mouse events, timestamps them, and o/inserts them into some kind of queue in the session, without firing anything so that it doesn't block and miss(time) further incoming events. Pulls from the event queue and fires only those rules that are triggered by those events that have been inserted, in the order they were inserted, with the state as it was at the time they were stamped. This presumably generates and inserts other facts, such as new TODO entries, TODO state changes, app mode changes etc. Fires all other rules that may have been matched by the insertion of facts by no. 2, such as creating/modifying/deleting GUI components according to TODO entry creation/modification/deletion above, reinserting those into the session's GUI-related state. o/querys the GUI-related state, topologically sorting them according to whether they overlap and passing each strand to be rendered in parallel by separate threads. Example entries made by no. 1 (in chronological order & grouped by their relevance, see below): {:key "n" :time 123}, {:key "esc" :time 124}, {:key "n" :time 125}, {:key "enter" :time 126}, {:key "tab" :time 127} {:key "n" :time 128} {:key "A" :time 129}, {:key "B" :time 130}, {:key "enter" :time 131} Example rules that are fired by no. 2 based on the above: If the state is normal and the key n is pressed, make a new empty TODO, set it as current, and set the state to insert. If the state is insert, the current TODO is empty, and the key esc is pressed, delete the TODO and set the state to normal. If the state is insert and the key enter is pressed, save the current TODO and set the mode to normal. If the state is normal and the key tab is pressed, set the current TODO's state to done. If the state is insert and any other key is pressed, type the character in the TODO's text. All of the above would result in: The creation and instant deletion of an empty TODO (which happened so fast that it would never reach thread no. 3 and therefore no GUI component would be needlessly created) The creation of an empty TODO, which is promptly marked as done so that I can pat myself on the back and take a well-deserved break to eat some cake. The creation of a TODO with the text "AB". More questions Is this sort of thing even possible? Obviously an overkill for a TODO app, but I can imagine this sort of architecture being useful for apps like IDEs that would love to be as-asynchronous-as-possible and have this kind of declarative logic. If it's not possible with a single session, then could it work with each thread having its own session and have a loop that looks like this: read events that previous thread left in the channel -> insert them into own session -> fire rules -> do stuff with the results -> put new facts on a channel for the next thread (I imagine this assumes that the rules for each thread's session don't have to interact, and that if they do that the state is duplicated and synced across the threads?) 3. What's the overhead in calling fire-rules? Is there a benefit in batching inserts before fire-rules is called, or would that cause problems? Would it help with avoiding unnecessary creation of GUI widgets, e.g. in the case where a TODO is created and then instantly deleted? Outro Thanks for writing and maintaining this awesome library, sorry for the barrage of questions and if half of them are obvious or don't make sense (or if they're a duplicate of #20 - that seemed more general than what I had in mind), I'd love to contribute to it at some point, writing async GUIs is hard, cheers & bye. If you want to process state changes on separate threads but still cleanly merge their results together, I think you need a CRDT. They are designed for exactly that purpose: merging independent changes in a way that produces a deterministic result. Rules can theoretically trigger any other rule, so I don't see a way to automatically separate them into silos that execute independently. If you managed to do this, it would sound quite similar to just using separate sessions with different rulesets. To answer your question about the overhead of fire-rules, it is generally better to just call it once per "frame" rather than after each insert. In fact, I made an early mistake in pararules by making it fire rules automatically after every insert. You have to manually turn autoFire off to disable this behavior. In odoyle I don't even have an "auto fire" feature, because it's just wasteful and provides a really minor convenience.
2025-04-01T04:34:57.338446
2020-12-11T04:43:02
761895474
{ "authors": [ "leejanee", "ryanzhang-oss", "wonderflow" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9207", "repo": "oam-dev/kubevela", "url": "https://github.com/oam-dev/kubevela/pull/768" }
gharchive/pull-request
Make the definitionRef field of traitDef optional [x] Support Definition Reference be Optional #749 Modified TraitDefinition CRD, and removed the definitionRef field from required list The patch type traitDefinition does not need the definitionRef field, and only the application will use it. Is this PR still valid? closing as it's already in the code base. closing as it's already in the code base.
2025-04-01T04:34:57.343058
2023-05-22T08:42:06
1719195665
{ "authors": [ "JkSelf" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9208", "repo": "oap-project/gluten", "url": "https://github.com/oap-project/gluten/pull/1723" }
gharchive/pull-request
[VL] Add parquet write benchmark What changes were proposed in this pull request? This PR add the parquet writer benchmark in velox backend. (Fixes: #ISSUE-ID) How was this patch tested? (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests) (If this patch involves UI changes, please attach a screenshot; otherwise, remove this) @jinchengchenghh Do you have any further comment?
2025-04-01T04:34:57.362392
2023-11-03T06:00:42
1975506380
{ "authors": [ "GlutenPerfBot", "JkSelf", "jinchengchenghh", "zhli1142015" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9209", "repo": "oap-project/gluten", "url": "https://github.com/oap-project/gluten/pull/3610" }
gharchive/pull-request
[GLUTEN-3594] [VL] Allow users to set bloom filter configurations What changes were proposed in this pull request? Fixes: #3594 How was this patch tested? (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests) (If this patch involves UI changes, please attach a screenshot; otherwise, remove this) Please also update the document, https://github.com/oap-project/gluten/blob/main/docs/Configuration.md Please also update the document, https://github.com/oap-project/gluten/blob/main/docs/Configuration.md Added, thanks. @zhli1142015 Can you help to rebase again? Thanks. @zhli1142015 Can you help to rebase again? Thanks. Rebased, thanks. ===== Performance report for TPCH SF2000 with Velox backend, for reference only ==== query log/native_3610_time.csv log/native_master_11_07_2023_e3eff1d8f_time.csv difference percentage q1 34.24 34.38 0.137 100.40% q2 25.01 25.03 0.012 100.05% q3 39.76 38.14 -1.612 95.95% q4 37.72 37.57 -0.148 99.61% q5 70.81 71.50 0.692 100.98% q6 7.98 6.26 -1.720 78.45% q7 84.36 82.22 -2.134 97.47% q8 85.73 86.95 1.222 101.43% q9 120.45 119.81 -0.639 99.47% q10 52.57 51.26 -1.306 97.52% q11 19.94 19.73 -0.213 98.93% q12 27.37 24.39 -2.980 89.11% q13 48.50 50.30 1.793 103.70% q14 16.80 17.67 0.865 105.15% q15 31.96 30.35 -1.609 94.97% q16 16.34 16.20 -0.141 99.14% q17 102.47 101.51 -0.961 99.06% q18 145.74 148.26 2.519 101.73% q19 14.79 16.17 1.386 109.37% q20 30.14 30.31 0.173 100.57% q21 222.04 224.88 2.842 101.28% q22 13.84 14.08 0.245 101.77% total 1248.56 1246.98 -1.576 99.87%
2025-04-01T04:34:57.369674
2024-08-21T07:16:11
2477261272
{ "authors": [ "jamietanna", "venkat-pandey" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9210", "repo": "oapi-codegen/oapi-codegen", "url": "https://github.com/oapi-codegen/oapi-codegen/issues/1732" }
gharchive/issue
Change json tag fields case For one of our use-case, we need to have the json tag fields generated in snake_case, by default codegen generates camelCase, i didnt find any toggle flag to alter this behaviour. current: type Foo struct { BarBar int 'json:barBar' } required: type Foo struct { BarBar int 'json:bar_bar' } is there way to override the part where it creates the tag fields and add your own implementation? Could you please share the snippet of the OpenAPI definition for the object you're trying to represent? oapi-codegen will generally follow the name of the fields within the source schema, so we generate a tag that corresponds with the field name https://github.com/oapi-codegen/oapi-codegen?tab=readme-ov-file#changing-the-names-of-generated-types may be possible to use, but I'd like to see the relevant part of your spec first Thanks for reply, LiveData: description: Defines the structure of Live Data type: object properties: id: type: string x-go-name: ID x-go-type: uuid.UUID x-go-type-import: path: github.com/google/uuid name: uuid measuredTimestamp: type: string x-go-type-skip-optional-pointer: true format: date-time with this spec i get json tag as `measuredTimestamp` where as it would be great to have `measured_timestamp` Thanks, can you please edit the πŸ‘†πŸΌ so it correctly shows as a multi-line code block, for readability? Out of interest, why do you want a different JSON tag used? The measuredTimestamp indicates that the request/response body will include a property with that name. If you're integrating with an API that's given you this spec, modifying the tag will result in you incorrectly receiving the property Out of interest, why do you want a different JSON tag used? The measuredTimestamp indicates that the request/response body will include a property with that name. If you're integrating with an API that's given you this spec, modifying the tag will result in you incorrectly receiving the property we want a REST endpoint for an existing service using protobuf schema with json tags defined in snake_case (cannot be changed). so to make it consistent across for the users. In that case, can you rename the field in your OpenAPI spec instead? That'll generate the right tags for you In that case, can you rename the field in your OpenAPI spec instead? That'll generate the right tags for you I guess that can work with conjunction to using x-go-name. thanks for help.
2025-04-01T04:34:57.378328
2017-11-18T16:46:28
275092058
{ "authors": [ "depakaakula", "oarriaga" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9211", "repo": "oarriaga/face_classification", "url": "https://github.com/oarriaga/face_classification/issues/43" }
gharchive/issue
Issue with training dataset Do I have to let it train 10000 times really? No, the model checkpoint will stop saving the models whenever the validation loss function stops decreasing. Thanks. Is it possible to generate the output file for whatever is getting logged in terminal? Thank you. On Mon, Nov 20, 2017 at 04:42, Octavio Arriaga<EMAIL_ADDRESS>wrote: No, the model checkpoint will stop saving the models whenever the validation loss function stops decreasing. β€” You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/oarriaga/face_classification/issues/43#issuecomment-345642466, or mute the thread https://github.com/notifications/unsubscribe-auth/Ae_KFHIXdSjbsL7vrWAVTLBkYBbl8AYgks5s4UmhgaJpZM4QjECd . There is a callback from keras that allows you to save the losses in a csv file. I believe the current training script contains such callback and is generating the log files while running. I see the log is kinda unstructured. For the people who are not technical enough to understand these numbers, is there a way to tune it a bit? Like when and which each emotion is displayed. Also, does each expression has unique value? Thank you. On Mon, Nov 20, 2017 at 8:10 AM, Octavio Arriaga<EMAIL_ADDRESS>wrote: Closed #43 https://github.com/oarriaga/face_classification/issues/43. β€” You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/oarriaga/face_classification/issues/43#event-1349830682, or mute the thread https://github.com/notifications/unsubscribe-auth/Ae_KFMze-KBgxahxAToEDqmEW7jtUn5jks5s4Xo_gaJpZM4QjECd . Dear oarriaga. The log file (fer2013_emotion_training.log) doesn't seem to be capturing my recent terminal report. Also, the output file data is kinda unstructured. Can you help me understand how to relate the expression and digits in the output file? Thank you.
2025-04-01T04:34:57.420803
2016-10-14T04:06:03
182959041
{ "authors": [ "obedm503", "urmilparikh" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9212", "repo": "obedm503/bootmark", "url": "https://github.com/obedm503/bootmark/issues/6" }
gharchive/issue
Option to disable source highlighting How do I disable source highlighting for code blocks? It wrongly assumes // in http://xyz as comment marker There is strange highlighting on 2nd line you can disable it or just set the language, this might be your problem for a list of supported languages see the code prettify faq but if you really want to disable to disable code highlighting (thru code prettify) you can: if using the <bootmark> element add the html attribute and set it equal to an object with prettify set to false <bootmark html="{ prettify: false }"> markdown here </bootmark> if using any element just add data- <div class="bootmark" data-html="{ prettify: false }"> markdown here </div> or use global config options $.fn.bootmark.options = { prettify: false }; this is all mentioned in the docs and the examples The docs page looked like lot of code, so I overlooked it. Anyway, prettify: false worked, thanks a lot!
2025-04-01T04:34:57.429964
2024-11-21T22:42:52
2681222087
{ "authors": [ "0crat", "deemp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9213", "repo": "objectionary/eo-phi-normalizer", "url": "https://github.com/objectionary/eo-phi-normalizer/pull/542" }
gharchive/pull-request
541 some files weren't committed on the PR Closes #541 @deemp Thank you for your contribution! Your effort is appreciated, but there's room for improvement. You've earned 4 points: +4 base, -16 for no code review, -16 for no reviewer comments, -8 for only 6 hits-of-code, and +40 adjusted to meet the minimum. To maximize your bonus next time, aim for more hits-of-code, ensure code reviews, and encourage reviewer comments. Remember, quality and speed are both crucial. Keep the contributions coming! πŸ’»
2025-04-01T04:34:57.438127
2024-07-12T15:43:00
2405887422
{ "authors": [ "0pdd", "rultor", "volodya-lombrozo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9214", "repo": "objectionary/opeo-maven-plugin", "url": "https://github.com/objectionary/opeo-maven-plugin/issues/344" }
gharchive/issue
Constructor.java:42-45: Continue enable 'spring-fat'... The puzzle 329-93a273ea from #329 has to be resolved: https://github.com/objectionary/opeo-maven-plugin/blob/c2af3985f231c712e8b53ede8b6e102a3747c6d4/src/main/java/org/eolang/opeo/ast/Constructor.java#L42-L45 The puzzle was created by @volodya-lombrozo on 12-Jul-24. Estimate: 90 minutes, role: DEV. If you have any technical questions, don't ask me, submit new tickets instead. The task will be "done" when the problem is fixed and the text of the puzzle is removed from the source code. Here is more about PDD and about me. @0pdd the puzzle #361 is still not solved. @0pdd 2 puzzles #361, #363 are still not solved. @0pdd 3 puzzles #361, #363, #366 are still not solved. @rultor release, tag is 0.3.1, title is Version 0.3.1. @rultor release, tag is 0.3.1, title is Version 0.3.1. @volodya-lombrozo OK, I will release it now. Please check the progress here @rultor release, tag is 0.3.1, title is Version 0.3.1. @volodya-lombrozo Done! FYI, the full log is here (took me 10min)
2025-04-01T04:34:57.444888
2016-09-12T08:20:46
176318754
{ "authors": [ "igorspasic", "titanseason" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9215", "repo": "oblac/jodd", "url": "https://github.com/oblac/jodd/issues/336" }
gharchive/issue
NullPointerException when use @PetiteInject in application's ready method How this happened: class AppWebAplication extends PetiteWebApplication Overwrite ready() method in AppWebAplication @Override protected void ready() { super.ready(); IndexGenerator generator = new IndexGenerator(); generator.generate(); } In IndexGenerator: @PetiteInject UserDao userDao; public void generate() { List<User> list = userDao.listAll(User.class); } Thus, NullPointerException happens (userDao is null) I want to do something when the application is started Heya @titanseason ! You can't do this: IndexGenerator generator = new IndexGenerator(); Because then Petite is not in the control of the bean lifecycle. You have created an instance manually, and Petite can' inject anything. So what you can do? You can register IndexGenerator as PetiteBean and lookup for it (getBean), instead of calling new. Or you can just wire it: petiteContainer.wire(generator), without adding instance in the container. Of course, you can addBean later. Hope this helps! Tks, @igorspasic , now my app back to normal nomal ! I overwrite ready() method in AppWebAplication @Override protected void ready() { super.ready(); PetiteContainer petite = providePetiteContainer(); IndexGenerator generator = petite.getBean(IndexGenerator.class); generator.generate(); } Perfect :)))
2025-04-01T04:34:57.453278
2020-07-15T08:13:07
657155041
{ "authors": [ "25juan", "AlessioCamm", "EmreDereli", "Harisene", "arkanos", "cisplRejjak", "fabiendem", "lucjan", "saxenanickk", "tambegaurav", "vishwanath-ogit" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9216", "repo": "oblador/react-native-vector-icons", "url": "https://github.com/oblador/react-native-vector-icons/issues/1202" }
gharchive/issue
Unable to resolve module ./create-icon-source-cache Does someone has the same problem ? I have this problem with my project when I build on XCode Unable to resolve module "./create-icon-source-cache" from "/Users/eonixpro/Git/Bonobo/Safetify/SafetifyMobileWeb/node_modules/react-native-vector-icons/lib/create-icon-set.js" : The module "./create-icon-source-cache" could not be found from "/Users/eonixpro/Git/Bonobo/Safetify/SafetifyMobileWeb/node_modules/react-native-vector-icons/lib/create-icon-set.js". I'm using react-native : 0.60.6 & react-native-vector-icons: 6.6.0 @AlessioCamm I am also facing same issue but on Android Platform. I didn't try iOS. react-native: 0.61.0 react-native-vector-icons: 6.6.0 Previously I was getting some issue related to toolbar-android. I fixed that by removing "^" from react-native-vector-icons: ^6.6.0. Still not solved yet any update on this issue? Experiencing same issue on 6.6.0, having previously been fighting with missing toolbar-android issues. me too, How to solve this problem I removed ^ from react-native-vector-icons: ^6.6.0 then npm start --reset-cache It worked for me I am facing this issue. Anyone with solution please reply... error: Error: Unable to resolve module ./create-icon-source-cachefromnode_modules\react-native-vector-icons\lib\create-icon-set.js`: None of these files exist: node_modules\react-native-vector-icons\lib\create-icon-source-cache(.native|.android.js|.native.js|.js|.android.json|.native.json|.json|.android.ts|.native.ts|.ts|.android.tsx|.native.tsx|.tsx) node_modules\react-native-vector-icons\lib\create-icon-source-cache\index(.native|.android.js|.native.js|.js|.android.json|.native.json|.json|.android.ts|.native.ts|.ts|.android.tsx|.native.tsx|.tsx)` I removed ^ from react-native-vector-icons: ^6.6.0 Then deleted pakage-lock and node modules and did npm i again. It worked for me πŸ˜„ Nothing worked for me but removing the package from package.json and doing a fresh install with npm i --save react-native-vector-icons for getting the latest version (7.1.0) I wiped the node_modules and re-did the install (yarn install for me). That did the job. changing the node version to 14.16.0 resolved the issue for me.
2025-04-01T04:34:57.462403
2015-11-05T13:41:45
115285722
{ "authors": [ "marsprince", "oblador", "satya164", "timzaak" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9217", "repo": "oblador/react-native-vector-icons", "url": "https://github.com/oblador/react-native-vector-icons/issues/49" }
gharchive/issue
Not working with RN master I know that it requires a stable version, but would appreciate you can help me fixing the issue. Error: Invariant Violation: RawText "" must be wrapped in an explicit <Text> component. stack: assign.mountComponent index.android.bun…:15980 Object.ReactReconciler.mountComponent index.android.bun…:4000 ReactMultiChild.Mixin.mountChildren index.android.bun…:20896 ReactNativeBaseComponent.Mixin.initializeChildren index.android.bun…:17119 ReactNativeBaseComponent.Mixin.mountComponent index.android.bun…:17243 Object.ReactReconciler.mountComponent index.android.bun…:4000 ReactCompositeComponentMixin.mountComponent index.android.bun…:11297 ReactPerf.measure.wrapper [as mountComponent] index.android.bun…:2792 Object.ReactReconciler.mountComponent index.android.bun…:4000 URL: undefined line: undefined message: Invariant Violation: RawText "" must be wrapped in an explicit <Text> component.handleException @ ExceptionsManager.js:68handleError @ InitializeJavaScriptAppEngine.js:37ErrorUtils.reportFatalError @ error-guard.js:28guard @ MessageQueue.js:41callFunctionReturnFlushedQueue @ MessageQueue.js:82messageHandlers.executeJSCall @ debuggerWorker.js:25onmessage @ debuggerWorker.js:42 Hi, I'll wait until there's at least a release candidate :-) Just tried with 0.14 that was released yesterday without problems. Next release will have core support for custom fonts so then we don't need our custom Text component anyway. @oblador Awesome. Closing this :) BTW, one problem, the folder name used this project is assets/Fonts, while the one used by RN is assets/fonts. With OS X treating both folders as same as it's case insensitive (as well as windows), it'll cause problems. Yeah, it might since android probably is case sensitive but it's just a matter of renaming the folder really. Annoying with three simultaneous naming conventions :-) A bigger problem might be git on existing installations though, doesn't handle case changes to well in my experience. @oblador Yeah, the problem was that before adding RN vector icons, I had it as assets/fonts, and after adding I renamed it to assets/Fonts. But because of OS X's case insensitiveness, I couldn't commit to GIT, as it saw no changes. I had to remove the folder from GIT, and then re-add, to get it working. Worst is, I didn't notice it, until someone else told me that he can't build after pulling from GIT. is there any plan to support 'v0.15.0-rc'? Yes, just very busy ATM :see_no_evil: :+1: same problems in RN 0.15,I tyied assets/Fonts and assets/fonts,all error. Fixed in v0.8.5. Btw it seems 0.15 won't support custom fonts in core after all, it didn't make it into the RC so don't think it will in the final release either. Lets hope for 0.16 instead. @oblador thank you,guys ! It worked it , however it showed nothing , just blanks.My version is RN 0.15 , android 4.4,any suggestions? @marsprince: are you using a regular <Icon> component or getImageSource()? @oblador just icon like <Icon name="arrow-back" size={30} color="#FFFFFF"></Icon> @marsprince: Don't know, you might not have the fonts installed correctly. For now it should still be Fonts with capital F. If you want to use the core font support, then you can try this fork. As of RN 0.16 and 1.0.0-rc, the custom Text component is no more!
2025-04-01T04:34:57.465954
2018-08-01T19:32:17
346737635
{ "authors": [ "cmungall", "shawntanzk" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9218", "repo": "obophenotype/cell-ontology", "url": "https://github.com/obophenotype/cell-ontology/issues/519" }
gharchive/issue
Single cell transcriptome profiling of retinal ganglion cells identifies cellular subtypes https://www.nature.com/articles/s41467-018-05134-3 40 proposed RGC subtypes Unclear how these would be represented but making a ticket anyway https://github.com/obophenotype/provisional_cell_ontology/issues/36
2025-04-01T04:34:57.472812
2023-05-12T12:40:51
1707567427
{ "authors": [ "aleixpuigb" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9219", "repo": "obophenotype/cell-ontology", "url": "https://github.com/obophenotype/cell-ontology/pull/1979" }
gharchive/pull-request
Addresses #1908 to create an upper cell type slim Addresses #1908 to create an upper cell type slim I have tested the coverage for animal cell (CL:0000548), and it is 90%, so not bad (Full report here: #####Coverage percentage##### | Β  -- | -- 90.46% | Β  #####Number of terms covered by each term in the slim##### | Β  exocrine cell | 23 bone cell | 39 smooth muscle cell | 59 striated muscle cell | 75 endocrine cell | 89 stem cell | 103 connective tissue cell | 206 neural cell | 595 hematopoietic cell | 659 epithelial cell | 721 I think I'd like something a bit more granular That is closer to my first proposal https://github.com/obophenotype/cell-ontology/issues/1908#issuecomment-1545658060, and then we decided to go to a higher level. I can use that one with some edits discussed with @bvarner-ebi. This list adds some granularity and still has a coverage above 83%: #####Coverage percentage##### | Β  -- | -- 83.37% | Β  #####Number of terms covered by each term in the slim##### | Β  fat cell | 12 melanocyte | 15 monocyte | 20 exocrine cell | 23 natural killer cell | 28 extraembryonic cell | 29 cell of skeletal muscle | 29 germ line cell | 33 bone cell | 39 glial cell | 41 blood cell | 42 stromal cell | 52 macrophage | 55 smooth muscle cell | 59 retinal cell | 74 dendritic cell | 77 fibroblast | 83 endocrine cell | 89 B cell | 94 stem cell | 103 T cell | 156 neuron | 493 epithelial cell | 721 More work can be done to improve it, but it might be a good first set. Would you be happy to use this subset instead of the previous more general one @dosumis ?
2025-04-01T04:34:57.504582
2016-04-02T18:56:27
145404377
{ "authors": [ "3rd-Eden", "glen-nicol", "jonalvarezz", "tcboles", "valorize" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9220", "repo": "observing/pre-commit", "url": "https://github.com/observing/pre-commit/issues/66" }
gharchive/issue
Incorrectly reports failure in scripts I started with the following in my package.json "pre-commit": [ "jshint app.js" ], This resulted in the following when I commit: pre-commit: pre-commit: We've failed to pass the specified git pre-commit hooks as the `jshint app.js` pre-commit: hook returned an exit code (1). If you're feeling adventurous you can pre-commit: skip the git pre-commit hooks by adding the following flags to your commit: pre-commit: pre-commit: git commit -n (or --no-verify) pre-commit: pre-commit: This is ill-advised since the commit is broken. pre-commit: Running jshint from the command line returns 0. So under some expert advice I changed my package.json to "pre-commit": [ "true" ], and it reports the same error as before except it says true returned exit code 1. So something is obviously wrong or misconfigured. Any help to troubleshoot? I am running OS X 10.11.4 and the following: node v5.5.0 npm 3.5.3 pre-commit 1.1.2 same behavior occurs using Fish shell and bash Same here. It happened to me after I upgrade NPM2 to NPM3. Any update? Has anyone found a solution to this. I am having the same problem in NPM 4.3 Same problem with npm 5.3.0. I guess it's rather a problem of the commit hook, or? Is there a test case that consistent reproduces this issue? Sry. I'm now sure that my issue is not related. I use npm via a script which runs npm in a docker image. The commit hook doesn't know about that and tries to execute npm on my host machine (windows). Sorry guys I forgot about this and I don't even remember what repo had the trouble. I might have even deleted it. Should I go ahead and close this issue?
2025-04-01T04:34:57.520148
2023-10-24T19:29:55
1959937816
{ "authors": [ "andrewmwhite", "claremacrae" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9221", "repo": "obsidian-tasks-group/obsidian-tasks", "url": "https://github.com/obsidian-tasks-group/obsidian-tasks/pull/2369" }
gharchive/pull-request
Fix typo in docs (filenameWithoutExtension) Description The docs refer to task.file.filename.filenameWithoutExtension, which is undefined. The correct path is task.file.filenameWithoutExtension. Motivation and Context n/a How has this been tested? Locally. Screenshots (if appropriate) Types of changes Changes visible to users: [ ] Bug fix (prefix: fix - non-breaking change which fixes an issue) [ ] New feature (prefix: feat - non-breaking change which adds functionality) [ ] Breaking change (prefix: feat!! or fix!! - fix or feature that would cause existing functionality to not work as expected) [x] Documentation (prefix: docs - improvements to any documentation content for users) [ ] Sample vault (prefix: vault - improvements to the Tasks-Demo sample vault) [ ] Contributing Guidelines (prefix: contrib - any improvements to documentation content for contributors - see Contributing to Tasks) Internal changes: [ ] Refactor (prefix: refactor - non-breaking change which only improves the design or structure of existing code, and making no changes to its external behaviour) [ ] Tests (prefix: test - additions and improvements to unit tests and the smoke tests) [ ] Infrastructure (prefix: chore - examples include GitHub Actions, issue templates) Checklist [x] My code follows the code style of this project and passes yarn run lint. [x] My change requires a change to the documentation. [x] I have updated the documentation accordingly. [x] My change has adequate Unit Test coverage. Terms [x] My contribution follow this project's contributing guide [x] I agree to follow this project's Code of Conduct Good spot - and thank you very much for the fix... This is interesting to me on many levels... tl;dr I will accept the change - thanks! The text is definitely a typo - but surprisingly it does actually behave as documented... After I've merged the PR, I'll go ahead and update the bit of the Tasks automated tests that generates this text, to prevent the error coming back again next time I update the docs... 1. The documentation is machine-generated from the automated tests.. This is one of a lot of pieces in the Tasks documentation that is machine-generated and automatically tested. Your correction is inside these lines - see the include: on the first line, and the endInclude on the last.. https://github.com/obsidian-tasks-group/obsidian-tasks/blob/dacaf099d00431f812c4a61799d99abe503e8025/docs/Queries/Grouping.md?plain=1#L833-L847 And here is the bit of the source-code that contains the data that generates the text that later gets inserted in the docs.... https://github.com/obsidian-tasks-group/obsidian-tasks/blob/dacaf099d00431f812c4a61799d99abe503e8025/tests/Scripting/ScriptingReference/CustomGrouping/CustomGroupingExamples.test.ts#L266-L276 2. The output of the tests looks good... That test, run on a selection of tasks, generates this output: https://github.com/obsidian-tasks-group/obsidian-tasks/blob/dacaf099d00431f812c4a61799d99abe503e8025/tests/Scripting/ScriptingReference/CustomGrouping/CustomGroupingExamples.test.file_properties_task.file.filename_results.approved.txt#L1-L23 So yes you're definitely correct that the text is not what was intended... And also, looking at the files generated by running that test, is shows that the code does actually work (surprisingly...) 3. Oh the irony... I've given a few conference lightning talks on the technique I'm using... Code Samples that Actually Compile Code Samples That Compile Even More Easily - Clare Macrae - CppCon 2020 So yes, the tests work and the sample in the documentation works - but it is also still not 'correct'! Irony on irony.... And also, looking at the files generated by running that test, is shows that the code does actually work (surprisingly...) Having updated the source code I see the changed output... I was looking past all the undefined... The output was clearly wrong. I will now need to search through the rest of the test output and make sure there are no more undefined that crept through! 😊
2025-04-01T04:34:57.711761
2024-08-18T03:39:54
2471771889
{ "authors": [ "LuckyPickleZZ", "PeachThinking" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9222", "repo": "oceanbase/odc", "url": "https://github.com/oceanbase/odc/issues/3134" }
gharchive/issue
[Bug]: export and get table detail would fail if there are pseudocolumns in oracle table ODC version 4.x OB version oracle What happened? export would fail if there are pseudocolumns in oracle table What did you expect to happen? How can we reproduce it (as minimally and precisely as possible)? CREATE TABLE stats_test ( "X" NUMBER(*,0), "Y" NUMBER(*,0), "Z" CHAR(2 CHAR) ); SELECT DBMS_STATS.CREATE_EXTENDED_STATS('JINGTIAN2', 'STATS_TEST', '(X, Y)') FROM DUAL; create a table with pseudocolumns. you could refer https://www.askmac.cn/archives/oracle-virtual-column.html Anything else we need to know? No response p r:https://github.com/oceanbase/odc/pull/4026
2025-04-01T04:34:57.714475
2023-09-12T12:22:13
1892402820
{ "authors": [ "LuckyPickleZZ", "yhilmare" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9223", "repo": "oceanbase/odc", "url": "https://github.com/oceanbase/odc/pull/317" }
gharchive/pull-request
refactor(data-transfer): transforming the Data-Transfer module to support multiple transfer tools What type of PR is this? type-feature module-data-transfer What this PR does / why we need it: This PR is a pre modification of the import and export module to support MySQL. The main work is to refactor the code structure of the data transfer module, enabling it to support various export strategies including ob loader dumper, datax, and so on. your code is not formatted, cause a check item failed: Closed.
2025-04-01T04:34:57.719403
2020-10-26T17:59:00
729784905
{ "authors": [ "SanjoDeundiak", "mrinalwadhwa" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9224", "repo": "ockam-network/ockam", "url": "https://github.com/ockam-network/ockam/issues/555" }
gharchive/issue
Rust default vault encryption test panics occasionally Sometimes test run fails thread 'software::tests::encryption' panicked at 'assertion failed: res.is_err()', vault/src/software/mod.rs:815:9 test software::tests::encryption ... FAILED We saw this again https://github.com/ockam-network/ockam/runs/1779559344#step:8:646 We saw this again https://github.com/ockam-network/ockam/runs/1779559344#step:8:646 Log of the failed run logs_5272.zip Log of the failed run logs_5272.zip Fixed
2025-04-01T04:34:57.794077
2020-02-03T08:52:12
558934065
{ "authors": [ "Holger-Will", "Smilebags", "Xwhat" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9225", "repo": "octalmage/robotjs", "url": "https://github.com/octalmage/robotjs/issues/545" }
gharchive/issue
colors mixed up with robot.screen.capture the colors in the image returned by robot.screen.capture are mixed up. The Bytes returned are in reversed order (little-endian?). and alpha is always 0 (fully transparent), should be 0xff by default. so, for example, the color #112233 becomes <Buffer 33 22 11 00, this behavior prevents using tools like image-encode or image-compare Expected Behavior for the color #112233 robot.screen.capture should return <Buffer 11 22 33 ff Current Behavior instead, it returns <Buffer 33 22 11 00 Possible Solution not sure if this behavior is intended. on big-endian systems, the bitmap returned should also be big-endian (or vice versa) Steps to Reproduce (for bugs) let bmp = robot.screen.capture(x, y, 1, 1) console.log(bmp) Context playing around with image recognition, I'm trying to compare images from screen capture to existing images using image-compare. currently, I'm fixing the bitmap like this: for (var i = 0; i < bmp.image.length; i += 4) { let r = bmp.image[i] let b = bmp.image[i + 2] let a = bmp.image[i + 3] bmp.image[i] = b bmp.image[i + 2] = r bmp.image[i + 3] = 255 } Your Environment RobotJS version: 0.6.0 Node.js version: 13.5.0 npm version: 6.13.4 Operating System: ubuntu 18.10 In the mean time if you want to just fix the colors, just do this for (var i = 0; i < captured.image.length; i += 4) { let r = captured.image[i] let g = captured.image[i + 1] let b = captured.image[i + 2] let a = captured.image[i + 3] captured.image[i] = b captured.image[i + 1] = g captured.image[i + 2] = r captured.image[i + 3] = 255 } let g = captured.image[i + 1] and captured.image[i + 1] = g is redundant... you are setting g to g ;-) Has this got a fix? Even if there's no support from the OS, being able to specify a channel order in the function and having the conversion done in a faster language than JS would be great. Something like screen.capture([x], [y], [width], [height], [channelOrder]) where channelOrder could be'RGBA'. I'd love to be able to screen-capture and send the image buffer directly over the network into an HTML Canvas (which is all working apart from the channel order).
2025-04-01T04:34:57.847182
2020-04-21T00:16:41
603615758
{ "authors": [ "sarahmonster", "tofumatt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9226", "repo": "octopusthink/kelpie", "url": "https://github.com/octopusthink/kelpie/pull/6" }
gharchive/pull-request
feat: Add live reload for Sass Sorry for the delay; forgot about this one! ES6 JS is coming next! ❀️❀️❀️❀️❀️❀️❀️❀️❀️❀️❀️
2025-04-01T04:34:57.849438
2018-02-20T16:19:26
298656280
{ "authors": [ "dfaust", "octplane", "passcod", "phaazon" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9227", "repo": "octplane/fsevent-rust", "url": "https://github.com/octplane/fsevent-rust/issues/13" }
gharchive/issue
Latest update breaks downstream crates It seems you released a breaking change without bumping to a new major version. This is causing problems downstream, see: passcod/notify#153 Could you yank the latest version on crates.io and release a new major version? Indeed, it is a breaking change... I have yanked both crates. Thank you <3 Great, thank you. Thank you all :)
2025-04-01T04:34:57.851267
2024-07-15T12:22:43
2408607692
{ "authors": [ "AntiLibrary5", "wang-ps" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9228", "repo": "octree-nn/octformer", "url": "https://github.com/octree-nn/octformer/issues/29" }
gharchive/issue
Request to add instructions to work with a new dataset Hi, I ran into this work recently and really liked the writing and visuals. Great to see the reproducibility stamp too on the codebase. We've been trying to use OctFormer on a dataset for 3d medical segmentation. So far I am trying to scratch my way around to have the dataset as close to scannet which could work. But, it would be great to the community if generic instructions were added to be able to work with new datasets. Kindly take it under consideration. Thanks, Best. Thanks for your great suggestion! I will try to add instructions in the future. And your solution of following the scannet dataset is practical and smart.
2025-04-01T04:34:57.873069
2019-05-28T04:13:38
449056493
{ "authors": [ "DennisAlund", "anandaprabawa" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9230", "repo": "oddbit/tanam", "url": "https://github.com/oddbit/tanam/pull/212" }
gharchive/pull-request
Feature/linting hooks Adding husky to do git hooks for linting before commit. If fails, it doesn't do file commits. Good idea
2025-04-01T04:34:57.875115
2016-04-26T00:41:37
151014228
{ "authors": [ "jonno84", "odeke-em" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9231", "repo": "odeke-em/drive", "url": "https://github.com/odeke-em/drive/issues/628" }
gharchive/issue
No way to turn off checksumming? I've got a remote drive I mount via SSHFS with large files (several GB) that is then pushed to my Google Drive. Trying to drive push with small, incremental backup will force MD5 check against the existing files on the remote drive - I was able to disable this behaviour in previous releases with the -ignore-checksum option but it appears this option has been removed in favour of an automatic checksum if the file size is the same. With the scarce resources due to SSHFS I would request reintroduction of the -ignore-checksum flag. Thanks! Hello @jonno84, thanks for reporting this and welcome to drive. For starters, what does drive version say? Ping! Please get the latest code, --ignore-checksum has been in for over a year.
2025-04-01T04:34:58.086881
2022-08-25T01:47:37
1350208671
{ "authors": [ "coveralls", "irainia" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9232", "repo": "odpf/optimus", "url": "https://github.com/odpf/optimus/pull/538" }
gharchive/pull-request
feat: add migration implementation This PR is to address this issue. In the implementation, there are some minor differences compared to the mentioned approach. Differences: some additional checks are done, such as the version of Optimus that does migration with the currently active one at some states, migration can be skipped, such as when there's no changes on Optimus version on migration Rollback, if the previous version is not found, it's not returning error, but only a warning changes the table that will store the migration version from migration_versions into migration_steps for more clarity and to avoid conflict name during its implementation code add record removal for migration_steps when migration Rollback is successful Note: additional detail may be added later Pull Request Test Coverage Report for Build<PHONE_NUMBER> 0 of 0 changed or added relevant lines in 0 files are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 75.813% Totals Change from base Build<PHONE_NUMBER>: 0.0% Covered Lines: 7413 Relevant Lines: 9778 πŸ’› - Coveralls
2025-04-01T04:34:58.090473
2019-08-06T12:58:11
477365490
{ "authors": [ "popa-raluca" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9233", "repo": "odpi/egeria", "url": "https://github.com/odpi/egeria/pull/1341" }
gharchive/pull-request
Data engine - update process creation to set the status to draft/active Create processes with DRAFT status and update them to ACTIVE after all the entities and relationships inside the process payload are created Extend RepositoryHandler to allow update of entity status Throw NoSchemaAttributeException when a schema attribute can not be retrieved based on the qualifiedName Only question I have is that I see we're doing parallel processing of the processes that are being sent through... Since within the payload of a given process we may include PortImplementation and PortAlias objects, it's expected that one process's PortAlias may refer to another's PortImplementation: if we're loading them all in parallel, how can we ensure that the lowest-level one (with the PortImplementation) is fully created before attempting to create the higher-level one (with a PortAlias that refers to the PortImplementation of the lower-level process)? Hi Chris, For now, I only process the responses in parallel, after the all processes have been created. The processes are being created in the method createProcesses(String userId, String serverName, List processes) - which does not create them in parallel yet. I'm working on this part now, I'll do a new pull request these days.
2025-04-01T04:34:58.093218
2024-03-15T14:38:18
2188681573
{ "authors": [ "kenconnor", "liborw", "samuelsadok" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9234", "repo": "odriverobotics/ros_odrive", "url": "https://github.com/odriverobotics/ros_odrive/pull/14" }
gharchive/pull-request
Add service to clear errors. Service that sends Clear_Errors command, fixes #13 and #9. Usage: $ ros2 service call /odrive_axis0/clear_errors odrive_can/srv/ClearErrors "identify: 0" @samuelsadok , @liborw Could you give me the permission to push so I can take over this PR? I think I have fixed the conflict and I would like you to review my commits. Thanks for taking over. This PR seems to be abandoned, so I would recommend to just open a new PR from your own fork and we can close this one when done. I see you already did that but then closed it again for some reason? (Btw quick comment from a preliminary look: I saw quite a bit of formatting noise (whitespace changes). It's better to keep that out of the PR, or if your IDE insists, put it in a separate commit at least.) @samuelsadok I closed #25 because it is better to use existing PR (and I did not use merge, making logs messy). And now I opened new PR #27 . I would appreciate it if you take a look at it.
2025-04-01T04:34:58.140568
2021-07-06T13:20:47
937892413
{ "authors": [ "AndreaCimminoArriaga" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9235", "repo": "oeg-upm/wot-hive", "url": "https://github.com/oeg-upm/wot-hive/issues/2" }
gharchive/issue
Improve listing -Add sort_by & sort_order flags Update etag This flags have become optional, and thus, they will not be implemented
2025-04-01T04:34:58.185248
2022-04-18T03:37:02
1206620809
{ "authors": [ "zhankangbao" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9236", "repo": "ofabry/go-callvis", "url": "https://github.com/ofabry/go-callvis/issues/114" }
gharchive/issue
could not import C (no metadata for C) I got this error while scanning some package /Users/kangbao.zhan/code/service/external/github.com/confluentinc/confluent-kafka-go/kafka/00version.go:48:8: could not import C (no metadata for C) fixed after install librdkafka
2025-04-01T04:34:58.225654
2015-11-03T23:56:54
114938272
{ "authors": [ "offirgolan" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9237", "repo": "offirgolan/ember-cp-validations", "url": "https://github.com/offirgolan/ember-cp-validations/pull/65" }
gharchive/pull-request
Handle ember-data promise proxy based instances in presence Regarding #63 This is ready for review but Im not really sure if it's the best way to handle this. Suggestions are definitely welcome. Using detectInstance seemed to get the job done :smile: @stefanpenner good to merge? alright made the final changes. Feel free to merge.
2025-04-01T04:34:58.266436
2022-10-28T06:59:19
1426784668
{ "authors": [ "ohah", "yangyeongwon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9238", "repo": "ohah/vue3-tui-grid", "url": "https://github.com/ohah/vue3-tui-grid/issues/5" }
gharchive/issue
Error when retrieving data of focused row Hi, An error occurs when calling the 'getRow' function to get the data of the focused row #version vue 3.2.27 tui-grid 4.21.2 my code const GridTable = ref(); let grid = ref(); let instance = ref(); onMounted(() => { grid = GridTable.value; instance = grid.gridInstance; }); const handleEdit = () => { const {rowKey} = instance.getFocusedCell(); console.log(instance.getRow(0)); } error thank you fix over the weekend Hi, Is the problem resolved? @yangyeongwon fixed v0.151
2025-04-01T04:34:58.293554
2016-05-30T12:07:28
157490391
{ "authors": [ "amitu", "ohbarye" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9239", "repo": "ohbarye/markdown-server", "url": "https://github.com/ohbarye/markdown-server/issues/1" }
gharchive/issue
No such file or directory: 'markdownserver/resources/css/github.css' After a fresh install. markdown-server==0.1.1 Python 2.7.11 Traceback (most recent call last): File "/Users/amitu/.pyenv/versions/home/bin/mdsvr", line 7, in <module> from markdownserver import main File "/Users/amitu/.pyenv/versions/home/lib/python2.7/site-packages/markdownserver/__init__.py", line 5, in <module> converter = MarkdownConverter() File "/Users/amitu/.pyenv/versions/home/lib/python2.7/site-packages/markdownserver/markdown_converter.py", line 9, in __init__ css = codecs.open(css_root + css_name,encoding=ms_encoding,mode='r') File "/Users/amitu/.pyenv/versions/home/lib/python2.7/codecs.py", line 896, in open file = __builtin__.open(filename, mode, buffering) IOError: [Errno 2] No such file or directory: 'markdownserver/resources/css/github.css' Ah, I assume that I'm dealing in those paths with wrong way... I will investigate. @amitu I fixed the bug and released 0.1.3. Now, I think it would work well with the following command. (I confirmed only on my end, though.) $ pip install markdown-server $ markdownserver $ open http://localhost:8009/sample.md I'm very sorry for my late reply, but I truly appreciate you that you created this issue! Thanks for the release. Not fixed :-( Traceback (most recent call last): File "/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/bin/bottle.py", line 862, in _handle return route.call(**args) File "/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/bin/bottle.py", line 1732, in wrapper rv = callback(*a, **ka) File "/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/lib/python2.7/site-packages/markdownserver/__init__.py", line 11, in gfmize html_file_name = os.path.basename(converter.convert(resource)) File "/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/lib/python2.7/site-packages/markdownserver/markdown_converter.py", line 31, in convert code = md.markdown(self.read_md(src), extensions=[markdown_type]) File "/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/lib/python2.7/site-packages/markdownserver/markdown_converter.py", line 35, in read_md md_file = codecs.open(os.path.join(markdown_dir, file_name),encoding=ms_encoding,mode='r') File "/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/lib/python2.7/codecs.py", line 896, in open file = __builtin__.open(filename, mode, buffering) IOError: [Errno 2] No such file or directory: '/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/lib/python2.7/site-packages/markdownserver/resources/markdown/foo.md' ` Does your markdown file which you want to convert locate the following path? '/Users/amitu/.pyenv/versions/2.7.11/envs/burstic/lib/python2.7/site-packages/markdownserver/resources/markdown/foo.md' No. Its in current folder, ~/projects/burstic/foo.md.
2025-04-01T04:34:58.307065
2024-11-09T08:28:09
2645793542
{ "authors": [ "AlgoFoe", "Jacobjeevan", "modamaan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9240", "repo": "ohcnetwork/care_fe", "url": "https://github.com/ohcnetwork/care_fe/issues/9071" }
gharchive/issue
Warranty Card UI: Improve Responsiveness and Text Overflow Handling for Multi-Language Support Describe the bug The warranty card UI encounters issues with layout and text overflow when displayed in different languages or on smaller screens. When certain languages are selected, text length increases, causing overflow outside the container. Additionally, the layout lacks proper responsiveness, leading to overlapping or misaligned elements on smaller screens. To Reproduce Steps to reproduce the behavior: In the assets section, select any asset. Change the language to one with longer text, such as Tamil. View the warranty card on various screen sizes, particularly smaller screens. Notice any text overflow and layout issues. Expected behavior The warranty card should adjust to accommodate different languages and screen sizes. Text should wrap properly, and elements should align responsively without overflow or overlapping. Screenshots https://github.com/user-attachments/assets/576bc706-c8b8-44a9-ab6d-4b0ad52c56fe Desktop (please complete the following information): OS: Windows 11 Browser: Chrome Version: 130 Smartphone (please complete the following information): Device: iPhone XR Browser: Chrome Version: 130 Additional context Further work is needed to ensure the warranty card UI is fully responsive, with text wrap and layout adjustments for improved language support and responsiveness across different screen sizes. @Jacobjeevan can u assign me this issue, so I can start working on it. @AlgoFoe Any updates with this? @Jacobjeevan i would like to take it up
2025-04-01T04:34:58.309250
2018-01-24T23:25:13
291397044
{ "authors": [ "melor", "nisc-acooper" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9241", "repo": "ohmu/pghoard", "url": "https://github.com/ohmu/pghoard/issues/248" }
gharchive/issue
1.6.0 on pypi Do you intend to release 1.6.0 to pypi? Latest version now available at https://pypi.python.org/pypi/pghoard/1.7.0
2025-04-01T04:34:58.327130
2017-01-15T14:05:10
200870575
{ "authors": [ "ohoservices" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9243", "repo": "ohoservices/remoteNodeMcu", "url": "https://github.com/ohoservices/remoteNodeMcu/issues/3" }
gharchive/issue
configurable http location for lua files updates might be held in credentials.lua Wip done v0.3 (in settings.lua)
2025-04-01T04:34:58.332420
2022-10-16T06:29:57
1410390071
{ "authors": [ "QinomeD", "oierbravo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9244", "repo": "oierbravo/createsifter", "url": "https://github.com/oierbravo/createsifter/issues/5" }
gharchive/issue
Deployer (important for balance) Meshes can be used in Deployers and will sift items, making actual Sifter obsolete. Why? Using Deployers here are sifting items at insane speeds when on max (256) rpm. Thanks for the report! Theres no specific reason, it's a side effect. I'll look to fix it. It was inactive for 10 months you dum dum bot xd
2025-04-01T04:34:58.344593
2023-10-04T14:10:13
1926308785
{ "authors": [ "vrutkovs" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9246", "repo": "okd-project/okd", "url": "https://github.com/okd-project/okd/pull/1754" }
gharchive/pull-request
FAQ.md: update upgrade section Add tips about unavailable upgrades cc @LorbusChris Right, fixed it in https://github.com/okd-project/okd/commit/e188aa37866d3f7036ba3a2ce2044bbff1f1c10f
2025-04-01T04:34:58.410422
2016-01-10T11:44:08
125815058
{ "authors": [ "Stephen-Gates", "pwalsh" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9247", "repo": "okfn/datapackagist", "url": "https://github.com/okfn/datapackagist/issues/204" }
gharchive/issue
Unable to select licence object after name entered When the edit data package first appears you can choose if a licence is a object (i.e. name and url) or a string (name only e.g. CC BY). As soon as you type in a data package name, this option disappears Further, if you select a licence object, you can't revert to a licence name if you change your mind. thanks. FIXED.
2025-04-01T04:34:58.417274
2022-02-20T23:03:53
1145160595
{ "authors": [ "giannif", "iiroj", "xcambar" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9248", "repo": "okonet/lint-staged", "url": "https://github.com/okonet/lint-staged/issues/1109" }
gharchive/issue
monorepo and empty config at root Description I am working on a monorepo where all the lint-staged config are in the packages (or workspaces, depending or your preferred nomenclature). In other words, the package.json at the root of the repo does not contain any config for lint-staged. With the current implementation, lint-staged throws an error if an empty config is found. See: https://github.com/okonet/lint-staged/blob/2604ac7dbca20194b6b8a417335166ccf311a769/lib/validateConfig.js#L50-L52 I've had to add an empty, never matched rule to be able to use lint-staged. I'd like to understand what are my alternatives here and open the discussion whether this check above is really necessary in lint-staged codebase. Steps to reproduce In package.json, use no lint-staged config or an empty config. Error code is always 1. Add any rule, error code is as expected. Debug Logs expand to view ``` lint-staged:bin Running<EMAIL_ADDRESS>+0ms lint-staged:bin Options parsed from command-line: { allowEmpty: false, concurrent: true, configPath: undefined, cwd: undefined, debug: true, maxArgLength: 65536, quiet: false, relative: false, shell: false, stash: true, verbose: true } +1ms lint-staged:validateOptions Validating options... +0ms lint-staged:validateOptions Validated options! +0ms lint-staged Unset GIT_LITERAL_PATHSPECS (was `undefined`) +0ms lint-staged:runAll Running all linter scripts... +0ms lint-staged:runAll Using working directory `/REDACTED` +0ms lint-staged:resolveGitRepo Resolving git repo from `/REDACTED` +0ms lint-staged:resolveGitRepo Unset GIT_DIR (was `undefined`) +0ms lint-staged:resolveGitRepo Unset GIT_WORK_TREE (was `undefined`) +0ms lint-staged:execGit Running git command [ 'rev-parse', '--show-prefix' ] +0ms lint-staged:resolveGitRepo Resolved git directory to be `/REDACTED` +18ms lint-staged:resolveGitRepo Resolved git config directory to be `/REDACTED/.git` +0ms lint-staged:execGit Running git command [ 'log', '-1' ] +18ms lint-staged:execGit Running git command [ 'diff', '--staged', '--diff-filter=ACMR', '--name-only', '-z' ] +7ms lint-staged:runAll Loaded list of staged files in git: lint-staged:runAll [ lint-staged:runAll '/REDACTED/.husky/pre-commit', lint-staged:runAll '/REDACTED/README.md', lint-staged:runAll '/REDACTED/package-lock.json', lint-staged:runAll '/REDACTED/package.json' lint-staged:runAll ] +32ms lint-staged:getConfigGroups Grouping configuration files... +0ms lint-staged:getConfigGroups Grouping staged files by their directories... +0ms lint-staged:getConfigGroups Grouped staged files into 2 directories: +0ms lint-staged:getConfigGroups { lint-staged:getConfigGroups '/REDACTED/.husky': [ '/REDACTED/.husky/pre-commit' ], lint-staged:getConfigGroups '/REDACTED': [ '/REDACTED/README.md', '/REDACTED/package-lock.json', '/REDACTED/package.json' ] lint-staged:getConfigGroups } +2ms lint-staged:getConfigGroups Searching config files... +0ms lint-staged:loadConfig Searching for configuration from `/REDACTED`... +0ms lint-staged:loadConfig Successfully loaded config from `/REDACTED/package.json`: lint-staged:loadConfig {} +2ms lint-staged:getConfigGroups Found new config "/REDACTED/package.json" from "/REDACTED"! +2ms lint-staged:validateConfig Validating config from `/REDACTED/package.json`... +0ms ✘ exit code: 1 </details> ### Environment <!-- Tell us about your development environment --> - **OS:** Ubuntu 20.04 - **Node.js:** v16.13.2 - **`lint-staged`:** ^12.3.4 Hey, this will be fixed by https://github.com/okonet/lint-staged/pull/1106 Version 12.3.5 is now released and hopefully fixed this issue. Can you confirm? @iiroj this approach of an empty root lint-staged config really helped me set up eslint in my monorepo. Each packages runs its own eslint command. Now it seems obvious, but maybe it can be mentioned in the monorepo section that putting things all in the root can be tricky Added in https://github.com/okonet/lint-staged/pull/1200/commits/8ef274720baf70efccb6938b588b6d6b695d480f
2025-04-01T04:34:58.421197
2016-10-07T13:07:39
181664277
{ "authors": [ "TheWolfNL", "ai", "okonet" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9249", "repo": "okonet/lint-staged", "url": "https://github.com/okonet/lint-staged/pull/65" }
gharchive/pull-request
Running Lint-staged from within a sub directory I've added the possibility to set the git-root which if set will be used to resolve the filepaths accordingly this should fix: #41 Awesome! Thank you! I'll ask @ai to test with his setup. Wow! Great work. I will try it in Wednesday (moving to next city :() It doesn’t work :(. I tested lint-staged from master: ➀ git status On branch master Your branch is up-to-date with 'origin/master'. Changes to be committed: (use "git reset HEAD <file>..." to unstage) modified: blocks/intro-notices/intro-notices.js Changes not staged for commit: (use "git add <file>..." to update what will be committed) (use "git checkout -- <file>..." to discard changes in working directory) modified: package.json ➀ git commit -m 'Test' > amplifr@ precommit /home/ai/Dev/amplifr/front > lint-staged Error: fatal: Not a git repository: '.git' at ChildProcess.<anonymous> (/home/ai/Dev/amplifr/front/node_modules/staged-git-files/index.js:85:19) at emitTwo (events.js:106:13) at ChildProcess.emit (events.js:191:7) at maybeClose (internal/child_process.js:877:16) at Socket.<anonymous> (internal/child_process.js:334:11) at emitOne (events.js:96:13) at Socket.emit (events.js:188:7) at Pipe._handle.close [as _onclose] (net.js:493:12) /home/ai/Dev/amplifr/front/node_modules/lint-staged/src/index.js:27 const filePaths = results.map(file => file.filename) ^ TypeError: Cannot read property 'map' of undefined at /home/ai/Dev/amplifr/front/node_modules/lint-staged/src/index.js:27:30 at /home/ai/Dev/amplifr/front/node_modules/staged-git-files/index.js:13:13 at /home/ai/Dev/amplifr/front/node_modules/staged-git-files/index.js:38:13 at ChildProcess.<anonymous> (/home/ai/Dev/amplifr/front/node_modules/staged-git-files/index.js:88:9) at emitTwo (events.js:106:13) at ChildProcess.emit (events.js:191:7) at maybeClose (internal/child_process.js:877:16) at Socket.<anonymous> (internal/child_process.js:334:11) at emitOne (events.js:96:13) at Socket.emit (events.js:188:7) Ouh, I forgot about "git-root": "../", :) Everything works right now. Waiting for release! I've renamed the options to gitDir for consistency with git and coding standard. Released as 3.1.0: https://github.com/okonet/lint-staged/releases/tag/3.1.0
2025-04-01T04:34:58.464859
2021-12-06T07:48:52
1071831708
{ "authors": [ "bryanapellanes-okta", "wlumetsberger" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9250", "repo": "okta/samples-blazor", "url": "https://github.com/okta/samples-blazor/issues/11" }
gharchive/issue
Authentication Loop in IFrame If the Balzor-Seite is embedded in an IFrame (Microsoft Teams WebsiteTab) the Authentication ends up in a Loop. Teams uses also Federated Authentication with Okta. Using the Site Standalone the Authentication Works as aspect. AccountController is called -> Challange is Called -> Okta Login does appear OnTokenValidateEvent gets fired containing the authenticated Userprincipal in args.Identity AccountController is called -> HttpContext.User is authenticated Using the same Site in a Team Websitetab the steps are the following: AccountController is called -> Challenge is Called -> no Okta Login appears (Token is already here) OnTokenValidateEvent gets fired containing the authenticated Userprincipal AccountController is called -> but now User is not authenticated! --> Loop This is the Configuration used: Issue is also during use in a standard IFrame: @wlumetsberger, Thanks for reaching out to report this! To help us reproduce and investigate the issue further can you share a project/solution that reproduces the issue as you describe? I have added an item for internal tracking for further review. @bryanapellanes-okta here is a link to the repo: https://github.com/wlumetsberger/okta-blazor-server-side-tests
2025-04-01T04:34:58.487378
2024-11-03T12:00:48
2631167807
{ "authors": [ "afkcodes", "bacarybruno", "gunnartorfis", "okwasniewski" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9251", "repo": "okwasniewski/react-native-bottom-tabs", "url": "https://github.com/okwasniewski/react-native-bottom-tabs/issues/114" }
gharchive/issue
Lucide / SVG icons Just noticed that documentation has been added for vector icons. Is it possible to use Lucide/SVG icons? +1 for this, specially because expo/vector-icons doesn't expose / support getImageSource. Hey, Thanks for opening the issue. At this point it's not possible. You can't pass custom React Native component into the tab bar because of native limitations so this won't be supported (Lucide only supports icon components). Regarding SVG Icons I'm looking into adding support for this but at this point Im using same functionality that React Native Core is using to load Images in the component, handing SVGs would require going custom. Regarding expo/vector-icons, you can read more about this library here: https://twitter.com/notbrent/status/1851718411128041714 (It's going to be deprecated soon). Most of the icons for the tab bar should be static so if possible I would suggest getting them as static assets and using them as usual the svgs look blurry if supplied externally @okwasniewski did you face this ? @afkcodes There shouldn't be any degradation if they are external - can you open an issue with reproduction?
2025-04-01T04:34:58.497069
2023-12-18T07:12:52
2045886787
{ "authors": [ "h3th-IV", "inidaname", "olakunlevpn", "sudophils" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9252", "repo": "olakunlevpn/Nigeria-Bank-Account-Number-API", "url": "https://github.com/olakunlevpn/Nigeria-Bank-Account-Number-API/issues/4" }
gharchive/issue
Is this endpoint still valid? It returns 404 Kindly take a look at it, https://nubaoi.com/verify or the one with code same error copied from your documentation, but the test on your try it works, which endpoint is that one pointing to Hey there, could you kindly take a look at this documentation on how to get started? You can find it at https://maylancer.org/docs/nuban/2.0/introduction/getting-started. Tried the endpoints in the tutorial and got the error above This page isn’t workingnubapi.com redirected you too many times. [Try deleting your cookies](https://support.google.com/chrome?p=rl_error&hl=en-GB). ERR_TOO_MANY_REDIRECTS Hey there, could you kindly take a look at this documentation on how to get started? You can find it at https://maylancer.org/docs/nuban/2.0/introduction/getting-started. Good day, the link(s) from the Documentation link don't seem to be working no more https://nubapi.com/register returns... " Secure Connection Failed An error occurred during a connection to nubapi.com. Peer reports it experienced an internal error. Error code: SSL_ERROR_INTERNAL_ERROR_ALERT The page you are trying to view cannot be shown because the authenticity of the received data could not be verified. Please contact the website owners to inform them of this problem. " the others too return the same
2025-04-01T04:34:58.528557
2017-10-25T08:20:12
268306670
{ "authors": [ "arpanimage", "oleksiyk" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9253", "repo": "oleksiyk/kafka", "url": "https://github.com/oleksiyk/kafka/issues/210" }
gharchive/issue
NokafkaConnectionError - Connection Timeout If I get a no connection timeout error then does that mean that the request from node server has reached Kafka servers. Is there a possibility that connection timeout can occur when the request has not left node server? Id so what could be the reason for this. It can be both, its an underlying network timeout. It can be TCP socket timeout for example and it might actually happen somewhere on the route from Node server to Kafka server. I suggest you use tcpdump or similar tool to analyse your network connection.
2025-04-01T04:34:58.529915
2015-12-27T03:36:37
123958227
{ "authors": [ "hxgqh", "nmrao" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9254", "repo": "olensmar/soapui-swagger-plugin", "url": "https://github.com/olensmar/soapui-swagger-plugin/issues/27" }
gharchive/issue
POM Dependency Error: ready-api-soapui-pro version not right version 1.2.0 of ready-api-soapui-pro could not be find in repository now. Should change it to be 1.2.2 or higher version. <dependency> <groupId>com.smartbear</groupId> <artifactId>ready-api-soapui-pro</artifactId> <version>1.2.2</version> </dependency> Thank you @hxgqh
2025-04-01T04:34:58.556107
2016-01-06T19:59:30
125250709
{ "authors": [ "olivere", "sundarv85" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9255", "repo": "olivere/elastic", "url": "https://github.com/olivere/elastic/issues/194" }
gharchive/issue
Complicated Query Examples Hello, I’m trying to wrap my head on how to write such the below two querires as below using elastic. It would be helpful if you could write up such examples in the Wiki as well. Thanks in advance. GET /industry/music/_search { "query": { "filtered": { "filter": { "term": { "kind": "rock" } }, "query": { "has_child": { "type": "song", "score_mode": "sum", "query": { "filtered": { "query": { "match_all": {} }, "filter": { "bool": { "must": [ { "term": { "singer": "Michael" } }, { "term": { "group": "no" } } ] }, "range": { "date": { "gte": "1995", "lt": "2000" } } } } } } } } } } Aggregate Example GET /industry/music/_search?search_type=count { "query": { "filtered": { "filter": { "term": { "kind": "rock" } } } }, "aggs": { "events": { "children": { "type": "song" }, "aggs": { "filtered": { "filter": { "range": { "date": { "gte": "1995", "lt": "2000" } } }, "aggs": { "browser": { "terms": { "field": "singer", "size": 15 } } } } } } } } @sundarv85 Yeah. Documentation could be better. Here are some tips that might help. First, read the tests. They serve as example code. Second, once you got the JSON right, build it from the inside out. Third, once you think you got it right and it still fails, enable the trace log and/or use the _.Source() func and serialize to JSON. It should give you an idea what's wrong. Regarding your query, notice that ES has deprecated filtered query as of 2.0 (and it's not in elastic.v3). You should use a bool query instead, as described here. HTH I've set up some documentation regarding the QueryDSL.
2025-04-01T04:34:58.579473
2018-03-28T11:38:44
309330138
{ "authors": [ "autonomobil", "olkal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9256", "repo": "olkal/LCD_ST7032", "url": "https://github.com/olkal/LCD_ST7032/issues/2" }
gharchive/issue
Schematic for connection ? I have also this type of ERC1602FYG-4 display (https://www.buydisplay.com/default/i2c-cog-16x2-character-lcd-display-module-st7032-serial-black-on-yg). How do you connect it to an Arduino Nano with SDA on Pin A4 und SCL on PIN A5? I connected like shown here, but the LCD stays dark: http://www.buydisplay.com/download/interfacing/ERC1602-4_Interfacing.pdf Hi Yes, A4 SDA and A5 SCL. I used the same schematic exept that I tied the RESET pin from the display to VDD. I'm sure you have checked everything but it really should work if the wiring and components is correct and the contrast is within the range.
2025-04-01T04:34:58.586938
2024-01-26T21:22:30
2102921772
{ "authors": [ "giannisak", "pdevine" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9257", "repo": "ollama/ollama", "url": "https://github.com/ollama/ollama/issues/2217" }
gharchive/issue
Message vs Template vs System What is the difference between message, template and system if I want to do few-shot prompting? I mean, I could pass the example of release(v0.1.21) to a model in three different ways: Few-shot using Message: SYSTEM You are a friendly assistant that only answers with 'yes' or 'no' MESSAGE user Is Toronto in Canada? MESSAGE assistant yes (etc..) Few-show using Template: TEMPLATE """ <|im_start|>system {{ .System }} <|im_end|> <|im_start|>user Is Toronto in Canada? <|im_end|> <|im_start|>assistant yes <|im_end|> (etc..) You will be given questions about whether a city is located in a specific country. Example 1: Is Toronto in Canada? yes Example 2: (etc..) """ I am running some tests using llama index in a similar topic on 7B models and I am getting better results in System format compared to Template format (I was expecting the opposite). I will test message format too, but I am trying to understand the differences and the expected behavior of each. Hey @giannisak will work. You can alternatively put in MESSAGE system You are a friendly assistant that only answers with 'yes' or 'no' instead of using SYSTEM. Both ways are supported. won't work, because the template is repeated each time you send a message. The template is supposed to define the format for how data gets transformed into whatever format the model is expecting. will probably work, but not as well as 1. It depends more on the LLM if it can understand what you're trying to pass to it. I wouldn't recommend doing it this way vs. 1. Keep in mind that the MESSAGE commands only work with the /api/chat endpoint and do not work with /api/generate. If there's enough demand, we can look at adding it for /api/generate, but it'll take a lot more effort than it was to make it work with the chat endpoint. Going to close this, but feel free to reopen it.
2025-04-01T04:34:58.589293
2023-08-04T14:14:17
1836825516
{ "authors": [ "Kylejustknows", "jmorganca" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9258", "repo": "ollama/ollama", "url": "https://github.com/ollama/ollama/issues/282" }
gharchive/issue
CORS requests don't work when origin is <IP_ADDRESS> When connecting to localhost:11434 from a host served on <IP_ADDRESS>, requests are blocked due to CORS checks – which are important to have, but we should consider opening them up to <IP_ADDRESS> as well OMG, running Ollama is so easy, but connect to it is so hard and require super deep advanced network/server/html knowledge. There are tons of posts all asking the same thing. I consider myself a pretty good old-school server manager. I spent like 2 hours tuning the env OLLAMA_ORIGINS and the caddy + HTML on a windows server, still getting the CORS. If running a non-standard service (simple handmade etc.) trying to connect to Ollama, it is a nightmare! Could Ollama just remove all the CORS restrict things, just run as a plaint service, let users handle their network port security etc please?
2025-04-01T04:34:58.592665
2024-08-12T02:22:03
2459914511
{ "authors": [ "garyyang85", "jmorganca", "rick-github" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9259", "repo": "ollama/ollama", "url": "https://github.com/ollama/ollama/issues/6316" }
gharchive/issue
ollama create will use a large amount of disk space in the /tmp What is the issue? ollama create cmd will use a large amount of disk space in the /tmp directory by default. Is there a way to change the /tmp to other directory? OS Linux GPU Nvidia CPU Intel Ollama version latest Hi there, this should be fixed in https://github.com/ollama/ollama/releases/tag/v0.3.5. If you'd like to install it ahead of it being fully released you can run: curl -fsSL https://ollama.com/install.sh | OLLAMA_VERSION=0.3.5 sh Hi @jmorganca I want to find a way to use another disk directory instead of /tmp when run ollama create. I have a lot of space in other directories but not /tmp. Can I do this? https://github.com/ollama/ollama/blob/01b80e9ffcd05825b413b61a91f9656723ce8afa/envconfig/config.go#L260
2025-04-01T04:34:58.598404
2024-11-16T05:38:35
2663848463
{ "authors": [ "arbusam", "dkarthicks27", "rick-github" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9260", "repo": "ollama/ollama", "url": "https://github.com/ollama/ollama/issues/7698" }
gharchive/issue
Pulling llama3.1 from ollama results in json error What is the issue? When I try to run the following command: ollama pull llama3.1 I get the following issue: Error: invalid character '}' looking for beginning of object key string. OS: Mac M1 (14.1.1) ollama version: 0.3.12 OS macOS GPU Apple CPU Apple Ollama version 0.3.12 Have you tried updating to the newest version and seeing if this still happens? Have you successfully pulled models before? What's the output of the following command: curl -D - https://registry.ollama.ai/v2/library/llama3.1/manifests/latest Have you tried updating to the newest version and seeing if this still happens? I'm on an M2 Mac mini on a newer version and this doesn't happen to me. I have tried other models, like mistral and gemma, and it does work. Manifest is pulled successfully, I observed this specific to llama3.1 Does this work any better: ollama pull llama3.1:8b-instruct-q4_0 Have you successfully pulled models before? What's the output of the following command: curl -D - https://registry.ollama.ai/v2/library/llama3.1/manifests/latest Hey @rick-github, I find it surprising, as I tried pulling the model, with a different wifi, and this time around it works, without any issues. What could be the underlying issue, as I am not able replicate the same issue. While the other day, it happened, even after updating ollama using pip, and reoccurred multiple times. But today its able to pull. Does this work any better: ollama pull llama3.1:8b-instruct-q4_0 This didn't work the other day, I tried this, but worked with ollama pull llama3.2:3b-instruct-fp16
2025-04-01T04:34:58.616254
2023-03-05T00:03:18
1609993344
{ "authors": [ "adli100", "lprot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9261", "repo": "olli991/mib-std2-pq-zr-toolbox", "url": "https://github.com/olli991/mib-std2-pq-zr-toolbox/issues/237" }
gharchive/issue
navigation map without street shapes dump.zip I have MIB-STD2-PQ Technisat EU SW 254 patched with toolbox, it recognizes the non-original sd card as map and starts navigation but there is no street drawing, it says: "B126C F0 (009)" Navigation Database Faulty dump.zip While the same card works perfectly on another MIB-STD2-PQ Technisat US SW 367 pictures attached Rename \maps\01 to \maps\00 hope this helps https://www.drive2.com/b/640748923375325654/ Rename \maps\01 to \maps\00 hope this helps https://www.drive2.com/b/640748923375325654/ The unit comes from Australia it's basically based on European firmware and the map includes Australia, North Africa etc. As I need Algerian map, so I solved the problem by renaming 27 to 22 Thank you what is the coding of 5F on the first unit and on the second unit? Friday,03,March,2023,22:36:13:60195 VCDS Version: Release 22.10.0 Address 5F: Information Electr. Control Module Part Number: 5C0 035 680 B Component and/or Version: MU-S-N-ER H22 0254 Software Coding: 06330708FF0000001121000100080000AF0100D401200101C6 Work Shop Code: WSC 29588 VCID: 204F93553B28C08DA2-8074 Advanced Identification/FAZIT Serial number: Identification: YD7-002 Date: 01.10.15 Manufacturer number: 0312 Test stand number: 0202 Flash Status Programming Attempts(application): 0/6 Programming Attempts(data): 65535/0/1/1/1/1/65535/65535/65535/1/1/65535/65535/65535/65535/65535/65535/0/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535 Successful Attempts(data): 65535/0/1/1/1/1/65535/65535/65535/1/1/65535/65535/65535/65535/65535/65535/0/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535/65535 Flash Date: 00.01.01 Flash Tool Code(data): 00015 790 00066 Misc. Hardware number: 5C0 035 680 B Workshop System Name: J794 Equipment/PR Code: FFFFFFFFFFFFFFFFFFFFFFFF Dataset Number: 5K7CV1v__A_ Dataset Version: 0001 ASAM Dataset: EV_MUStd4CTSAT ASAM Dataset Revision: 001022 Friday,03,March,2023,22:38:58:60195 VCDS Version: Release 22.10.0 Running on Windows 7 x86 www.Ross-Tech.com Address 5F: Information Electr. Labels: 5G0-035-MIB-STD2.clb Control Module Part Number: 5C0 035 680 B HW: 5C0 035 680 B Component and/or Version: MU-S-N-ER H22 0254 Software Coding: 06330708FF0000001121000100080000AF0100D401200101C6 Work Shop Code: WSC 29588 002 104857 ASAM Dataset: EV_MUStd4CTSAT 001022 (VN35) ROD: EV_MUStd4CTSAT_SE36.rod VCID: 204F93553B28C08DA2-8074 2 Faults Found: 1555 - Check Software Version Management B201A 00 [009] - - [New feature! Extended UDS fault detail is only supported by current gen. interfaces] Confirmed - Tested Since Memory Clear Freeze Frame: Fault Status: 00000001 Fault Priority: 6 Fault Frequency: 1 Reset counter: 186 Mileage: 143272 km Date: 2064.14.14 Time: 14:35:08 Voltage terminal 30: 12.5 V Status_Software_Version_Management-module_name: 15360768 - Component Protection U1101 00 [009] - Active [New feature! Extended UDS fault detail is only supported by current gen. interfaces] Confirmed - Tested Since Memory Clear Freeze Frame: Fault Status: 00000001 Fault Priority: 4 Fault Frequency: 1 Reset counter: 186 Mileage: 143272 km Date: 2064.14.14 Time: 14:35:09 Voltage terminal 30: 12.5 V Friday,03,March,2023,23:07:03:60195 VCDS Version: Release 22.10.0 Running on Windows 7 x86 www.Ross-Tech.com Address 5F: Information Electr. Labels: 5G0-035-MIB-STD2.clb Control Module Part Number: 5C0 035 684 B HW: 5C0 035 684 B Component and/or Version: MU-S-NS-US H31 0367 Software Coding: 06330702FF0000001111000100080000AF0100D401200101C6 Work Shop Code: WSC 12345 123 61029 ASAM Dataset: EV_MUStd4CTSAT 001022 (VN35) ROD: EV_MUStd4CTSAT_SE36.rod VCID: 24479F453730DCADC6-8070 Fault Codes have been Erased 2 Faults Found: 1555 - Check Software Version Management B201A 00 [009] - - [New feature! Extended UDS fault detail is only supported by current gen. interfaces] Confirmed - Tested Since Memory Clear Freeze Frame: Fault Status: 00000001 Fault Priority: 6 Fault Frequency: 1 Reset counter: 186 Mileage: 143272 km Date: 2064.14.14 Time: 15:08:47 Voltage terminal 30: 13.2 V Status_Software_Version_Management-module_name: 15360768 - Component Protection U1101 00 [009] - Active [New feature! Extended UDS fault detail is only supported by current gen. interfaces] Confirmed - Tested Since Memory Clear Freeze Frame: Fault Status: 00000001 Fault Priority: 4 Fault Frequency: 1 Reset counter: 186 Mileage: 143272 km Date: 2064.14.14 Time: 15:08:47 Voltage terminal 30: 13.2 V
2025-04-01T04:34:58.632105
2020-08-10T13:27:21
676136211
{ "authors": [ "olton", "projeffboy" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9262", "repo": "olton/Metro-UI-CSS", "url": "https://github.com/olton/Metro-UI-CSS/issues/1596" }
gharchive/issue
For dropdown element, is there a way to click on the dropped down element and not disappear? Is your feature request related to a problem? Please describe. Basically, I don't want the red box below to disappear when I click on it (only "Drop block" can make it disappear/reappear): Describe alternatives you've considered I tried modifying $('#filters').data('no-close') but it seems to be read-only. I then tried event.preventDefault and return false when using the event handler onUp. Maybe I have to tinker with $('#filters').data('dropdown')? I'm quite new to your framework, so if I'm using it incorrectly please let me know. Thanks for the framework! Use collapse component Ugh, I didn't think outside the box. Thank you.
2025-04-01T04:34:58.685024
2017-12-09T14:36:36
280724555
{ "authors": [ "omenking" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9263", "repo": "omenking/swap-n-pop", "url": "https://github.com/omenking/swap-n-pop/issues/161" }
gharchive/issue
ensure electron distribute works with typescript conversion Since we changed the structure of our app, I don't think electron builder will know what to do. Its also possible since we don't utilize two package json files, the dependencies might not get bundled into the asar. Will find out as I go. This is a high priority ticket since without it fixed we can't release binaries @RauliL we might have to go back to two package json. There are no examples on how to configure electron-builder to use one package.json structure https://github.com/electron-userland/electron-builder/issues/601 I tried to configure as such: "build": { "directories" : { "app" : ".", "buildResources" : "static" }, There is also something suggesting that native modules will not work unless you have a two package json structure. I would like one package.json file, though I would also just like this thing to work. I see this as I build a dist No native production dependencies Maybe this indicates we can get away with just having one package.json. I removed electron-webpack and getting fewer errors but only after I changed back to two package json.
2025-04-01T04:34:58.707574
2019-08-09T12:37:57
478965455
{ "authors": [ "DavidMansolino", "omichel" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9264", "repo": "omichel/webots", "url": "https://github.com/omichel/webots/issues/791" }
gharchive/issue
Linux installation instructions are broken Describe the Bug I tried to follow the Linux installation instructions on a fresh Ubuntu 18.04 VM and it fails badly. Steps to Reproduce Follow the instructions from the user guide to install Webots from apt. When I typed: sudo apt-get install webots I got: Reading package lists... Done Building dependency tree Reading state information... Done E: Unable to locate package webots I can reproduce this on a fresh native Ubuntu 18.04 too :disappointed: When downloading the .deb file and opening it with the default application for .deb file, it displays "License: proprietary" whereas it should display Apache 2.0 / open source instead. Moreover, a screenshot is missing... Partially fixed in #792. It remains the missing screenshot and wrong license to fix. It is not possible to have a screenshot until we get officially accepted in the main ubuntu repos: See https://askubuntu.com/questions/117130/how-to-include-a-screenshot-of-my-application-in-the-deb-ppa-to-show-it-in-softw It is not possible to display a free open source license in the ubuntu software center if our package is not part of the official ubuntu repositories in main or universe sections, see: https://stackoverflow.com/questions/18017084/copyright-file-in-deb-package-how-to I started the procedure to get Webots into the official debian repositories. This is the procedure to follow to get it in ubuntu repositories as well. I just found a debian sponsor who will help us getting things done. I am closing this issue as the resolution of it implies getting Webots into Ubuntu universe which is described in #810.
2025-04-01T04:34:58.726577
2018-11-26T16:09:30
384401000
{ "authors": [ "kfichter", "paulperegud" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9265", "repo": "omisego/plasma-contracts", "url": "https://github.com/omisego/plasma-contracts/pull/42" }
gharchive/pull-request
[WIP] MoreVP flow Please review, but do not merge yet. Done: all the things, branch dev-morevp (by Kelvin) fixing discrepancies between master and dev-morevp ERC20 tokens in transaction format bug which caused mature in-flight exits to be processed after not-yet-mature standard exits TODO: ERC20 handling for in-flight exits high exit priority for deposits (the one and only failing test ATM) bug which caused mature in-flight exits to be processed after not-yet-mature standard exits What caused this? Just curious. Most significant bit of priority uint256 value set to 1 for in-flight exits. This changed the order from one sorted by {exitable_timestamp} to one sorted by {type, exitable_timestamp}. To fix this, bits were changed from: use bits [0-191] as exit id (192 bits) use bits [192-254] as exitable timestamp (63 bits) use most-significant bit [255] as a in-flight flag (1 bit) to: use least-significant bit [0] as a in-flight flag (1 bit) use bits [1-191] as exit id (191 bits) use bits [255-192] as exitable timestamp (64 bits)
2025-04-01T04:34:58.734460
2018-07-11T04:23:15
340086994
{ "authors": [ "Benjamin-L", "omni-viral" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9266", "repo": "omni-viral/xfg-rs", "url": "https://github.com/omni-viral/xfg-rs/issues/10" }
gharchive/issue
ui example uses hardcoded font path The path used for the font in the ui example is hardcoded to somewhere in your home directory. It would probably be better to just put the font in the repository and use a relative path, so that people don't have to go manually change it. Oh. Sure. I didn't mean to commit hardcoded paths 😦
2025-04-01T04:34:58.742861
2023-01-17T22:26:37
1537115445
{ "authors": [ "dvdgomez", "jamesbeedy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9267", "repo": "omnivector-solutions/slurmrestd-operator", "url": "https://github.com/omnivector-solutions/slurmrestd-operator/pull/1" }
gharchive/pull-request
Add initial unit tests and CI pipeline Hey folks, this PR establishes the initial unit tests and CI pipeline for slurmrestd-operator. Summary of Change Add initial templates for bug reports, enhancement requests, and pull requests Add a basic CI pipeline using GitHub actions - Inclusive language check - Lint - A separate PR will cover linting changes, currently linting has not been run yet. Those changes will stem from the results of running tox -e lint. - Unit tests Formatting - Changes to the source code in this PR stem from running tox -e fmt. These changes result from using black and ruff. Unit Tests - These tests only test the typical expected behavior of the charm. Integration and functional tests will follow in later PRs however those are not covered here. Please take a look and ensure these changes are satisfactory and raise any questions, comments, or concerns! Going to merge this now, as all comments are resolved. Thanks @dvdgomez!
2025-04-01T04:34:58.792070
2018-07-27T10:51:25
345181545
{ "authors": [ "maxkachalin", "ondras" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9268", "repo": "ondras/TeaJS", "url": "https://github.com/ondras/TeaJS/issues/12" }
gharchive/issue
Build crashes on ACTION tools_gyp_v8_gyp_v8_snapshot_target_run_mksnapshot with segmentation fault I'm trying to build latest version, process crashes on v8 building with segfault. Versions: Ubuntu 18.04 (Linux 95bcb97072dd 4.9.87-linuxkit-aufs #1 SMP Wed Mar 14 15:12:16 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux) g++ (Ubuntu 7.3.0-16ubuntu3) 7.3.0 root@95bcb97072dd:/TeaJS/deps/v8# make native werror=no PYTHONPATH="/TeaJS/deps/v8/tools/generate_shim_headers:/TeaJS/deps/v8/build::/TeaJS/deps/v8/build/gyp/pylib:" \ GYP_GENERATORS=make \ build/gyp/gyp --generator-output="out" build/all.gyp \ -Ibuild/standalone.gypi --depth=. -S.native -Dv8_enable_backtrace=1 -Dwerror='' -Darm_fpu=default -Darm_float_abi=default make[1]: Entering directory '/TeaJS/deps/v8/out' ACTION tools_gyp_v8_gyp_v8_snapshot_target_run_mksnapshot /TeaJS/deps/v8/out/native/obj.target/v8_snapshot/geni/snapshot.cc Segmentation fault tools/gyp/v8_snapshot.target.native.mk:13: recipe for target '/TeaJS/deps/v8/out/native/obj.target/v8_snapshot/geni/snapshot.cc' failed make[1]: *** [/TeaJS/deps/v8/out/native/obj.target/v8_snapshot/geni/snapshot.cc] Error 139 make[1]: Leaving directory '/TeaJS/deps/v8/out' Makefile:301: recipe for target 'native' failed make: *** [native] Error 2 Hi @maxkachalin, I am sorry to inform you that TeaJS is officially discontinued. I have not updated nor built this project in last ~6 years, so it is kind of expected that it does not work with recent V8. I will archive the project at GitHub as well, soon. Hi @ondras, I see. :( By the way I have been compiling against included in your distro v8 deps.
2025-04-01T04:34:58.793862
2017-04-12T20:05:11
221375157
{ "authors": [ "devshell", "ondras" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9269", "repo": "ondras/wwwsqldesigner", "url": "https://github.com/ondras/wwwsqldesigner/pull/253" }
gharchive/pull-request
Change Serial to Big Serial for IDs This will ensure the ID never exceeds the max size. With Serial, the DB can only handle 2 billion rows, which is not that much for a db. By switching to Big Serial, there is no longer any reason to worry about the limit being reached. All for the small cost of using an additional 4bytes per ID. See this article Thanks!
2025-04-01T04:34:58.806649
2023-07-02T22:06:48
1784940978
{ "authors": [ "ArtemSBulgakov", "danmaninc" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9270", "repo": "one-zero-eight/InNoHassle-Website", "url": "https://github.com/one-zero-eight/InNoHassle-Website/issues/37" }
gharchive/issue
Add more features to the user icon [x] Add the Β«View ProfileΒ» button [x] Add the icons for the buttons That's how it looks now: Now it looks like this:
2025-04-01T04:34:58.810053
2020-08-13T05:32:19
678164069
{ "authors": [ "michael-smirnov" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9271", "repo": "oneapi-src/oneDAL", "url": "https://github.com/oneapi-src/oneDAL/pull/818" }
gharchive/pull-request
Data management folder restructure Changes proposed in this PR: According to recent agreements, data folder shall be divided into more granular pieces (table, io, graph, etc.). At now, only table component exists, so data renamed to table table.hpp divided onto common.hpp (table interface) and homogen.hpp (homogen_table class) accessor.hpp divided onto row_accessor.hpp and column_accessor.hpp array.hpp moved to the root folder type traits for tables moved from utils to table folder and named like the entities which they belong to (common_type_traits, homogen_type_traits) :hourglass: files table_metadata and table_builder were not changed because are planned to change in next PRs /intel-ci: run /intelci: run
2025-04-01T04:34:58.905929
2019-08-14T23:17:24
480928963
{ "authors": [ "Keats", "bovee", "luizirber" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9272", "repo": "onecodex/taxonomy", "url": "https://github.com/onecodex/taxonomy/issues/4" }
gharchive/issue
traverse can be slow Running e.g. prune_to on a large-ish taxonomy (>100k nodes) takes minutes mostly because the traverse method takes a long time. I think this is mostly because we're doing a ton of cloning in this method and malloc is taking a lot of time. I tried writing a impl<'t> Iterator for TaxonomyIterator<'t, usize, f32> ... specialized method, but I keep getting conflicting implementation for 'taxonomy::TaxonomyIterator<'_, usize, f32>' errors. I tried rearranging the clones, but I was also having issues with lifetimes with that. Unrelated: what tool is the screenshot from? @Keats instruments, you can launch it from the command line: https://help.apple.com/instruments/mac/10.0/#/devba105ecc Thanks! It does look like it's missing the number of allocations (temp or not) as well as the memory usage. @bovee I have an Ubuntu VM with heaptrack set up. If you can get me a runnable example, I can send you the profiling for allocations @Keats Thanks! There's a separate mode for allocations; it looks like it's making a quarter million or so transient small ones per second (not super adjusted to the interface to say where they are, but guessing they're copying keys around): I'm running a fairly simple example using one of our database build.json's: import json from taxonomy import Taxonomy tax = Taxonomy.from_json(json.dumps(json.load(open('.../build.json'))['taxonomy']['node_link_data'])) _ = tax.prune(keep=['562']) # this line basically kills the process I think this may also be responsible for how hideously slow iterating over the taxonomy in Python is (although there may be other issues there with how I construct the Python iterator). With debug=True in the release profile of Cargo.toml, heaptrack can tell you exactly which line is allocating, how many allocations and their total/peak values. Maybe instruments does the same, I'll try tomorrow with the proxy. Can I get a build.json from s3? @keats s3://refgenomics-datafiles/dbs/mg_complete_20181218_extra/build.kmerdb.json hmm, but taxonomy doesn't work with the current mgo db? @luizirber That error was fixed in b5794ad9e33065ceff8e4e73f7b3130089dc4c1c; once we get something for this issue we can deploy both fixes as 0.3.2 on pypi, etc. Okay, this turned out to be a bug where tax.children('1') returned e.g. ['1', '2', '10239'] so we were infinitely recursing (every time we cleared the queue root was still there). The actual speed of the code is fine; it takes ~47 ms to prune the taxonomy now.
2025-04-01T04:34:58.910858
2023-02-27T22:21:19
1602045560
{ "authors": [ "NorbertHuethmayr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9274", "repo": "onesine/react-tailwindcss-select", "url": "https://github.com/onesine/react-tailwindcss-select/issues/20" }
gharchive/issue
PostCss Error Hi, when adding the following line to my tailwind.config.js file, my next app no longer compiles. "./node_modules/react-tailwindcss-select/dist/index.esm.js" The error I'm seeing is as follows error - ./node_modules/next/dist/build/webpack/loaders/css-loader/src/index.js??ruleSet[1].rules[2].oneOf[8].use[1]!./node_modules/next/dist/build/webpack/loaders/postcss-loader/src/index.js??ruleSet[1].rules[2].oneOf[8].use[2]!./styles/globals.cssSyntaxError: Unexpected token (8:4) Have you seen this before? This item can be closed. It started working after recompiling...
2025-04-01T04:34:58.915899
2023-12-21T01:16:09
2051514317
{ "authors": [ "Heeesang", "onevcat" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9275", "repo": "onevcat/Kingfisher", "url": "https://github.com/onevcat/Kingfisher/pull/2184" }
gharchive/pull-request
Comment Modification - Adding Space Added space in the comment: "//Flip image one more time if needed to, this is to prevent flipped image" Added a space between "//" and the text in the comment to maintain consistency with the existing comment style. This change aims to improve code readability and maintain consistency. I hope this helps. Please review when you get a chance. Thanks!
2025-04-01T04:34:58.917912
2021-09-07T15:18:17
990096215
{ "authors": [ "rabelloo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9276", "repo": "onfido/castor", "url": "https://github.com/onfido/castor/pull/962" }
gharchive/pull-request
chore(deps): update jest Purpose Update Jest related dependencies. Approach npx npm-upgrade then yarn Testing CI Risks None. Looks good, so it wasn't picked up by Dependabot? It was, but it couldn't automatically resolve the Node version bump.
2025-04-01T04:34:58.921728
2023-05-18T13:45:25
1715667395
{ "authors": [ "m-Peter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9277", "repo": "onflow/cadence-tools", "url": "https://github.com/onflow/cadence-tools/pull/114" }
gharchive/pull-request
WIP: Retrieve Blockchain logs Work towards: https://github.com/onflow/developer-grants/issues/148 Description Provides a way to retrieve logs from the Blockchain. Also improves the look & feel for logs that come from test files. [ ] Targeted PR against master branch [ ] Linked to Github issue with discussion and accepted design OR link to spec that describes this work [ ] Code follows the standards mentioned here [ ] Updated relevant documentation [ ] Re-reviewed Files changed in the Github PR explorer [ ] Added appropriate labels Logs from test files are consistent with the blockchain logs. The INF logs are the blockchain logs, which we could return to the test file
2025-04-01T04:34:58.927758
2020-06-11T17:40:07
637208002
{ "authors": [ "orodio", "psiemens", "sideninja" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9278", "repo": "onflow/flow-go-sdk", "url": "https://github.com/onflow/flow-go-sdk/pull/53" }
gharchive/pull-request
Update flow.Transaction to implement json.Marshaler and json.Unmarshaler Closes: https://github.com/dapperlabs/flow-go/issues/3314 Description This PR updates flow.Transaction to implement the json.Marshaler and json.Unmarshaler interfaces defined by the built-in encoding/json package. I made the following design decisions: Empty identifiers (0x0000...) are marshalled to null Empty addresses (0x0000...) are marshalled to null An empty proposal key (flow.ProposalKey{}) is marshalled to null Identifiers and addresses are encoded as hexadecimal strings All other byte types (Transaction.Script, TransactionSignature.Signature) are encoded as base64 strings Cadence values (Transaction.Arguments) are encoded as JSON-CDC that is directly embedded in the transaction JSON structure Example output: { "script": "dHJhbnNhY3Rpb24geyBleGVjdXRlIHsgbG9nKCJIZWxsbywgV29ybGQhIikgfSB9", "arguments": [ { "type": "Int", "value": "42" } ], "referenceBlockID": "f0e4c2f76c58916ec258f246851bea091d14d4247a2fc3e18694461b1816e13b", "gasLimit": 42, "proposalKey": { "address": "0000000000000001", "keyID": 4, "sequenceNumber": 10 }, "payer": "0000000000000001", "authorizers": [ "0000000000000001" ], "payloadSignatures": [ { "address": "0000000000000001", "keyID": 4, "signature": "9yJTiMHWnVfmJRyf2lDLv54FEx5a24HlqgQiQC8EgWI=" } ], "envelopeSignatures": null } @orodio @JeffreyDoyle This PR was originally added to unblock users of the Go SDK who wanted to transport transactions as JSON. However, I'm realizing that the JSON representation of a transaction should be consistent across the SDKs. Can either of you shed some light on how you're representing transactions in the JS SDK? I'd like to update this PR to follow the format you're already using or define a common format if there isn't one already. This isn't a priority, so no rush to spend any time looking at this this week. @psiemens we represent transactions internally with this data structure: https://github.com/onflow/flow-js-sdk/tree/master/packages/interaction#internal-properties Is this still relevant? Should we merge it? @sideninja No, there is no longer a strong need for this so we can close it.
2025-04-01T04:34:58.957361
2015-09-23T17:43:11
107970278
{ "authors": [ "CGMLab", "TheAtomicNord", "onitake", "sde1000" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9279", "repo": "onitake/daliserver", "url": "https://github.com/onitake/daliserver/issues/9" }
gharchive/issue
Tridonic BM RS-232 with USB adapter Hello, First of all, thank you for maintaining this awesome library. It really is great. I've just got one small problem. I'm using a Tridonic BM RS-232 (which is is usually through serial) plugged into a RS-232 to USB adapter in hopes of using this library. Unfortunately, the daliserver library can't seem to find it no matter what I try. However, the official DALI Tool finds it and operates it perfectly fine. I'm just curious if you have any input or advice on this. Thanks! The BM RS-232 is a completely different device to the DALI-USB. daliserver doesn't have any code in it to drive it. It's likely that the BM RS-232 is using the same serial protocol as the RS-232 PS/S. This is documented in the data sheet - see the link here. The protocol looks quite straightforward, so writing a driver ought to be fairly simple. What kind of software do you want to use it with? I'm writing a simple voice controlled system to control the lighting in a media studio. Unfortunately, I'm simply a first year student in a computer science major so I'm very inexperienced when it comes to these things. I'll look into the link you gave me and if push comes to shove, we'll just buy the Tridonic USB Dali. Thanks! I've written an outline of a driver for the RS-232 device for use with the python-dali library. See the file dali/serial.py in the "serialdriver" branch here. Note that I don't have one of these, the code is completely untested, and I haven't included any code to handle unexpected responses from the device - that's up to you to do! Wow thanks a lot! I'll give this a shot as soon as I can! @CGMLab I'm going to close this issue now, as it's out of focus of daliserver. Perhaps I'll look into supporting the RS-232 protocol in the future, but I currently don't have access to the corresponding hardware.
2025-04-01T04:34:58.967049
2017-12-01T21:20:33
278599962
{ "authors": [ "Akin909", "bryphe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9280", "repo": "onivim/oni", "url": "https://github.com/onivim/oni/issues/1051" }
gharchive/issue
Theming: Allow specifying of token colors in json themes TextMate themes give us a finer granularity through 'scopes' in terms of colorized tokens, as opposed to the default conventional groups in Vim themes. We should take advantage of this by allowing specification of token colors in themes, and giving real fine-grained control. Two specific pieces of work: Today, TextMate themes only use existing highlight groups - we need to extend this capability so that they can specify new highlight groups TextMate themes already use the editor.tokenColors to colorize, but we need to pull that from the theme as well. @bryphe I've made some progress based on your advice and I'm able to add colors to the hover menu based on its types but am a bit confused (still sorry) re the desired outcome re the json settings my understanding of the desired functionality would be that user can set "editor.tokenColors": { "variable.object": "SomeColor" } and this token wherever it is used will be set to that color, if it is not set then we would want to take it from the theme.json, this though is only valid for the other ui pieces and not vim as it does have access to all these tokens so those are assigned to vim highlight groups. Where I'm at with this is I've currently got syntax highlighting based on vim highlight groups working ish still a work in progress which I can go into when I make a PR but for the theming is that the correct flow? which leads on to the current data structure which is an array of tokens, this means I can't reference the color values easily in styled components without looping inside the component 😦 so I've converted it to an object which looks like "token.name": { "scope": "someScopeAkaSameAsTokenName", "settings": "thisLooksLikeADefaultValue", "color": "actualHex" } My plan was then that when the theme is built up user values override any matching props, theme values next and finally anything without a value still but has a "setting" of a vim highlight group is populated with that so Ideally all tokens are assigned some default color. Does this sound right?, apologies for the wall of text. Wow, that's looking awesome, @Akin909 ! The syntax highlighting really takes the hover to another level πŸ‘ @bryphe I've made some progress based on your advice and I'm able to add colors to the hover menu based on its types but am a bit confused (still sorry) re the desired outcome re the json settings my understanding of the desired functionality would be that user can set Yep, it looks like we have the same understanding here! The precedence is the interesting part - it should go as follows: User configuration overrides should are highest priority Theme configuration in the JSON should be next highest Implicit themes that come from Vim should be the lowest My reasoning for the priority is that, if the user or the theme is explicitly setting a token, they are overriding the default behavior. In addition, the textmate scopes actually allow for finer-grained selection, so we'd want them to take precedence. I hope that makes sense. still a work in progress which I can go into when I make a PR but for the theming is that the correct flow? which leads on to the current data structure which is an array of tokens, this means I can't reference the color values easily in styled components without looping inside the component 😦 so I've converted it to an object Seems reasonable for our code that leverages it. I would prefer it to be an array for the actual files - the reason for this is to maintain compatibility with VSCode's theme json files. My hope is that we could leverage VSCode themes too - an example is https://github.com/Microsoft/vscode/blob/master/extensions/theme-solarized-light/themes/solarized-light-color-theme.json. Imagine how cool it would be if we could have both Vim and VSCode plugins in Oni, or at least a subset of both! But we can convert that to whatever internal representation we'd like. Converting it a dictionary is definitely more convenient for rendering so we don't have to keep re-iterating over it. Hope that helps, let me know if you have any questions. So right now the priority works as you've elaborated above πŸ‘ re the array in the files, I can merge the objects inside the array with the theme object we build in Oni so that internally the theme object is has keys that point towards a token which makes it easier to access, regarding the userConfig though, I think the utility of an array if offset by the unwieldiness of users having to right out JS arrays to add settings? (personal opinion alert)
2025-04-01T04:34:58.975594
2017-12-04T20:09:38
279131610
{ "authors": [ "Bretley", "badosu", "bryphe", "samvv" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9281", "repo": "onivim/oni", "url": "https://github.com/onivim/oni/pull/1072" }
gharchive/pull-request
[WIP] Externalize CommandLine and maybe wildmenu (?) This is the initial work I've done so far. I kind of just took bits and pieces and threw things at the wall to see what stuck. I'm sure this is going to need a ton of revision and a ton of extra work. [ ] Handle recursion with the CommandLine [ ] Handle the cursor/position within CommandLine [ ] Find out what values should be configurable [ ] Determine what CommandLine should look like/ where it should be placed. Sorry this is super bare in terms of commits, I very nearly deleted all of the changes I made... :man_shrugging: Wow, excellent progress @Bretley ! Looking really cool from the gif. And the code changes look great so far. I like the way you wired everything up - looks like you have a good handle on React + Redux πŸ˜‰ Left a few minor comments. I saw the cases for wildmenu_show, wildmenu_select, and wildmenu_hide - are you planning on including the wildmenu externalization in this change too? I'm okay either way. @bryphe Wildmenu might or might not get done depending on how your vision of this looks. There are three ways I could see this going: CommandLine becomes a Menu: Neovim sends over a cmdline event and the wildmenu functionality is instead replaced by fuzzy-style completion based on some list of keywords we have. This keeps 95% of the work in Oni and only requires that we have some precomputed list of help terms. We keep neovim's version and have it in the usual spot, but instead it's a cool little textbox instead of a part of the actual neovim screen. Even if we don't do much, externalizing the 'static' parts of the neovim screen is a step we need to take in order to get more advanced scrolling/ externalize the actual text display. Some hybrid of these two. It would be pretty easy to make cases for search vs ex mode vs. whatver ? means because neovim lets us know which commandline functionality is being used Personally, I don't want the search to be over the screen because it might be covering up text, but I do think that a popup commandline would be cool. This is really a design choice for you. Cool, nice summary @Bretley ! Initially, I think it makes sense to keep the command line separate (implementation-wise) from the quick open / command palette menu. There's two 'classes' of users I want to appeal to - Vim veterans who expect/know the command line, and new users who want to get start with modal editing - may be more comfortable in the quick open/command palette. Keeping that distinction now can satisfy both 'classes' of users, and then we can look towards ways to reconcile them later, if it makes sense. I do like the way you've structured the UI right now - having the command line pop in an overlay is really nice! Good point about the search, though. I wonder if, in that case, we could put it in the corner - sort of like how in Chrome it pops up in the corner? Regarding the wildmenu, if we use the UI-overlay like you've structured it, we could make the look and feel similiar to the quick open / command palette menu (even if the implementation is different) - I think it would help both classes of our users (for the vim veteran, getting adjusted to some of the new UX paradigms in Oni, and for the new user, it won't feel that different / scary from quick open or command palette). Hope that helps and wasn't too vague... Great work so far! Hope we can have a configuration to disable this without the overhead of it's implementation. Although I think this is a great feature it's very obstrusive visually for the edititing flow. I'll try to smoothen this out so that it can be merged. First need to see I can get it working. I'll close this out since we have #1177 and #1179 in now
2025-04-01T04:34:58.977266
2018-08-31T20:21:11
356112714
{ "authors": [ "Akin909" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9282", "repo": "onivim/oni", "url": "https://github.com/onivim/oni/pull/2536" }
gharchive/pull-request
Feature/vcs manager styling improvements Adds Icons to the VCS pane and replaces statusbar icons with react octicons πŸ‘ had to reign myself in not to put icons everywhere πŸ˜†
2025-04-01T04:34:58.983400
2020-02-17T15:58:31
566389607
{ "authors": [ "Sheepolution", "bryphe", "glennsl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9283", "repo": "onivim/oni2", "url": "https://github.com/onivim/oni2/issues/1338" }
gharchive/issue
Default navigation menu font is ugly The font used for the navigation menu, tabs, command palette and more, has always looked weird to me. I took a screenshot along with Sublime Text to compare the fonts. To me, Sublime Text's menu font looks way cleaner and I have less trouble reading it (besides that it's darker). At first I thought this was because the Onivim font is smaller, but when I compare the fonts better, I found that they're the same size.. Except that Onivim's font has wider spacing. Still, I don't like the current font. Especially the 'e'. I'd like to open this issue as a way for others to share their opinion on this. Perhaps I'm the only one who thinks this. If that's the case I hope in the future it will be possible to configure this font. You seem to be using a version of Oni2 that's at least several months old. We've recently switched rendering engine, partly to get better text rendering. And I think we've even changed font since then. Would you try with the latest nightly to see how that works for you? Whoops, sorry about that. I updated Onivim, but the problem still stands. Here are some more screenshots: Onivim2 / Sublime Text / VS Code Well, I'm not sure how this happened, but it looks even worse. The 'e' is now all blurry. And at the top/bottom there are these pixels sticking out. The latter also appears in the font for extensions (which seems to be a different and bigger font). I have to say that the extension icons could use a clean up as well, but that should probably be its own issue. Yeah, that's definitely worse. The font was changed about a month ago, in #1182, but just at a convenient way to get additional styles. The legibility of the UI font doesn't seem to have been given much weight yet, and I think the font rendering improvements with Skia has mostly been focused on the editor font. I'm not sure what requirements we have for the UI font (@bryphe?), but suggestions for good alternative fonts to try would certainly be appreciated. Preferably ones that are freely licensed. @glennsl I downloaded the newest version and see changes have been made to the font. In my opinion it looks fine and this issue can be closed. Thanks for the update, @Sheepolution !
2025-04-01T04:34:58.991923
2015-10-21T20:40:08
112674390
{ "authors": [ "chosenken", "jackboberg" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9284", "repo": "onmodulus/docker-run-php", "url": "https://github.com/onmodulus/docker-run-php/pull/3" }
gharchive/pull-request
Bug/bootstrap Will build an image with PHP 5.3, 5.4, 5.5, and 5.6 in /opt/modulus/php. Does this replace #2? Yea, it can replace #2. I closed that one. Working build! Docker image now has php-5.6.15 with gdm and mongodb support added. Updated start script to correctly update config files.
2025-04-01T04:34:59.012285
2022-05-11T05:22:50
1232024543
{ "authors": [ "jenkins-droid", "whitneywhtsang" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9285", "repo": "onnx/onnx-mlir", "url": "https://github.com/onnx/onnx-mlir/pull/1421" }
gharchive/pull-request
Upgrade LLVM level to 23aa5a744666 Updating LLVM to https://github.com/llvm/llvm-project/commit/23aa5a744666b281af807b1f598f517bf0d597cb causes a number of build issues. This patch attempts to modify onnx-mlir to use LLVM at https://github.com/llvm/llvm-project/commit/23aa5a744666b281af807b1f598f517bf0d597cb. This PR replaces uses from the MLIR Standard dialect to the new Func dialect (https://mlir.llvm.org/docs/Dialects/Func/). We want to upgrade LLVM to https://github.com/llvm/llvm-project/commit/9778ec057cf4214241e4a970f3e764e3cf150181 at the end. Signed-off-by: Whitney Tsang<EMAIL_ADDRESS> Jenkins Linux s390x Build #5728 [push] Upgrade LLVM level to 23... started at 13:32 Jenkins Linux ppc64le Build #4850 [push] Upgrade LLVM level to 23... passed after 1 hr 56 min
2025-04-01T04:34:59.029494
2021-10-15T17:30:46
1027652605
{ "authors": [ "mikehaertl", "onsails" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9286", "repo": "onsails/lspkind-nvim", "url": "https://github.com/onsails/lspkind-nvim/issues/23" }
gharchive/issue
Document font requirements Please forgive my ignorance but I'm probably missing something very obvious: The icons don't work at all for me. What's weird is that I can not even see any icons in the browser (e.g. on the project's start page on github): Could you maybe add some notes on which font is required to make this work? OS: Linux Mint Neovim 0.5.1 I understand your confusion, please take attention to this: -- can be either 'default' (requires nerd-fonts font) or -- 'codicons' for codicon preset (requires vscode-codicons font) in configuration example. There is also an issue #18 to have preset which doesn't require extra font and awesome input from @fenux, however I didn't have a time to properly test this approach. Thanks. I've seen this comment but somehow missed its meaning. Maybe a short introductory note would still be helpful. Anyway this can be closed then.
2025-04-01T04:34:59.036863
2024-06-06T01:16:40
2337105692
{ "authors": [ "ThatOneCalculator", "onuratakan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9289", "repo": "onuratakan/gpt-computer-assistant", "url": "https://github.com/onuratakan/gpt-computer-assistant/issues/48" }
gharchive/issue
Why just Ubuntu? This seems to work fine on any Linux distro that can run Python 3.9 (tested on Arch), so I'm not sure why the documentation is "Windows, macOS, and Ubuntu" instead of "Windows, macOS, and Linux" Actualy, you are so right. I just think about native app support. But linux users just like me and you can easily install it via terminal. Yeah, I just think that saying it's "for Ubuntu" instead of "for Linux" might discourage other Linux users from using it.
2025-04-01T04:34:59.038607
2023-09-23T07:49:59
1909757905
{ "authors": [ "diomed", "prototypa" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9290", "repo": "onwidget/astrowind", "url": "https://github.com/onwidget/astrowind/issues/277" }
gharchive/issue
back to top template would benefit from back to top button I would very much like if back to top would be in form of this SVG icon <svg xmlns="http://www.w3.org/2000/svg" class="icon icon-tabler icon-tabler-square-rounded-chevron-up" width="24" height="24" viewBox="0 0 24 24" stroke-width="2" stroke="currentColor" fill="none" stroke-linecap="round" stroke-linejoin="round"> <path stroke="none" d="M0 0h24v24H0z" fill="none"></path> <path d="M9 13l3 -3l3 3"></path> <path d="M12 3c7.2 0 9 1.8 9 9s-1.8 9 -9 9s-9 -1.8 -9 -9s1.8 -9 9 -9z"></path> </svg> Moved to discussions/ideas since it's not a bug with AstroWind
2025-04-01T04:34:59.070572
2019-10-11T20:25:45
506042987
{ "authors": [ "bassosimone", "hellais", "sarathms" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9291", "repo": "ooni/probe-desktop", "url": "https://github.com/ooni/probe-desktop/pull/53" }
gharchive/pull-request
Show middlebox measurement details Implements views for the two middlebox tests. This is how we show it on the mobile. However, the early desktop mockups suggest we wanted to also show details like the side-by-side comparison of what headers were sent and received etc. Do we want to do this now? @hellais @sarathms I believe we should create an issue for enhancing the view as you describe but we should make sure other higher priority endeavours are addressed first. (I.e. "sprint backlog"?) The green on the side looks a bit off to me in these screenshots as if they are two slightly different green colors. Yes. I had noticed it too. I tried debugging, but couldn't figure out immediately what is causing it. You can see that in the anomaly hero too. Will open an issue for this.
2025-04-01T04:34:59.088032
2018-05-11T13:54:33
322305793
{ "authors": [ "prashbabu123", "schukerzheng" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9292", "repo": "op/go-logging", "url": "https://github.com/op/go-logging/issues/127" }
gharchive/issue
log rotate functionality Is log rotate functionality is provided. If yes let me know how we can do it When you new a LogBackend , you can pass a rotated io.Writer, such as dailyrotate shown at https://github.com/kjk/dailyrotate.
2025-04-01T04:34:59.100291
2024-12-06T08:36:53
2722409810
{ "authors": [ "artem-astafev" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9293", "repo": "opea-project/GenAIExamples", "url": "https://github.com/opea-project/GenAIExamples/pull/1233" }
gharchive/pull-request
Added compose example for MultimodalQnA deployment on AMD ROCm systems Description Changes for MultimodalQnA example deployment to AMD ROCm systems. Type of change Added compose example for MultimodalQnA deployment on AMD ROCm systems. Added Readme file for MultimodalQnA compose deployment example on AMD ROCm systems. Added CI tests for MultimodalQnA compose deployment example on AMD ROCm systems. Tests For manual test s go to GenAIExamples/MultimodalQnA/tests dir and run test_compose_on_rocm.sh script (./test_compose_on_rocm.sh) Hi @artem-astafev , thanks for your contribution. Please check the comment here and refine the set_env part in your readme. Hi @letonghan , I've removed manual environment set part and add set_env.sh description in Readme.md in last 2 commits. I can't get what else can be change in Readme.md for better description. Is these enough or something else need to be changed ?
2025-04-01T04:34:59.105087
2021-08-11T09:15:45
966305150
{ "authors": [ "SayBender", "amirhesamyazdi", "kinshuk" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9294", "repo": "open-climate-tech/firecam", "url": "https://github.com/open-climate-tech/firecam/issues/10" }
gharchive/issue
Dataset is missing The dataset in the github includes only 37 unique locations from 72 unique days of fire, total of 1740 image frames. This is not same as reported in the paper. Do you happen to have the full smoke dataset used in the paper? And if so is it possible to share it? It is a subset. Unfortunately, all the data used in the paper cannot be shared publicly. However, there are some other datasets that are available. Also, if you are willing to volunteer some time, I can point you to some half-processed data and tools to label new data. If interested, please email the address listed in the contact section of website https://openclimatetech.org/ Thank you so much Dr. Govil for your quick response, I have another question. Is it possible to separately release the 250 image test set used to obtain the result reported in (Table 1. Confusion matrix on test set.) ? The paper says 250 images were used (100 smoke + 150 non-smoke) to get the result in table.1. I wonder if you could release this in the repo. (just the 250 images, not the entire dataset) Thank you so much Dr. Govil for your quick response, I have another question. Is it possible to separately release the 250 image test set used to obtain the result reported in (Table 1. Confusion matrix on test set.) ? The paper says 250 images were used (100 smoke + 150 non-smoke) to get the result in table.1. I wonder if you could release this in the repo. (just the 250 images, not the entire dataset)
2025-04-01T04:34:59.107345
2021-09-28T01:50:35
1008984948
{ "authors": [ "jnpacker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9295", "repo": "open-cluster-management/multicloud-integrations", "url": "https://github.com/open-cluster-management/multicloud-integrations/pull/13" }
gharchive/pull-request
Change roleBinding type (reduce scope of query to namespace) Signed-off-by: Joshua Packer<EMAIL_ADDRESS> The get is for the incorrect ClusterRoleBinding type, which is showing an error message and trying to create a duplicate resource. This doesn't cause a real failure today because we absorb the create error. The code should not be trying to create the roleBinding each time. Added a log message for a problem QE was seeing when the APIVersion was incorrect. Currently there is no error message Even though the ClusterRole is changed to a Role, the serviceAccount already has access (we create a RoleBinding in the next step) Issue Reference: https://github.com/open-cluster-management/backlog/issues/16627
2025-04-01T04:34:59.117870
2023-02-08T10:56:24
1575906024
{ "authors": [ "Skarlso" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9296", "repo": "open-component-model/ocm-controller", "url": "https://github.com/open-component-model/ocm-controller/pull/120" }
gharchive/pull-request
feat: add snapshot checking with conditions for controllers We have three main controllers that create snapshots after reconciliation is done. Since these names are predetermined they can be used to pick up if a snapshot has already been created or not. The next stage, uses the Ready condition state to check if a snapshot has been reconciled or not. If not, it's possible that it has to be recreated or pushed over because content is incorrect. In that case, reconciliation will proceed. If the snapshot exists and it's Ready we requeue the {Localization,Configuration,Resource}. Manual testing the tracking of component version: Before version update: 2023-02-09T11:50:54Z INFO resource-controller successfully pushed resource {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "8e0dea7a-cc1d-4616-a744-7543d8bea457", "resource": "deployment"} 2023-02-09T11:50:54Z INFO resource-controller successfully reconciled resource {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "8e0dea7a-cc1d-4616-a744-7543d8bea457", "name": "podinfo-deployment"} 2023-02-09T11:50:54Z INFO snapshot-reconcile reconciling snapshot {"controller": "snapshot", "controllerGroup": "delivery.ocm.software", "controllerKind": "Snapshot", "Snapshot": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "408354be-4ec7-4256-bf8b-2117cc87bf86"} 2023-02-09T11:50:55Z INFO ocm-component-version-reconcile starting ocm component loop {"controller": "componentversion", "controllerGroup": "delivery.ocm.software", "controllerKind": "ComponentVersion", "ComponentVersion": {"name":"podinfo","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo", "reconcileID": "9b372593-24b1-4e41-a90b-d7104a7f0cde"} 2023-02-09T11:50:55Z LEVEL(-4) ocm-component-version-reconcile found component {"controller": "componentversion", "controllerGroup": "delivery.ocm.software", "controllerKind": "ComponentVersion", "ComponentVersion": {"name":"podinfo","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo", "reconcileID": "9b372593-24b1-4e41-a90b-d7104a7f0cde", "component": {"apiVersion": "delivery.ocm.software/v1alpha1", "kind": "ComponentVersion", "namespace": "ocm-system", "name": "podinfo"}} 2023-02-09T11:50:55Z LEVEL(-4) ocm-component-version-reconcile current reconciled version is {"controller": "componentversion", "controllerGroup": "delivery.ocm.software", "controllerKind": "ComponentVersion", "ComponentVersion": {"name":"podinfo","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo", "reconcileID": "9b372593-24b1-4e41-a90b-d7104a7f0cde", "reconciled": "2.0.2"} 2023-02-09T11:50:56Z LEVEL(-4) ocm-component-version-reconcile got newest version from component {"controller": "componentversion", "controllerGroup": "delivery.ocm.software", "controllerKind": "ComponentVersion", "ComponentVersion": {"name":"podinfo","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo", "reconcileID": "9b372593-24b1-4e41-a90b-d7104a7f0cde", "version": "v2.0.2"} 2023-02-09T11:50:56Z INFO ocm-component-version-reconcile latest version already reconciled {"controller": "componentversion", "controllerGroup": "delivery.ocm.software", "controllerKind": "ComponentVersion", "ComponentVersion": {"name":"podinfo","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo", "reconcileID": "9b372593-24b1-4e41-a90b-d7104a7f0cde"} 2023-02-09T11:51:04Z INFO resource-controller starting resource reconcile loop {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "eb07ba08-7779-450a-961a-86b0c80354c7"} 2023-02-09T11:51:04Z INFO resource-controller component version already reconciled {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "eb07ba08-7779-450a-961a-86b0c80354c7"} We can see resource controller is saying it already reconciled said version... now pushing new version: 2023-02-09T11:53:47Z INFO resource-controller successfully pushed resource {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "49d1426e-c84d-41f1-b161-de4c2dbd6908", "resource": "deployment"} 2023-02-09T11:53:47Z INFO resource-controller successfully reconciled resource {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "49d1426e-c84d-41f1-b161-de4c2dbd6908", "name": "podinfo-deployment"} The resource controller detected a version diff, and reconciled it. Then: 2023-02-09T11:53:57Z INFO resource-controller starting resource reconcile loop {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "18307781-23e9-49df-ba02-e03928369f89"} 2023-02-09T11:53:57Z INFO resource-controller component version already reconciled {"controller": "resource", "controllerGroup": "delivery.ocm.software", "controllerKind": "Resource", "Resource": {"name":"podinfo-deployment","namespace":"ocm-system"}, "namespace": "ocm-system", "name": "podinfo-deployment", "reconcileID": "18307781-23e9-49df-ba02-e03928369f89"} Detected everything is fine. Noted two things: component version already reconciled should include the component version the snapshot controller however, said that the snapshot is already reconciled. so that should re-run the reconcile on the snapshot.
2025-04-01T04:34:59.129470
2022-05-10T10:11:59
1230928044
{ "authors": [ "jpmckinney", "sabahfromlondon" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9297", "repo": "open-contracting/data-registry", "url": "https://github.com/open-contracting/data-registry/issues/215" }
gharchive/issue
Monthly buckets of data The registry currently is only able to provide users with data in year buckets. This is not ideal for a two main reasons: In the interface users are asked to provide start and end dates for data they are search for i.e. start month and year and end month and year, so there is a mismatch between what they request and what they receive We end up providing users with too much data, which will lead to a poor user experience and users having to go through the dataset to delete large swathes of it potentially e.g. if a user selected a start date of Dec 2018 and end date of Jan 2019 (so data spanning two months only), then currently they will receive data for all of 2018 and all of 2019! As a result, we would like there to be monthly buckets of data, so that a user can receive only the data from the start month and year and up to only the end month and year. For example: IF a user selects a start date of Dec 2018 and end date of Jan 2019, THEN they will only receive data for those two months, BUT all in one sheet for CSV or one JSON file. The UI options in the screenshot below will need updated as a result of this change to monthly buckets. IF the user selected a date range in the faceted search, THEN the first option, automatically selected for the users should now present that date range. IF the users stick with this option and clicks "Download data", THEN they should be provided with the appropriate monthly buckets of data. We can still keep the current download options as second and third options for the user. IF the user did not select any date range in the faceted search (the default would then be "All"), THEN the current UI options should remain. IMPORTANT: We should have the same set of options for JSON and CSV while the Flatten tool is not ready to be launched. Noting that there are a couple places in the code that test for the presence of "_" in a filename (separator between year and month): in the flattener callback and files_available. When we're ready to convert monthly files, we can update the flattener callback to not convert files if the output already exists. That way, we can just publish messages to the queue to fill in the month files. As for the frontend, if I understand @sabahfromlondon, users can: Download data for all time Download data for a specific year Specify a start date and end date, and download all intervening months as one file We can't simply display all months like we do years, because, for example, job 713 has 210 files (2005-12 to 2022-06). https://github.com/open-contracting/data-registry/pull/235#discussion_r976714505 I figure users would just want all the data from start date to end date. If they need access to individual months, then we would need a new design. For reference, here is the current design at https://data.open-contracting.org/en/publication/22: Users can get all-time or a year in one click. For start/end, I figure we can have a small form with start date, end date and "download" button. The date fields can be pre-populated according to the search filters. In this way, if a user searches without any date filters, they still have the opportunity to set a specific range. The form can enforce a minimum/maximum according to the known date range, and it can perhaps repeat the date range, so that users don't need to scroll to the top to remember. @jpmckinney I'm a bit confused about what has happened because the requirement was for the user to select the date range as part of thier search. In the requirements document as part of the faceted search I had the options: Past month, Past 6 months, Past year, Past 5 years, All time - as part of a drop down. Currently the UI is only showing: All, Past year, Past six months using radio buttons - which I can live with. There was also a requirement for a custom date range option in the design using a calendar selector. I'm pretty sure this was developed but is not longer available. I'm not sure what happened to it! The date range for the download should be for what the user selected back on the search page. It's why we added the feedback labels on the datasets in case there was a partial coverage issue. I think it's odd for the user to have to re-select the date range again here. Is it becuase of a technical difficulty? The UI for a date range needs to be added back to the search page. It is temporarily missing. I had to rip out Vue in order to fix a variety of bugs, and didn't have time to add that functionality back yet. Now tracked in #249 We can add Past month and Past 5 years once #234 is closed. Now tracked in #250. I'm suggesting that if the user did not indicate a date range on the search page, then they can have the opportunity to set one on the detail page. Also, if they change their mind, they can do so without going back to the search page. I do see where you are coming from and editing the date range on the details page is not unprecendeted, but we did do extra work to support the user on the search page with the feedback labels for the coverage. The goal was for users to feel confident that the could access data for the date range they wanted. I hoped that there would be no need to the user to have to select again, expect in the rarest of circumstances. If the change as you say is to make the date range editable in the Access data section, then the options for the user should mirror what is on the search page. Thanks for adding the additional tickets :) Yup, the partial overlap logic will be restored along with the filter, and the options will be synced to avoid data re-entry for the common case. Sounds perfect! Relevant for large annual Excel files.
2025-04-01T04:34:59.150434
2020-05-28T11:18:20
626428360
{ "authors": [ "attwad" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9298", "repo": "open-covid-data/healthmap-gdo-temp", "url": "https://github.com/open-covid-data/healthmap-gdo-temp/pull/199" }
gharchive/pull-request
Enable authorization There's some good, bad and ugly in here. The UI needs to display errors more nicely because right now it just says 304 request unauthorized when you don't have perms to list something (Allyson and I will be looking into it in another PR, that's a general issue not related to authz). Passport (or its types) has a bug https://github.com/jaredhanson/passport/issues/776 which can lock you out of the UI until you remove your cookies manually (easy to do in chrome dev tools but still a bummer), to avoid this: test are now clearing out all users after they run, Allyson added an in-memory mongo to not affect the dev users but I believe passport should do the right thing instead of us having to work around it. Sessions are deleted as well, which is what was causing the issue to start with. Added a small documentation section that's meant as a way to quickly ramp-up on the authn/authz concepts used by the curator service. Integration tests were trivial, they use the same method as the unit tests to send a post request to the handler that creates a user and logs them in. ran npm run lint so there are small fixes unrelated to authz in there as well. In case you missed my PR from yesterday, you can run from the verification/scripts dir: mongo "mongodb://localhost:27017/covid19" --eval 'var email="${EMAIL}"; var roles=["admin", "curator", "reader"];' roles.js to give you full power in dev. Add authorization logic (i.e. require curator role to CRUD data)
2025-04-01T04:34:59.170513
2023-09-06T15:46:52
1884297384
{ "authors": [ "DaryaYuk", "beeme1mr" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9299", "repo": "open-feature/dotnet-sdk", "url": "https://github.com/open-feature/dotnet-sdk/issues/145" }
gharchive/issue
[DOC] Update readme to follow new template Change in the documentation Update the readme to follow the new template. Once this is ready, the documentation can be configured to pull from the repo automatically. Examples from other repos: https://github.com/open-feature/java-sdk https://github.com/open-feature/js-sdk/tree/main/packages/client Hi there! I am a junior technical writer and I want to help you. I'm excited to be assigned to work on it
2025-04-01T04:34:59.181617
2023-01-27T22:07:39
1560489363
{ "authors": [ "Kavindu-Dodan", "toddbaert" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9300", "repo": "open-feature/flagd", "url": "https://github.com/open-feature/flagd/pull/337" }
gharchive/pull-request
chore: add sbom generation for releases This PR Fixes #329 Adding SBOM generation through go releaser. SBOM is generated per archive (see image for reference). Along with the PR, changing deprecated archive naming [1] How to test Install goreleaser [2] and run following command tog get artefacts & sbom generates in local dist folder goreleaser release --skip-publish --skip-validate --rm-dist [1] - https://goreleaser.com/deprecations/#archivesreplacements [2] - https://goreleaser.com/install/ @Kavindu-Dodan should we also add SBOM generation for the container image, like it OFO? https://github.com/open-feature/open-feature-operator/blob/312e91e6f9c1c44b9d642c74e031742f78c0f7f9/.github/workflows/release-please.yml#L124 @Kavindu-Dodan Also I think the title should be "chore" or something like that. This is a "feature" in terms of security features, but not a new functional feature in flagd in the sense of semver or compatibility. @Kavindu-Dodan should we also add SBOM generation for the container image, like it OFO? https://github.com/open-feature/open-feature-operator/blob/312e91e6f9c1c44b9d642c74e031742f78c0f7f9/.github/workflows/release-please.yml#L124 Per my understanding, we are not scanning the image for SBOM generation in OFO but rather scanning only the source [1]. I will update this PR to introduce image scanning [2] . We might have to do the same (after validating things here) for OFO later on. [1] - https://github.com/anchore/sbom-action#basic-usage [2] - https://github.com/anchore/sbom-action#scan-a-container-image @Kavindu-Dodan should we also add SBOM generation for the container image, like it OFO? https://github.com/open-feature/open-feature-operator/blob/312e91e6f9c1c44b9d642c74e031742f78c0f7f9/.github/workflows/release-please.yml#L124 Per my understanding, we are not scanning the image for SBOM generation in OFO but rather scanning only the source [1]. I will update this PR to introduce image scanning [2] . We might have to do the same (after validating things here) for OFO later on. [1] - https://github.com/anchore/sbom-action#basic-usage [2] - https://github.com/anchore/sbom-action#scan-a-container-image Oh, interesting. I think you're right based on the doc. Your plan sounds good. @Kavindu-Dodan should we also add SBOM generation for the container image, like it OFO? https://github.com/open-feature/open-feature-operator/blob/312e91e6f9c1c44b9d642c74e031742f78c0f7f9/.github/workflows/release-please.yml#L124 Per my understanding, we are not scanning the image for SBOM generation in OFO but rather scanning only the source [1]. I will update this PR to introduce image scanning [2] . We might have to do the same (after validating things here) for OFO later on. [1] - https://github.com/anchore/sbom-action#basic-usage [2] - https://github.com/anchore/sbom-action#scan-a-container-image Oh, interesting. I think you're right based on the doc. Your plan sounds good. Updated the PR. I validated the workflow in a test project - Release artefcats [1] & Image scan [2]. Let's see the workflow in action once merged 🀞 [1] - https://github.com/Kavindu-Dodan/flagd-grpc-sync/releases/tag/v0.6 [2] - https://github.com/Kavindu-Dodan/flagd-grpc-sync/actions/runs/4047491430/jobs/6961565162#step:6:18
2025-04-01T04:34:59.185425
2022-06-02T20:17:12
1258667962
{ "authors": [ "beeme1mr", "toddbaert" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9301", "repo": "open-feature/spec", "url": "https://github.com/open-feature/spec/pull/92" }
gharchive/pull-request
Add sample provider OpenAPI schema A basic OpenAPI spec for OpenFeature provider-compliant HTTP communication. I called the resource /flags initially, but /flags/{flag-key}/resolve is a more technically correct path, because we're not actually creating a flag resource instance here, but performing an evaluation. I subscribe to the thinking here... Not sure if that feels too awkward. Another option would be /flag-resolutions/{flag-key}, which is arguably more RESTful. The resource essentially just evaluates a flag given a flag-key (path param), default-value (query param), and optional context JSON object in body. It responds with an OpenFeature compliant JSON body representing an resolution details object Here's a rendering in a swaggerUI: Rename the file to provider.yml. Now available here: https://github.com/open-feature/schemas
2025-04-01T04:34:59.204846
2016-10-19T08:15:09
183894718
{ "authors": [ "FernandoMiguel", "QuinnyPig", "ehammond", "jlevy", "pingles" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9302", "repo": "open-guides/og-aws", "url": "https://github.com/open-guides/og-aws/pull/237" }
gharchive/pull-request
Encourage temporary MFA-authenticated credentials on the command-line Most developers are familiar with using the Access and Secret Keys downloaded from the AWS Console. Security can be improved by: Requiring MFA be used when performing privileged API operations (ec2:TerminateInstances, rds:Delete* etc.). Using temporary credentials for everything except requesting temporary keys. You can obviously just use the AWS CLI tools to achieve this (by invoking the different operations) but we wrote a little CLI tool to make it easier to do. πŸ‘ These seems like good additions. That looks like a useful tool β€” though I'm a little hesitant to link to a GitHub tool that's brand new. Would you be up for committing the tips without the tool until we get more feedback on the tool? Or sharing a bit more on its maturity? Glad to chat more in Slack #security too! @ThanosBaskous @richadams @ehammond thoughts/review? Totally understand wrt it being new. It's possible to achieve the same directly using the AWS command-line tools (albeit a little more fiddly- which is where the tool came in). I could update with an example of achieving the same through the AWS command-line and example policy if that helps? Also happy to wait for other people to take a look at the tool first. Since I was @ mentioned... I totally support the recommendation to use temporary credentials and MFA where possible. I also agree that non-Amazon software recommendations should be avoided, unless they are completely accepted by the marketplace (e.g., "serverless"), especially when it relates to security management. @pingles Given the feedback, could you remove the tool for now and handle the copy editing tweaks? Of you do that glad to merge. Thanks again. Also, please consider sharing your tool https://github.com/uswitch/stscreds on the Slack channel to get more feedback or others to test it. That could give it some validation! Yep, that's fine. I'll rework and update here. Good suggestion re:slack, will do that now. @pingles if you have time to revisit this, appreciate it. I should have time today, thanks for the reminder :-) On Tue, 22 Nov 2016 at 03:36, Joshua Levy<EMAIL_ADDRESS>wrote: @pingles https://github.com/pingles if you have time to revisit this, appreciate it. β€” You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/open-guides/og-aws/pull/237#issuecomment-262139095, or mute the thread https://github.com/notifications/unsubscribe-auth/AAAEfip1yYkodDTkCbyV86bXQcPbclS7ks5rAmNOgaJpZM4Karo1 . @pingles Love to see this get merged this week! I've been using aws-vault which uses STS and keychain to keep the keys with lots of success. Highly recommend it Going to close this for now until @pingles resurfaces; feel free to reopen.
2025-04-01T04:34:59.206682
2017-03-07T06:13:58
212340448
{ "authors": [ "bgdnlp", "jlevy" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:9303", "repo": "open-guides/og-aws", "url": "https://github.com/open-guides/og-aws/pull/390" }
gharchive/pull-request
Remove the limitation tip about NFSv4 file locking As per https://aws.amazon.com/about-aws/whats-new/2017/03/amazon-elastic-file-system-amazon-efs-now-supports-nfsv4-lock-upgrading-and-downgrading/ LGTM. Thanks @bgdnlp ! @ThanosBaskous I'll merge since bgdnlp is editor but feel free to post-hoc review. :)